Model Selection with Cross-Validations and Bootstraps Application to Time Series Prediction with RBFN Models

Size: px
Start display at page:

Download "Model Selection with Cross-Validations and Bootstraps Application to Time Series Prediction with RBFN Models"

Transcription

1 Model Selecton wth Cross-Valdatons and Bootstraps Applcaton to Tme Seres Predcton wth RBF Models Amaury Lendasse Vncent Wertz and Mchel Verleysen Unversté catholque de Louvan CESAME av. G. Lemaître 3 B-348 Louvan-la-euve Belgum {lendasse wertz}@auto.ucl.ac.be DCE pl. du Levant 3 B-348 Louvan-la-euve Belgum verleysen@dce.ucl.ac.be Abstract. Ths paper compares several model selecton methods based on expermental estmates of ther generalzaton errors. Experments n the context of nonlnear tme seres predcton by Radal-Bass Functon etworks show the superorty of the bootstrap methodology over classcal cross-datons and leave-one-out. ntroducton onlnear modelng has rased a consderable research effort snce decades ncludng n the feld of artfcal neural networks. onlnear modelng ncludes the necessty to compare models (for example of dfferent complextes) n order to select the best model among several ones. For ths purpose t s necessary to obtan a good approxmaton of the generalzaton error (or expected loss ) of each model (the generalzaton error beng the average error that the model would make on an nfntesze and unknown test set). n ths paper the terms model selecton wll be used when several models must be compared based on estmatons of ther generalzaton errors n order to select one of them. owadays there exst some well-known and wdely used methods able to fulfll ths task. Among them we can cte: œ the hold-out (HO) whch consst n removng data from the ng set and keepng them for daton; HO s also called daton [] or external daton for example n chemometrcs etc. œ the Monte-Carlo cross-daton (or smply cross-daton CV) where several HO daton sets are randomly and sequentally drawn; œ the k-fold cross-daton where the ntal set s randomly splt nto k roughly equal parts each one beng used successvely as a daton set; œ Leave-One-Out (LOO) s a k-fold cross-daton where the sze of the daton set s ; œ the bootstrap [ 3] whch conssts n drawng sets wth replacement from the orgnal sample and usng these sets to estmate the generalzaton errors (boostrap 63 and 63+ are mproved versons of the orgnal bootstrap): O. aynak et al. (Eds.): CA/COP 003 LCS 74 pp Sprnger-Verlag Berln Hedelberg 003

2 574 A. Lendasse V. Wertz and M. Verleysen All these methods of estmatng generalzaton errors have been shown to be asymptotcally roughly equent (see for example [4]) wth some exceptons and lmtatons: œ LOO s less based [5 ] but ts varance s unacceptable [6]; œ cross-daton s consstent (.e. converges to the generalzaton error when the sze of the sample ncreases) f the sze of the daton set grows nfntely faster than the sze of the ng set (whch s counter-ntutve!) [7]; œ cross-daton s almost unbased []; œ bootstrap s downward based but has a very low varance []; œ most recent bootstrap methods (63 and 63+) are almost unbased and also have a low varance []. Gven ths lst of possble model selecton methods and crtera the purpose of ths paper s to compare expermentally these methods n the context of () hghly nonlnear regresson (makng complex model structures unavodable) and () a small number of data ponts or nput-output pars (whch s often the case n real-world applcatons n medum- or hgh-dmensonal spaces). An mportant argument n favor of smple relable model selecton methods s the followng. n the context of nonlnear regresson or model dentfcaton the ng tme (or computatonal complexty) may be far from neglgble. n some cases for example when one tres a sngle model structure and has a large daton set at dsposal ths mght not be a concern. However n most stuatons frst one has to try many dfferent model structures (for example of dfferent complextes dfferent number of neurons n the hdden layers of mult-layer perceptrons etc.); and secondly the use of resamplng methods (CV LOO bootstrap ) s necessary because of the small sze of the orgnal sample at dsposal. Ths mght multply the ng tme by several hundreds or thousands makng computaton tme a real concern even on upto-date powerful machnes. Therefore ths paper wll consder the varous model selecton methods n the context of nonlnear regresson wth relatvely small sample from the pont of vew of performances and computatonal complexty. Ths wll be llustrated on a standard tme seres predcton benchmark the Santa Fe A seres [8]. Ths partcular example has been chosen for llustraton because [8] () the deal regressor s known avodng the supplementary problem of choosng ts sze () f the rules of the Santa Fe competton are followed we have a small set of data (000) (3) the relaton to s hghly nonlnear and (4) makng the hypothess of a statonary process we have a very large test set at dsposal (8000 data) used to obtan a very good approxmaton of the generalzaton error (ndependently from the model selecton method) for comparson purposes. Accordng to our prevous experence [9] we chose to approxmate ths seres by a Radal-Bass Functon network wth 00 Gaussan kernels the ng beng acheved by the method presented n [0]. Model Selecton Methods Ths secton presents the dfferent methods enumerated n the ntroducton. The ntal data used for the ng phase are the pars (x ) wth x representng the d-

3 Model Selecton wth Cross-Valdatons and Bootstraps 575 dmensonal nput vectors and y the scalar outputs. Theses pars form the ng dataset X. Each of the methods below computes an approxmaton of the generalzaton error obtaned wth a model g. The generalzaton error s defned by: E gen ( ( ) y) lm () gx wth (x ) the elements of an nfnte test dataset and g(x ) the approxmaton of y obtaned wth the model g. The selected model wll be the model mnmzng ths estmate of the generalzaton error.. Monte-Carlo Cross-Valdaton The consecutve steps of the Monte-Carlo cross-daton method are:. One randomly draws wthout replacement some elements of the dataset X; these elements form a new ng dataset X. The remanng elements of X form the daton set X. Usually two thrd of the elements of X are used n X and one thrd n X []; ths rule wll be used n the followng.. The tranng of the model g s done usng X and the error E k (g) s calculated accordng to: E k /3 ( gx ( ) y ) /3 () wth ( x ) the elements of X and gx ( ) the approxmaton of y by model g. 3. Steps and are repeated tmes wth as large as possble. The error E k (g) s computed for each repetton k. The average error s defned by: Eˆ gen k Ek. A partcular case of the Monte-Carlo cross-daton method s the Hold-Out method where the number of repettons s equal to. (3). k-fold Cross-Valdaton The k-fold cross-daton method s a varant of the Monte-Carlo cross-daton method. The consecutve steps of ths method are:. One dvdes the elements of the dataset X nto sets of roughly equal sze. The elements of the k th set form the daton set X. The other sets form a new ng dataset X.. The tranng of the model g s done usng X and the error E k (g) s calculated accordng to:

4 576 A. Lendasse V. Wertz and M. Verleysen E k / ( gx ( ) y ) / (4) wth ( x ) the elements of X and gx ( ) the approxmaton of y by model g. 3. Steps and are repeated for k varyng from to. The average error s computed accordng to (3). A partcular case of the k-fold cross-daton method s the Leave-One-Out where s equal to..3 Bootstrap The consecutve steps of the bootstrap method developed by Efron [] are:. n the dataset X one draws randomly samples wth replacement. These new samples form a new ng set X wth the same sze as the orgnal one. The daton set X s the orgnal ng set X. Ths process s called re-samplng.. The tranng of the model g s done usng X and the errors Ek ( g ) and Ek obtaned wth ths model are calculated accordng to the followng equatons: E k ( gx ( ) ) y (5) wth ( x ) the elements of X and gx ( ) the approxmaton of y obtaned by model g; E k ( gx ( ) ) y (6) wth ( x ) the elements of X and gx ( ) the approxmaton of y by model g. 3. The optmsm D k (g) a measure of performance degradaton (for the same model) between a ng and a daton set s computed accordng to: D E E. (7) k k k 4. Steps and 3 are repeated tmes wth as large as possble. The average optmsm Dg ˆ ( ) s computed by: Dg ˆ ( ) k Dk. (8)

5 Model Selecton wth Cross-Valdatons and Bootstraps Once ths average optmsm s computed a new tranng of the model g s done usng the ntal dataset X; the ng error E ( g ) s calculated accordng to: m E m ( gx ( ) ) y (9) wth ( x ) the elements of X and gx ( ) the approxmaton of y obtaned usng the model g. 6. Step 5 s repeated repeated M tmes wth M as large as possble. For each repetton m the ng error Em ( g ) s computed. The apparent error ˆ E ( g ) s defned as the average of errors Em ( g ) over the M repettons. n the case of a lnear model g ths repetton s not necessary; ng of a lnear model gves a unque set of parameters makng all ng errors Em ( g ) equal. Wth nonlnear models ths repetton performs a (Monte-Carlo) estmate of the most probable apparent error obtaned after tranng of g. 7. ow we have an estmate of the apparent error and of the optmsm ther sum gves an estmate of the generalzaton error: ˆ ˆ E E + Dˆ. (0) gen A partcular case of bootstrap method s the bootstrap 63 [3]. n ths method Ê gen (g) s calculated accordng to: ˆ ˆ ˆ 63 E.368 E +.63 D () gen 63 wth Dˆ ( g ) an optmsm estmated only on the data that are not selected durng the re-samplng (see [3] for detals). 3 Radal-Bass Functon etworks and Tme Seres Predcton The nonlnear models that have been chosen n ths paper are the RBF []. These approxmators have the well-known property of unversal approxmaton. Other approxmators as Mult-Layers Perceptrons could have been chosen. However the goal of ths paper s not to compare dfferent famles of approxmators: RBF have been chosen for the smplcty of ther tranng phase. ndeed a possble ng algorthm for RBF [0] conssts n a three-folded strategy (separated computaton of kernel centers wdths and weghts). As a consequence the weghts are smply calculated as the soluton of a lnear system. The problem of tme seres predcton s a common problem n many dfferent felds as fnance electrcty producton hydrology etc. A set of successve ponts (the tme seres) s known; the goal s to predct the next unknown ue. To perform ths task a model s needed that assocates some prevous ues of the tme seres to the next one. Ths model has the followng form:

6 578 A. Lendasse V. Wertz and M. Verleysen yt ( ) g( yt ( )t ( )...t ( n)) () where n s the lag order []. Model g can be lnear or nonlnear. n ths paper the class of models we have chosen s a set of RBF wth dfferent numbers C of Gaussan kernels but wth a gven sze for the regressor. The problem of model selecton s then to choose one of the models n the set.e.: how many Gaussan kernels must be used? 4 Expermental Results We llustrate the methods descrbed n the prevous secton on a standard benchmark n tme-seres predcton. The Santa Fe A tme seres [8] has been chosen manly for the large number of data aable for the tranng stage (000) as well as for the test stage (8000). These two numbers correspond to the rules of the Santa Fe competton as detaled n [8]. Accordng to [8] we also choose n 6. We traned 7 RBFs on the Santa Fe ng dataset for C and 40 respectvely. Ths number of models s kept small to avod a too large computatonal tme; ndeed the goal s not to fnd the best among an nfnte set of models but to compare the performances of the dfferent model selecton methods. Havng a huge (8000 samples) test set makes t possble to have a very good approxmaton of the true generalzaton error: 8000 ( gx ( ) y) ( gx ( ) y) Egen lm. (3) n practce for smlar reasons as those argued n favour of a Monte-Carlo euaton (step 6) of the apparent error ˆ E ( g ) n the bootstrap (use of nonlnear models) a Monte-Carlo method (wth M 00 repettons) s used to euate E ( ) gen g for each model. Then accordng to Fg. (a) the best RBF model has a number C of Gaussan kernels equal to Error 40 Error umber of Gaussan ernels (a) umber of Gaussan ernels Fg.. (a) Estmaton of the true generalzaton error; (b) Estmaton of the generalzaton error by a Monte-Carlo cross-daton method. (b) Fg. (b) shows the estmaton of the generalzaton error by a Monte-Carlo crossdaton method. The optmal number of Gaussan kernel s 80 (another number of

7 Model Selecton wth Cross-Valdatons and Bootstraps 579 kernels that could be chosen -accordng to the parsmony prncple- s 40 correspondng to a local mnmum). Fgure (a) shows the estmaton of the generalzaton error by a Leave-One-Out method. The optmal number of Gaussan kernels s 80. Fg.. (a) Estmaton of the generalzaton error by Leave-One-Out; (b) Estmaton of the generalzaton error by the classcal bootstrap. Fg. (b) shows the estmaton of the generalzaton error by the classcal bootstrap. The optmal number of Gaussan kernels s between 80 and 0. n ths case agan accordng to the parsmony prncple t s advsed to choose C 80. Fgure 3 shows the estmaton of the generalzaton error by the bootstrap 63. The optmal number of Gaussan kernels s 00. Error umber of Gaussan ernels Fg. 3. Estmaton of the generalzaton error by the bootstrap Concluson A number of smlar experments have been conducted wth the purpose of comparng model selecton methods based on emprcal estmatons of the generalzaton error: Monte-Carlo cross-daton Leave-One-Out bootstrap bootstrap 63. They all lead to smlar conclusons: bootstrap methods and n partcular the bootstrap 63 gve the best estmatons of the generalzaton error. Although theoretcal results (brefly summarzed n the ntroducton) show that all these methods are roughly asymptotcally equent ths s no more the case n real applcatons wth lmted number of data. Concernng the computaton tme the number of repettons n the cross-daton and bootstraps may be tuned whch s not the case wth the Leave-One-Out. Ths

8 580 A. Lendasse V. Wertz and M. Verleysen allows n our 000 ng data experments to strongly reduce the computatonal costs assocated wth cross-daton and Bootstraps. However n accordance wth theoretcal results cross-daton has a much hgher varance than bootstraps. Usng the same number of repettons n crossdaton and bootstraps hghlghts the advantages of the bootstraps. Such results have also been publshed n dfferent contexts namely classfcaton [5]. Ths paper extends these results to regresson and tme seres predcton n partcular. Acknowledgements. Mchel Verleysen s Senor research assocate at the Belgan atonal Fund for Scentfc Research (FRS). The work of A. Lendasse and V. Wertz s supported by the nterunversty Attracton Poles (AP) ntated by the Belgan Federal State Mnstry of Scences Technologes and Culture. The scentfc responsblty rests wth the authors. References. Ljung L.: System dentfcaton - Theory for the user nd ed Prentce Hall Efron B. Tbshran R. J. An ntroducton to the bootstrap Chapman & Hall Efron B Tbshran R.: mprovements on cross-daton: The.63+ bootstrap method. J. Amer. Statst. Assoc. (997) 9: Stone M.: An asymptotc equence of choce of model by cross-daton and Akake s crteron J. Royal. Statst. Soc. B ohav R.: A study of Cross-Valdaton and Bootstrap for Accuracy Estmaton and Model Selecton Proc. of the 4th nt. Jont Conf. on A.. Vol. Canada Efron B.: Estmatng the error rate of a predcton rule: mprovements on cross-daton. Journal of Amercan Statstcal Assocaton (983) 78(38): Shao J. Tu D.: The Jackknfe and Bootstrap Sprnger Seres n Statstcs Sprnger- Verlag ew York (995). 8. Wegend A. S. Gershenfeld. A.: Tmes Seres Predcton: Forecastng the future and Understandng the Past Addson-Wesley Publshng Company (994). 9. Smon G. Lendasse A. Wertz V. Verleysen M.: Fast approxmaton of the bootstrap for model selecton submtted to ESA03 Bruges (Belgum) Aprl Benoudjt. Archambeau C. Lendasse A. Lee J. Verleysen M.: Wdth optmzaton of the Gaussan kernels n Radal Bass Functon etworks ESA 00 European Symposum on Artfcal eural etworks Bruges (00) Moody J. Darken C.: Learnng wth localsed receptve felds Proceedngs of the 988 Connectonst Models Summer School San Mateo CA (989).

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

Mathematics 256 a course in differential equations for engineering students

Mathematics 256 a course in differential equations for engineering students Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

Simulation: Solving Dynamic Models ABE 5646 Week 11 Chapter 2, Spring 2010

Simulation: Solving Dynamic Models ABE 5646 Week 11 Chapter 2, Spring 2010 Smulaton: Solvng Dynamc Models ABE 5646 Week Chapter 2, Sprng 200 Week Descrpton Readng Materal Mar 5- Mar 9 Evaluatng [Crop] Models Comparng a model wth data - Graphcal, errors - Measures of agreement

More information

Classification / Regression Support Vector Machines

Classification / Regression Support Vector Machines Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM

More information

A Statistical Model Selection Strategy Applied to Neural Networks

A Statistical Model Selection Strategy Applied to Neural Networks A Statstcal Model Selecton Strategy Appled to Neural Networks Joaquín Pzarro Elsa Guerrero Pedro L. Galndo joaqun.pzarro@uca.es elsa.guerrero@uca.es pedro.galndo@uca.es Dpto Lenguajes y Sstemas Informátcos

More information

Programming in Fortran 90 : 2017/2018

Programming in Fortran 90 : 2017/2018 Programmng n Fortran 90 : 2017/2018 Programmng n Fortran 90 : 2017/2018 Exercse 1 : Evaluaton of functon dependng on nput Wrte a program who evaluate the functon f (x,y) for any two user specfed values

More information

Edge Detection in Noisy Images Using the Support Vector Machines

Edge Detection in Noisy Images Using the Support Vector Machines Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona

More information

Using Neural Networks and Support Vector Machines in Data Mining

Using Neural Networks and Support Vector Machines in Data Mining Usng eural etworks and Support Vector Machnes n Data Mnng RICHARD A. WASIOWSKI Computer Scence Department Calforna State Unversty Domnguez Hlls Carson, CA 90747 USA Abstract: - Multvarate data analyss

More information

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET 1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School

More information

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics Introducton G10 NAG Fortran Lbrary Chapter Introducton G10 Smoothng n Statstcs Contents 1 Scope of the Chapter... 2 2 Background to the Problems... 2 2.1 Smoothng Methods... 2 2.2 Smoothng Splnes and Regresson

More information

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS ARPN Journal of Engneerng and Appled Scences 006-017 Asan Research Publshng Network (ARPN). All rghts reserved. NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS Igor Grgoryev, Svetlana

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

EXTENDED BIC CRITERION FOR MODEL SELECTION

EXTENDED BIC CRITERION FOR MODEL SELECTION IDIAP RESEARCH REPORT EXTEDED BIC CRITERIO FOR ODEL SELECTIO Itshak Lapdot Andrew orrs IDIAP-RR-0-4 Dalle olle Insttute for Perceptual Artfcal Intellgence P.O.Box 59 artgny Valas Swtzerland phone +4 7

More information

SVM-based Learning for Multiple Model Estimation

SVM-based Learning for Multiple Model Estimation SVM-based Learnng for Multple Model Estmaton Vladmr Cherkassky and Yunqan Ma Department of Electrcal and Computer Engneerng Unversty of Mnnesota Mnneapols, MN 55455 {cherkass,myq}@ece.umn.edu Abstract:

More information

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration Improvement of Spatal Resoluton Usng BlockMatchng Based Moton Estmaton and Frame Integraton Danya Suga and Takayuk Hamamoto Graduate School of Engneerng, Tokyo Unversty of Scence, 6-3-1, Nuku, Katsuska-ku,

More information

Optimal Design of Nonlinear Fuzzy Model by Means of Independent Fuzzy Scatter Partition

Optimal Design of Nonlinear Fuzzy Model by Means of Independent Fuzzy Scatter Partition Optmal Desgn of onlnear Fuzzy Model by Means of Independent Fuzzy Scatter Partton Keon-Jun Park, Hyung-Kl Kang and Yong-Kab Km *, Department of Informaton and Communcaton Engneerng, Wonkwang Unversty,

More information

Backpropagation: In Search of Performance Parameters

Backpropagation: In Search of Performance Parameters Bacpropagaton: In Search of Performance Parameters ANIL KUMAR ENUMULAPALLY, LINGGUO BU, and KHOSROW KAIKHAH, Ph.D. Computer Scence Department Texas State Unversty-San Marcos San Marcos, TX-78666 USA ae049@txstate.edu,

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

Validating and Understanding Software Cost Estimation Models based on Neural Networks

Validating and Understanding Software Cost Estimation Models based on Neural Networks Valdatng and Understandng Software Cost Estmaton Models based on eural etworks Al Idr Department of Software Engneerng ESIAS, Mohamed V Unversty Rabat, Morocco E-mal : dr@ensas.ma Alan Abran École de Technologe

More information

2x x l. Module 3: Element Properties Lecture 4: Lagrange and Serendipity Elements

2x x l. Module 3: Element Properties Lecture 4: Lagrange and Serendipity Elements Module 3: Element Propertes Lecture : Lagrange and Serendpty Elements 5 In last lecture note, the nterpolaton functons are derved on the bass of assumed polynomal from Pascal s trangle for the fled varable.

More information

A Modified Median Filter for the Removal of Impulse Noise Based on the Support Vector Machines

A Modified Median Filter for the Removal of Impulse Noise Based on the Support Vector Machines A Modfed Medan Flter for the Removal of Impulse Nose Based on the Support Vector Machnes H. GOMEZ-MORENO, S. MALDONADO-BASCON, F. LOPEZ-FERRERAS, M. UTRILLA- MANSO AND P. GIL-JIMENEZ Departamento de Teoría

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour 6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the

More information

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization Problem efntons and Evaluaton Crtera for Computatonal Expensve Optmzaton B. Lu 1, Q. Chen and Q. Zhang 3, J. J. Lang 4, P. N. Suganthan, B. Y. Qu 6 1 epartment of Computng, Glyndwr Unversty, UK Faclty

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

INTELLECT SENSING OF NEURAL NETWORK THAT TRAINED TO CLASSIFY COMPLEX SIGNALS. Reznik A. Galinskaya A.

INTELLECT SENSING OF NEURAL NETWORK THAT TRAINED TO CLASSIFY COMPLEX SIGNALS. Reznik A. Galinskaya A. Internatonal Journal "Informaton heores & Applcatons" Vol.10 173 INELLEC SENSING OF NEURAL NEWORK HA RAINED O CLASSIFY COMPLEX SIGNALS Reznk A. Galnskaya A. Abstract: An expermental comparson of nformaton

More information

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 48 CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 3.1 INTRODUCTION The raw mcroarray data s bascally an mage wth dfferent colors ndcatng hybrdzaton (Xue

More information

X- Chart Using ANOM Approach

X- Chart Using ANOM Approach ISSN 1684-8403 Journal of Statstcs Volume 17, 010, pp. 3-3 Abstract X- Chart Usng ANOM Approach Gullapall Chakravarth 1 and Chaluvad Venkateswara Rao Control lmts for ndvdual measurements (X) chart are

More information

Tuning of Fuzzy Inference Systems Through Unconstrained Optimization Techniques

Tuning of Fuzzy Inference Systems Through Unconstrained Optimization Techniques Tunng of Fuzzy Inference Systems Through Unconstraned Optmzaton Technques ROGERIO ANDRADE FLAUZINO, IVAN NUNES DA SILVA Department of Electrcal Engneerng State Unversty of São Paulo UNESP CP 473, CEP 733-36,

More information

Reducing Frame Rate for Object Tracking

Reducing Frame Rate for Object Tracking Reducng Frame Rate for Object Trackng Pavel Korshunov 1 and We Tsang Oo 2 1 Natonal Unversty of Sngapore, Sngapore 11977, pavelkor@comp.nus.edu.sg 2 Natonal Unversty of Sngapore, Sngapore 11977, oowt@comp.nus.edu.sg

More information

Empirical Distributions of Parameter Estimates. in Binary Logistic Regression Using Bootstrap

Empirical Distributions of Parameter Estimates. in Binary Logistic Regression Using Bootstrap Int. Journal of Math. Analyss, Vol. 8, 4, no. 5, 7-7 HIKARI Ltd, www.m-hkar.com http://dx.do.org/.988/jma.4.494 Emprcal Dstrbutons of Parameter Estmates n Bnary Logstc Regresson Usng Bootstrap Anwar Ftranto*

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

A Similarity-Based Prognostics Approach for Remaining Useful Life Estimation of Engineered Systems

A Similarity-Based Prognostics Approach for Remaining Useful Life Estimation of Engineered Systems 2008 INTERNATIONAL CONFERENCE ON PROGNOSTICS AND HEALTH MANAGEMENT A Smlarty-Based Prognostcs Approach for Remanng Useful Lfe Estmaton of Engneered Systems Tany Wang, Janbo Yu, Davd Segel, and Jay Lee

More information

Correlative features for the classification of textural images

Correlative features for the classification of textural images Correlatve features for the classfcaton of textural mages M A Turkova 1 and A V Gadel 1, 1 Samara Natonal Research Unversty, Moskovskoe Shosse 34, Samara, Russa, 443086 Image Processng Systems Insttute

More information

The Man-hour Estimation Models & Its Comparison of Interim Products Assembly for Shipbuilding

The Man-hour Estimation Models & Its Comparison of Interim Products Assembly for Shipbuilding Internatonal Journal of Operatons Research Internatonal Journal of Operatons Research Vol., No., 9 4 (005) The Man-hour Estmaton Models & Its Comparson of Interm Products Assembly for Shpbuldng Bn Lu and

More information

Recommended Items Rating Prediction based on RBF Neural Network Optimized by PSO Algorithm

Recommended Items Rating Prediction based on RBF Neural Network Optimized by PSO Algorithm Recommended Items Ratng Predcton based on RBF Neural Network Optmzed by PSO Algorthm Chengfang Tan, Cayn Wang, Yuln L and Xx Q Abstract In order to mtgate the data sparsty and cold-start problems of recommendaton

More information

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints Australan Journal of Basc and Appled Scences, 2(4): 1204-1208, 2008 ISSN 1991-8178 Sum of Lnear and Fractonal Multobjectve Programmng Problem under Fuzzy Rules Constrants 1 2 Sanjay Jan and Kalash Lachhwan

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

A Robust LS-SVM Regression

A Robust LS-SVM Regression PROCEEDIGS OF WORLD ACADEMY OF SCIECE, EGIEERIG AD ECHOLOGY VOLUME 7 AUGUS 5 ISS 37- A Robust LS-SVM Regresson József Valyon, and Gábor Horváth Abstract In comparson to the orgnal SVM, whch nvolves a quadratc

More information

Feature Selection as an Improving Step for Decision Tree Construction

Feature Selection as an Improving Step for Decision Tree Construction 2009 Internatonal Conference on Machne Learnng and Computng IPCSIT vol.3 (2011) (2011) IACSIT Press, Sngapore Feature Selecton as an Improvng Step for Decson Tree Constructon Mahd Esmael 1, Fazekas Gabor

More information

TECHNIQUE OF FORMATION HOMOGENEOUS SAMPLE SAME OBJECTS. Muradaliyev A.Z.

TECHNIQUE OF FORMATION HOMOGENEOUS SAMPLE SAME OBJECTS. Muradaliyev A.Z. TECHNIQUE OF FORMATION HOMOGENEOUS SAMPLE SAME OBJECTS Muradalyev AZ Azerbajan Scentfc-Research and Desgn-Prospectng Insttute of Energetc AZ1012, Ave HZardab-94 E-mal:aydn_murad@yahoocom Importance of

More information

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for

More information

APPLICATION OF PREDICTION-BASED PARTICLE FILTERS FOR TELEOPERATIONS OVER THE INTERNET

APPLICATION OF PREDICTION-BASED PARTICLE FILTERS FOR TELEOPERATIONS OVER THE INTERNET APPLICATION OF PREDICTION-BASED PARTICLE FILTERS FOR TELEOPERATIONS OVER THE INTERNET Jae-young Lee, Shahram Payandeh, and Ljljana Trajovć School of Engneerng Scence Smon Fraser Unversty 8888 Unversty

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

An Ensemble Learning algorithm for Blind Signal Separation Problem

An Ensemble Learning algorithm for Blind Signal Separation Problem An Ensemble Learnng algorthm for Blnd Sgnal Separaton Problem Yan L 1 and Peng Wen 1 Department of Mathematcs and Computng, Faculty of Engneerng and Surveyng The Unversty of Southern Queensland, Queensland,

More information

Journal of Chemical and Pharmaceutical Research, 2014, 6(6): Research Article. A selective ensemble classification method on microarray data

Journal of Chemical and Pharmaceutical Research, 2014, 6(6): Research Article. A selective ensemble classification method on microarray data Avalable onlne www.ocpr.com Journal of Chemcal and Pharmaceutcal Research, 2014, 6(6):2860-2866 Research Artcle ISSN : 0975-7384 CODEN(USA) : JCPRC5 A selectve ensemble classfcaton method on mcroarray

More information

Learning-Based Top-N Selection Query Evaluation over Relational Databases

Learning-Based Top-N Selection Query Evaluation over Relational Databases Learnng-Based Top-N Selecton Query Evaluaton over Relatonal Databases Lang Zhu *, Wey Meng ** * School of Mathematcs and Computer Scence, Hebe Unversty, Baodng, Hebe 071002, Chna, zhu@mal.hbu.edu.cn **

More information

Lecture #15 Lecture Notes

Lecture #15 Lecture Notes Lecture #15 Lecture Notes The ocean water column s very much a 3-D spatal entt and we need to represent that structure n an economcal way to deal wth t n calculatons. We wll dscuss one way to do so, emprcal

More information

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,

More information

Complex System Reliability Evaluation using Support Vector Machine for Incomplete Data-set

Complex System Reliability Evaluation using Support Vector Machine for Incomplete Data-set Internatonal Journal of Performablty Engneerng, Vol. 7, No. 1, January 2010, pp.32-42. RAMS Consultants Prnted n Inda Complex System Relablty Evaluaton usng Support Vector Machne for Incomplete Data-set

More information

Fuzzy Model Identification Using Support Vector Clustering Method

Fuzzy Model Identification Using Support Vector Clustering Method Fuzzy Model Identfcaton Usng Support Vector Clusterng Method $\úhj OUçar, Yakup Demr, and Cüneyt * ]HOLú Electrcal and Electroncs Engneerng Department, Engneerng Faculty, ÕUDW Unversty, Elazg, Turkey agulucar@eee.org,ydemr@frat.edu.tr

More information

Incremental Learning with Support Vector Machines and Fuzzy Set Theory

Incremental Learning with Support Vector Machines and Fuzzy Set Theory The 25th Workshop on Combnatoral Mathematcs and Computaton Theory Incremental Learnng wth Support Vector Machnes and Fuzzy Set Theory Yu-Mng Chuang 1 and Cha-Hwa Ln 2* 1 Department of Computer Scence and

More information

Proper Choice of Data Used for the Estimation of Datum Transformation Parameters

Proper Choice of Data Used for the Estimation of Datum Transformation Parameters Proper Choce of Data Used for the Estmaton of Datum Transformaton Parameters Hakan S. KUTOGLU, Turkey Key words: Coordnate systems; transformaton; estmaton, relablty. SUMMARY Advances n technologes and

More information

GSLM Operations Research II Fall 13/14

GSLM Operations Research II Fall 13/14 GSLM 58 Operatons Research II Fall /4 6. Separable Programmng Consder a general NLP mn f(x) s.t. g j (x) b j j =. m. Defnton 6.. The NLP s a separable program f ts objectve functon and all constrants are

More information

The Shortest Path of Touring Lines given in the Plane

The Shortest Path of Touring Lines given in the Plane Send Orders for Reprnts to reprnts@benthamscence.ae 262 The Open Cybernetcs & Systemcs Journal, 2015, 9, 262-267 The Shortest Path of Tourng Lnes gven n the Plane Open Access Ljuan Wang 1,2, Dandan He

More information

Meta-heuristics for Multidimensional Knapsack Problems

Meta-heuristics for Multidimensional Knapsack Problems 2012 4th Internatonal Conference on Computer Research and Development IPCSIT vol.39 (2012) (2012) IACSIT Press, Sngapore Meta-heurstcs for Multdmensonal Knapsack Problems Zhbao Man + Computer Scence Department,

More information

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems A Unfed Framework for Semantcs and Feature Based Relevance Feedback n Image Retreval Systems Ye Lu *, Chunhu Hu 2, Xngquan Zhu 3*, HongJang Zhang 2, Qang Yang * School of Computng Scence Smon Fraser Unversty

More information

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1)

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1) Secton 1.2 Subsets and the Boolean operatons on sets If every element of the set A s an element of the set B, we say that A s a subset of B, or that A s contaned n B, or that B contans A, and we wrte A

More information

A Robust Method for Estimating the Fundamental Matrix

A Robust Method for Estimating the Fundamental Matrix Proc. VIIth Dgtal Image Computng: Technques and Applcatons, Sun C., Talbot H., Ourseln S. and Adraansen T. (Eds.), 0- Dec. 003, Sydney A Robust Method for Estmatng the Fundamental Matrx C.L. Feng and Y.S.

More information

Computer Animation and Visualisation. Lecture 4. Rigging / Skinning

Computer Animation and Visualisation. Lecture 4. Rigging / Skinning Computer Anmaton and Vsualsaton Lecture 4. Rggng / Sknnng Taku Komura Overvew Sknnng / Rggng Background knowledge Lnear Blendng How to decde weghts? Example-based Method Anatomcal models Sknnng Assume

More information

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

THE FUZZY GROUP METHOD OF DATA HANDLING WITH FUZZY INPUTS. Yuriy Zaychenko

THE FUZZY GROUP METHOD OF DATA HANDLING WITH FUZZY INPUTS. Yuriy Zaychenko 206 5 Knowledge Dalogue - Soluton THE FUZZY GROUP ETHOD OF DATA HANDLING WITH FUZZY INPUTS Yury Zaycheno Abstract: The problem of forecastng models constructng usng expermental data n terms of fuzzness,

More information

APPLICATION OF PREDICTION-BASED PARTICLE FILTERS FOR TELEOPERATIONS OVER THE INTERNET

APPLICATION OF PREDICTION-BASED PARTICLE FILTERS FOR TELEOPERATIONS OVER THE INTERNET APPLICATION OF PREDICTION-BASED PARTICLE FILTERS FOR TELEOPERATIONS OVER THE INTERNET Jae-young Lee, Shahram Payandeh, and Ljljana Trajovć School of Engneerng Scence Smon Fraser Unversty 8888 Unversty

More information

Network Intrusion Detection Based on PSO-SVM

Network Intrusion Detection Based on PSO-SVM TELKOMNIKA Indonesan Journal of Electrcal Engneerng Vol.1, No., February 014, pp. 150 ~ 1508 DOI: http://dx.do.org/10.11591/telkomnka.v1.386 150 Network Intruson Detecton Based on PSO-SVM Changsheng Xang*

More information

Random Kernel Perceptron on ATTiny2313 Microcontroller

Random Kernel Perceptron on ATTiny2313 Microcontroller Random Kernel Perceptron on ATTny233 Mcrocontroller Nemanja Djurc Department of Computer and Informaton Scences, Temple Unversty Phladelpha, PA 922, USA nemanja.djurc@temple.edu Slobodan Vucetc Department

More information

Module Management Tool in Software Development Organizations

Module Management Tool in Software Development Organizations Journal of Computer Scence (5): 8-, 7 ISSN 59-66 7 Scence Publcatons Management Tool n Software Development Organzatons Ahmad A. Al-Rababah and Mohammad A. Al-Rababah Faculty of IT, Al-Ahlyyah Amman Unversty,

More information

Dynamic Integration of Regression Models

Dynamic Integration of Regression Models Dynamc Integraton of Regresson Models Nall Rooney 1, Davd Patterson 1, Sarab Anand 1, Alexey Tsymbal 2 1 NIKEL, Faculty of Engneerng,16J27 Unversty Of Ulster at Jordanstown Newtonabbey, BT37 OQB, Unted

More information

Relevance Assignment and Fusion of Multiple Learning Methods Applied to Remote Sensing Image Analysis

Relevance Assignment and Fusion of Multiple Learning Methods Applied to Remote Sensing Image Analysis Assgnment and Fuson of Multple Learnng Methods Appled to Remote Sensng Image Analyss Peter Bajcsy, We-Wen Feng and Praveen Kumar Natonal Center for Supercomputng Applcaton (NCSA), Unversty of Illnos at

More information

From Comparing Clusterings to Combining Clusterings

From Comparing Clusterings to Combining Clusterings Proceedngs of the Twenty-Thrd AAAI Conference on Artfcal Intellgence (008 From Comparng Clusterngs to Combnng Clusterngs Zhwu Lu and Yuxn Peng and Janguo Xao Insttute of Computer Scence and Technology,

More information

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices Internatonal Mathematcal Forum, Vol 7, 2012, no 52, 2549-2554 An Applcaton of the Dulmage-Mendelsohn Decomposton to Sparse Null Space Bases of Full Row Rank Matrces Mostafa Khorramzadeh Department of Mathematcal

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

Some Advanced SPC Tools 1. Cumulative Sum Control (Cusum) Chart For the data shown in Table 9-1, the x chart can be generated.

Some Advanced SPC Tools 1. Cumulative Sum Control (Cusum) Chart For the data shown in Table 9-1, the x chart can be generated. Some Advanced SP Tools 1. umulatve Sum ontrol (usum) hart For the data shown n Table 9-1, the x chart can be generated. However, the shft taken place at sample #21 s not apparent. 92 For ths set samples,

More information

A CLASS OF TRANSFORMED EFFICIENT RATIO ESTIMATORS OF FINITE POPULATION MEAN. Department of Statistics, Islamia College, Peshawar, Pakistan 2

A CLASS OF TRANSFORMED EFFICIENT RATIO ESTIMATORS OF FINITE POPULATION MEAN. Department of Statistics, Islamia College, Peshawar, Pakistan 2 Pa. J. Statst. 5 Vol. 3(4), 353-36 A CLASS OF TRANSFORMED EFFICIENT RATIO ESTIMATORS OF FINITE POPULATION MEAN Sajjad Ahmad Khan, Hameed Al, Sadaf Manzoor and Alamgr Department of Statstcs, Islama College,

More information

Related-Mode Attacks on CTR Encryption Mode

Related-Mode Attacks on CTR Encryption Mode Internatonal Journal of Network Securty, Vol.4, No.3, PP.282 287, May 2007 282 Related-Mode Attacks on CTR Encrypton Mode Dayn Wang, Dongda Ln, and Wenlng Wu (Correspondng author: Dayn Wang) Key Laboratory

More information

Mixed Linear System Estimation and Identification

Mixed Linear System Estimation and Identification 48th IEEE Conference on Decson and Control, Shangha, Chna, December 2009 Mxed Lnear System Estmaton and Identfcaton A. Zymns S. Boyd D. Gornevsky Abstract We consder a mxed lnear system model, wth both

More information

Fuzzy Filtering Algorithms for Image Processing: Performance Evaluation of Various Approaches

Fuzzy Filtering Algorithms for Image Processing: Performance Evaluation of Various Approaches Proceedngs of the Internatonal Conference on Cognton and Recognton Fuzzy Flterng Algorthms for Image Processng: Performance Evaluaton of Varous Approaches Rajoo Pandey and Umesh Ghanekar Department of

More information

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009.

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009. Farrukh Jabeen Algorthms 51 Assgnment #2 Due Date: June 15, 29. Assgnment # 2 Chapter 3 Dscrete Fourer Transforms Implement the FFT for the DFT. Descrbed n sectons 3.1 and 3.2. Delverables: 1. Concse descrpton

More information

Learning an Image Manifold for Retrieval

Learning an Image Manifold for Retrieval Learnng an Image Manfold for Retreval Xaofe He*, We-Yng Ma, and Hong-Jang Zhang Mcrosoft Research Asa Bejng, Chna, 100080 {wyma,hjzhang}@mcrosoft.com *Department of Computer Scence, The Unversty of Chcago

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

Clustering-based feature selection for black-box weather temperature prediction

Clustering-based feature selection for black-box weather temperature prediction Clusterng-based feature selecton for black-box weather temperature predcton Zahra Karevan KU Leuven, ESAT-STADIUS Kasteelpark Arenberg 10 B-3001 Leuven, Belgum Emal : zahra.karevan@esat.kuleuven.be Johan

More information

HU Sheng-neng* Resources and Electric Power,Zhengzhou ,China

HU Sheng-neng* Resources and Electric Power,Zhengzhou ,China do:10.21311/002.31.6.09 Applcaton of new neural network technology n traffc volume predcton Abstract HU Sheng-neng* 1 School of Cvl Engneerng &Communcaton, North Chna Unversty of Water Resources and Electrc

More information

SURFACE PROFILE EVALUATION BY FRACTAL DIMENSION AND STATISTIC TOOLS USING MATLAB

SURFACE PROFILE EVALUATION BY FRACTAL DIMENSION AND STATISTIC TOOLS USING MATLAB SURFACE PROFILE EVALUATION BY FRACTAL DIMENSION AND STATISTIC TOOLS USING MATLAB V. Hotař, A. Hotař Techncal Unversty of Lberec, Department of Glass Producng Machnes and Robotcs, Department of Materal

More information

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng

More information

Evolutionary Support Vector Regression based on Multi-Scale Radial Basis Function Kernel

Evolutionary Support Vector Regression based on Multi-Scale Radial Basis Function Kernel Eolutonary Support Vector Regresson based on Mult-Scale Radal Bass Functon Kernel Tanasanee Phenthrakul and Boonserm Kjsrkul Abstract Kernel functons are used n support ector regresson (SVR) to compute

More information

Exercises (Part 4) Introduction to R UCLA/CCPR. John Fox, February 2005

Exercises (Part 4) Introduction to R UCLA/CCPR. John Fox, February 2005 Exercses (Part 4) Introducton to R UCLA/CCPR John Fox, February 2005 1. A challengng problem: Iterated weghted least squares (IWLS) s a standard method of fttng generalzed lnear models to data. As descrbed

More information

A classification scheme for applications with ambiguous data

A classification scheme for applications with ambiguous data A classfcaton scheme for applcatons wth ambguous data Thomas P. Trappenberg Centre for Cogntve Neuroscence Department of Psychology Unversty of Oxford Oxford OX1 3UD, England Thomas.Trappenberg@psy.ox.ac.uk

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS46: Mnng Massve Datasets Jure Leskovec, Stanford Unversty http://cs46.stanford.edu /19/013 Jure Leskovec, Stanford CS46: Mnng Massve Datasets, http://cs46.stanford.edu Perceptron: y = sgn( x Ho to fnd

More information

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms Course Introducton Course Topcs Exams, abs, Proects A quc loo at a few algorthms 1 Advanced Data Structures and Algorthms Descrpton: We are gong to dscuss algorthm complexty analyss, algorthm desgn technques

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information