Evolutionary Support Vector Regression based on Multi-Scale Radial Basis Function Kernel
|
|
- Bridget Jenkins
- 5 years ago
- Views:
Transcription
1 Eolutonary Support Vector Regresson based on Mult-Scale Radal Bass Functon Kernel Tanasanee Phenthrakul and Boonserm Kjsrkul Abstract Kernel functons are used n support ector regresson (SVR) to compute the nner product n a hgher dmensonal feature space. The performance of approxmaton depends on the chosen kernels. The radal bass functon (RBF) kernel s a Mercer s kernel that has been wdely used n many problems. Howeer, t stll has the restrcton n some complex problems. In order to obtan a more flexble kernel functon, the mult-scale RBF kernels are combned by nonnegate weghtng lnear combnaton. Ths proposed kernel s proed to be a Mercer s kernel. Then, the eolutonary strategy (ES) s appled for adjustng the parameters of SVR and kernel functon. Moreoer, subsets cross-aldaton s used for ealuatng these parameters. The optmum alues of these parameters are searched by (5+0)-ES. The expermental results show the ablty of the proposed method that outperforms the statstcal technques. S I. INTRODUCTION UPPORT ector machnes (SVMs) are learnng algorthms proposed by Vapnk et al. [], based on the dea of the emprcal rsk mnmzaton prncple. They hae been wdely used n many applcatons such as pattern recogntons and functon approxmatons. Bascall SVM operates a lnear hyperplane n an augmented space by means of some defned kernels satsfyng Mercer s condton [], [], [3]. These kernels map the nput ectors nto a ery hgh dmensonal space, possbly of nfnte dmenson, where a lnear hyperplane s more lkely [3]. There are many types of kernel functons such as lnear kernels, polynomal kernels, sgmod kernels, and RBF kernels. Each kernel functon s sutable for some tasks, and t must be chosen for the tasks under consderaton by hand or usng pror knowledge [4]. The RBF kernel s a most successful kernel n many problems, but stll has the restrctons n some complex problems. Therefore, we propose to mproe the effcency of approxmaton by usng the combnaton of RBF kernels at dfferent scales. These kernels are combned by ncludng weghts. The weghts, the wdths of the RBF kernels, the deaton Ths work was supported by the Thaland Research Fund and the Royal Golden Jublee Ph.D. Program. T. Phenthrakul s wth the Department of Computer Engneerng, Chulalongkorn Unerst Bangkok 0330 Thaland (phone: ; fax: ; e-mal: tanasanee@yahoo.com). B. Kjsrkul, Dr., s wth the Department of Computer Engneerng, Faculty of Engneerng, Chulalongkorn Unerst Bangkok 0330 Thaland (e-mal: boonserm.k@chula.ac.th). of approxmaton, and the regularzaton parameter of SVM are the adjustable parameters; these parameters are called hyperparameters. In general, these hyperparameters are usually determned by a grd search. The hyperparameters are ared wth a fxed step-sze n a range of alues, but ths knd of search consumes a lot of tme. Hence, we propose to use the eolutonary strateges (ESs) for choosng these hyperparameters. Howeer, the objecte functon s an mportant part n eolutonary algorthms. There are many ways to measure the ftness of the hyperparameters. In ths work, we propose to use subsets cross-aldaton for ealuatng the hyperparameters n the eolutonary process. We ge a short descrpton of support ector regresson n Secton II. In Secton III and Secton IV we propose the mult-scale RBF kernel and apply eolutonary strateges to determne the approprate hyperparameters, respectely. The proposed kernels wth the help of ES are tested n Secton V. Fnall the conclusons are gen n Secton VI. II. SUPPORT VECTOR REGRESSION The support ector machne s a learnng algorthm that can be dded nto support ector classfcaton and support ector regresson. Support Vector Regresson (SVR) s a powerful method that s able to approxmate a real alued functon n terms of a small subset (called support ectors) of the tranng examples. Suppose we are gen tranng data {( y), K, ( xl, y l )} X R, where X denotes the space of nput patterns. In ε SV regresson, our goal s to fnd a functon f (x) that has at most ε deaton from the actually obtaned target y for all the tranng data, and at the same tme s as flat as possble. In other words, we do not care about errors as long as they are less than ε, but we do not accept any deaton larger than ths [5]. We begn by descrbng the case of lnear functons f, takng n form: f ( x) = w, x + b wth w X, b R. () Flatness n ths case means that one seeks a small w. One way to ensure ths s to mnmze the norm,.e. w = w, w. We can wrte ths problem as a conex optmzaton problem:
2 mnmze subject to w y w, x b ε () w, x b y ε + Sometmes, we may want to allow for some errors. Soft margn loss functon was adapted to SV machnes; one can ntroduce slack arables ξ, ξ to cope wth otherwse nfeasble constrants of the optmzaton problem [5]. Hence we wll get mnmze subject to w + C y w, x l = ( ξ + ξ ) b, x + b y ε + ξ w ε + ξ ξ, ξ 0 The constant C > 0 determnes the trade-off between the flatness of f and the amount up to whch deatons larger than ε are tolerated. In most cases ths optmzaton problem can be soled more easly n ts dual formulaton [5]. An example of a lnear SVR s shown n Fg.. (3) Kernel Lnear Polynomal TABLE I COMMON KERNEL FUNCTIONS Formula K( = x y K ( = ( + x Sgmod K ( = tanh( α x y + β ) Exponental RBF K( = exp( γ x y ) Gaussan RBF K( = exp( γ x y ) Mult-quadratc K ( = x y + c III. MULTI-SCALE RBF KERNEL The Gaussan RBF kernel s wdely used n many problems. It uses the Eucldean dstance between two ponts n the orgnal space to fnd the correlaton n the augmented space [3]. Ths correlaton s rather smooth, and there s only one parameter for adjustng the wdth of RBF, whch s not powerful enough for some complex problems. In order to get a better kernel, the combnaton of RBF kernels at dfferent scales s proposed. The analytc expresson of ths kernel s the followng: d n Κ( = a K( γ ), (4) = Fg.. An example of the soft margn for a lnear SVR. Moreoer, seekng a proper lnear hyperplane n an nput space has the restrctons. There s an mportant technque that enables these machnes to produce complex nonlnear approxmaton nsde the orgnal space. Ths performs by mappng the nput space nto a hgher dmensonal feature space through a mappng functon Φ [6]. Ths can be acheed by substtutng Φ ( x ) nto each tranng example x. Howeer, a good property of SVM s that t s not necessary to know the explct form of Φ. Only the nner product n the feature space, called kernel functon K( = Φ( x) Φ(, must be defned. The functon whch can be a kernel must satsfy Mercer s condton [7]. Some of the common kernels are shown n Table I. Each kernel corresponds to some feature space and because no explct mappng to ths feature space occurs, optmal lnear separators can be found effcently n the feature space wth mllons of dmensons [8]. where n s a poste nteger, a for =,..., n are the arbtrary non-negate weghtng constants, and K γ ) = ( exp( γ x y ) s the RBF kernel at the wdth γ for =,...,n. In general, the functon whch maps the nput space nto the augmented feature space s unknown. Howeer, the exstence of such functon s assured by Mercer s theorem [4]. The Mercer s theorem [], [4] s shown n Fg.. Fg.. The Mercer s theorem.
3 The proposed kernel functons can be proed to be an admssble kernel by the Mercer s theorem. The prong process s shown n Fg. 3. The RBF s a well-known Mercer s kernel. Therefore, the non-negate lnear combnaton of RBFs n (4) can be proed to be the Mercer s kernel. ( µ + λ )-ES where µ parents produce λ offsprng. Both parents and offsprng compete equally for sural []. The (5+0)-ES s appled to adjust these hyperparameters, and ths algorthm s shown n Fg. 4. t = 0; ntalzaton(,..., 5,σ ); ealuate f ( ),..., f ( 5 ) ; whle (t < 500) do for = to 0 do = recombnaton(,..., 5 ); = mutate ( ) ; ealuate f ( ) ; end (,..., 5 ) = select(,..., 5,,..., 0 ) σ = mutate σ (σ ) ; t = t+; end Fg. 4. (5+0)-ES algorthm. Fg. 3. The proof of the proposed kernel. As shown n (4), there are n parameters when n terms of RBF kernels are used ( n parameters for adjustng weghts and n alues of the wdths of RBFs ). Howeer, we notce that the number of parameters can be reduced to n by fxng a alue of the frst parameter to. The mult-scale RBF kernel becomes as follows n + Κ( = K( γ 0 ) ak( γ ). (5) = Ths form of the mult-scale RBF kernel wll be used n the rest of ths paper. IV. EVOLUTIONARY STRATEGIES FOR SVR BASED ON MULTI-SCALE RBF KERNEL Eolutonary strateges (ES, [9]) are based on the prncples of adapte selecton found n the natural world. ES has been successfully used to sole arous types of optmzaton problems [0]. Each generaton (teraton) of the ES algorthm takes a populaton of ndduals (potental solutons) and modfes the problem parameters to produce offsprng (new solutons) []. Only the hghest ft ndduals (better solutons) sure to produce new generatons []. In order to obtan approprate alues of the hyperparameters, ES s consdered. There are seeral dfferent ersons of ES. Neertheless, we prefer to use the Ths algorthm uses 5 solutons to produce 0 new solutons by a recombnaton method. These new solutons are mutated and ealuated, but only the 5 fttest solutons are selected from 5+0 solutons to be the parents n the next generaton. These processes wll be repeated untl a fxed number of generatons hae been produced or the acceptance crteron s reached. A. Intalzaton Let be the non-negate real-alue ector of the hyperparameters that has n + dmensons. The ector s represented n the form: = ( C, ε, n, γ 0, a, γ, a, γ,, a n, γ n ), (6) where C s the regularzaton parameter, ε s the deaton of an approxmaton, n s the number of terms of RBFs, γ for = 0,..., n are the wdths of RBFs, and a for =,..., n are the weghts of RBFs. The (5+0)-ES algorthm starts wth the 0 th generaton (t=0) that selects 5 solutons,..., 5 and standard deaton + σ R n + usng randomzaton. These 5 ntal solutons are ealuated to calculate ther ftness. Our goal s to fnd that optmzes the objecte functon f (). B. Recombnaton The recombnaton functon wll create 0 new solutons. We use the global ntermedary recombnaton method for creatng these 0 new solutons. Ten pars of solutons are selected from conentonal 5 solutons. The aerage of each par of solutons, element by element, s a new soluton.
4 M = ( ) = ( ) = ( ) (7) C. Mutaton The for =,,0 are mutated by addng each of them wth ( z, z,, z n+ ), and z s a random alue from a normal dstrbuton wth zero mean and σ araton. mutate ( ) = ( C + z, ε + z, n + z3, γ 0 + z4,, a n z, γ n z ) + n + z ~ (0, σ ). N + n + In each generaton, the standard deaton wll be adjusted by z z z n + mutate σ (σ ) = ( σ e, σ e,, σ n+ e ) (9) z ~ (0, τ ), N when τ s an arbtrary constant. D. Ealuaton In general, percentage error s used to measure the effcency of regresson model. Howeer, ths functon may oerft tranng data. Sometmes, data contan a lot of nose, and thus f the model fts these nosy data, the learned concept may be wrong. Hence, we would lke to aldate a set of hyperparameters on many tranng sets. A good hyperparameters should perform well on these tranng data. Howeer, as we hae only a fxed amount of tranng data. Therefore, 5-subsets cross-aldaton s proposed to aod the oerft problems. At the begnnng, the tranng data are dded nto fe subsets, each of whch has almost the same number of data. For each generaton of ES, the classfers wth the same hyperparameters are traned and aldated fe tmes. In the j th teraton ( j =,, 3, 4, 5), the classfer s traned on all subsets except for the j th one. Then, the error of predcton s calculated for the j th subset. The aerage of these fe errors s used to be the objecte functon f (). These parttons are dsplayed n Fg. 5. (8) Fg. 5. Partton tranng data nto 5 subsets. V. EXPERIMENTAL RESULTS In order to erfy the performance of the proposed method, SVRs wth the mult-scale RBF kernels are traned and tested on tme seres datasets from Neural Forecastng Competton []. These datasets are monthly tme seres drawn from homogeneous populaton of emprcal busness tme seres []. Snce these datasets are used for competton, we can ensure that these data are not so easy. In each dataset, the data of pror perods are used as the tranng data for predctng the next perod. The eolutonary strateges are used to fnd the optmal hyperparameters. The alue of τ n ealuaton process of these experments s.0. The number of RBF terms s a poste nteger that s less than or equal to 0. The wdths of RBFs ( γ ), the weghts of RBFs ( a ), the deaton of an approxmaton ( ε ), and the regularzaton parameter ( C ) are real numbers between 0.0 and 0.0. These hyperparameters are nspected wthn 500 generatons of ES. The performance of the proposed method s ealuated by the symmetrc mean absolute percentage error (SMAPE) [3], whch s defned as where y for data, ŷ for SMAPE = l l = y yˆ 00 ( y + yˆ ) (0) =,..., l are the actually targets of the tranng =,..., l are the forecast alues, and l s the number of tranng data. The expermental results are shown n Table II.
5 TABLE II THE SMAPE VALUES ON TRAINING DATA OF EACH DATASET (%) Datasets Cumulate Mean Mong Aerage N=5 Exponental Smoothng α =0.8 ES-SVR m-rbf NN3_ NN3_ NN3_ NN3_ NN3_ NN3_ NN3_ NN3_ NN3_ NN3_ NN3_ Aerage The expermental results show the ablty of the proposed method by SMAPE that outperforms the statstcal technques,.e. the cumulate mean, the mong aerage, and the exponental smoothng. Moreoer, the proposed method that combnes two technques of ES and SVR based on the mult-scale RBF kernel yelds the SMAPE alues less than SVR wth the sngle RBF kernel n all datasets. The aerage SMAPE on datasets of the proposed method s the best when compared wth the other technques. The examples of the approxmatons are llustrated as graphs n Fg. 6. From these graphs, the proposed method s compared wth the cumulate mean and the mong aerage. The proposed method and the mong aerage yeld the results that are smlar to the emprcal data. Howeer, wth the help of ES, the proposed method can determne the optmal hyperparameters n a more conenent way. For the cumulate mean, the results of the approxmaton are not good enough; t does not perform well on these data, where there are both seasonal and trend characterstcs. Fg. 6. Graph of the predcton by the proposed method, the cumulate mean, and the mong aerage on NN3_ dataset.
6 VI. CONCLUSION The non-negate lnear combnaton of multple RBF kernels wth ncludng weghts s proposed for support ector regresson. The proposed kernel s proed to be an admssble kernel by Mercer s theorem. Then, the eolutonary strategy s appled to the adjustment of the hyperparameters of SVR. The optmum alues of these hyperparameters are searched. Moreoer, subsets crossaldaton on the error of predcton s consdered to be the objecte functon n eolutonary process to aod the oerft problems. The expermental results show the ablty of the proposed method through the symmetrc mean absolute percentage error (SMAPE). When SVR uses the proposed kernel, t s able to learn from data ery well. Furthermore, the eolutonary strategy s effecte n optmzng the hyperparameters. Hence, the proposed method s hghly sutable for the complex problems where we hae no pror knowledge about ther hyperparameters. Besdes, ths non-negate lnear combnaton can be appled to other Mercer s kernels such as sgmod, polynomal, or Fourer seres kernels, as the general form of lnear combnaton of the Mercer s kernels has been proed to be a Mercer s kernel already. REFERENCES [] V.N. Vapnk, Statstcal Learnng Theor John Wley and Sons, Inc., New York, USA, 998. [] B. Schölkopf, C. Burges, and A.J. Smola, Adances n Kernel Methods: Support Vector Machnes. MIT Press, Cambrdge, MA, 998. [3] N.E. Ayat, M. Cheret, L. Remak, and C.Y. Suen, KMOD-A New Support Vector Machne Kernel wth Moderate Decreasng for Pattern Recognton, Proceedngs on Document Analyss and Recognton, pp. 5-9, Seattle, USA, 0-3 Sept. 00. [4] V. Kecman, Learnng and Soft Computng: Support Vector Machnes, Neural Networks, and Fuzzy Logc Models, MIT Press, London, 00. [5] A.J. Smola and B. Schölkopf, A Tutoral on Support Vector Regresson, NEUROCOLT Techncal Report NC-TR , Royal Holloway College, London, 998. [6] B. Schölkopf, and A.J. Smola, Learnng wth Kernels: Support Vector Machnes, Regularzaton, Optmzaton, and Beyond, MIT Press, London, 00. [7] J.S. Taylor and N. Crstann, Kernel Methods for Pattern Analyss, Cambrdge Unersty Press, UK, 004. [8] S. Russell and P. Norg, Artfcal Intellgence: A Modern Approach, Prentce-Hall, 003. [9] H.-G. Beyer and H.-P. Schwefel, Eoluton strateges: A comprehense ntroducton, Natural Computng, Vol., No., pp. 3-5, 00. [0] D.B. Fogel, Eolutonary Computaton: Toward a New Phlosophy of Machne Intellgence. IEEE Press, Pscatawa NJ, 995. [] E. dedoncker, A. Gupta, and G. Greenwood, Adapte Integraton Usng Eolutonary Strateges, Proceedngs of 3rd Internatonal Conference on Hgh Performance Computng, pp , 9- December 996. [] BI 3 S-Lab, Artfcal Neural Network & Computatonal Intellgence Forecastng Competton, Hamburg, German datasets.htm [Accessed: March 007]. [3] R.J. Hyndman and A.B. Koehler, Another Look at Measures of Forecast Accurac Internatonal Journal of Forecastng, Vol., Issue 4, Netherlands, 006.
Support Vector Machines
/9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.
More informationClassification / Regression Support Vector Machines
Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM
More informationSupport Vector Machines
Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned
More informationLearning the Kernel Parameters in Kernel Minimum Distance Classifier
Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department
More informationMachine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)
Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes
More informationEVOLVING HYPERPARAMETERS OF SUPPORT VECTOR MACHINES BASED ON MULTI-SCALE RBF KERNELS
EVOLVING HYPERPARAMETERS OF SUPPORT VECTOR MACHINES BASED ON MULTI-SCALE RBF KERNELS Tanasanee Phienthrakul and Boonserm Kijsirikul Department of Conputer Engineering, Chulalongkorn Uniersity, Thailand
More informationMachine Learning 9. week
Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below
More informationThe Research of Support Vector Machine in Agricultural Data Classification
The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou
More informationSmoothing Spline ANOVA for variable screening
Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory
More information12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification
Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero
More informationEdge Detection in Noisy Images Using the Support Vector Machines
Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona
More informationOutline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1
4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:
More informationAnnouncements. Supervised Learning
Announcements See Chapter 5 of Duda, Hart, and Stork. Tutoral by Burge lnked to on web page. Supervsed Learnng Classfcaton wth labeled eamples. Images vectors n hgh-d space. Supervsed Learnng Labeled eamples
More informationCS246: Mining Massive Datasets Jure Leskovec, Stanford University
CS46: Mnng Massve Datasets Jure Leskovec, Stanford Unversty http://cs46.stanford.edu /19/013 Jure Leskovec, Stanford CS46: Mnng Massve Datasets, http://cs46.stanford.edu Perceptron: y = sgn( x Ho to fnd
More informationBOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET
1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School
More informationFace Recognition Method Based on Within-class Clustering SVM
Face Recognton Method Based on Wthn-class Clusterng SVM Yan Wu, Xao Yao and Yng Xa Department of Computer Scence and Engneerng Tong Unversty Shangha, Chna Abstract - A face recognton method based on Wthn-class
More informationFeature Reduction and Selection
Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components
More informationGSLM Operations Research II Fall 13/14
GSLM 58 Operatons Research II Fall /4 6. Separable Programmng Consder a general NLP mn f(x) s.t. g j (x) b j j =. m. Defnton 6.. The NLP s a separable program f ts objectve functon and all constrants are
More informationSVM-based Learning for Multiple Model Estimation
SVM-based Learnng for Multple Model Estmaton Vladmr Cherkassky and Yunqan Ma Department of Electrcal and Computer Engneerng Unversty of Mnnesota Mnneapols, MN 55455 {cherkass,myq}@ece.umn.edu Abstract:
More informationSum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints
Australan Journal of Basc and Appled Scences, 2(4): 1204-1208, 2008 ISSN 1991-8178 Sum of Lnear and Fractonal Multobjectve Programmng Problem under Fuzzy Rules Constrants 1 2 Sanjay Jan and Kalash Lachhwan
More informationLecture 5: Multilayer Perceptrons
Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented
More informationIncremental Learning with Support Vector Machines and Fuzzy Set Theory
The 25th Workshop on Combnatoral Mathematcs and Computaton Theory Incremental Learnng wth Support Vector Machnes and Fuzzy Set Theory Yu-Mng Chuang 1 and Cha-Hwa Ln 2* 1 Department of Computer Scence and
More informationThe Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique
//00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy
More informationUsing Neural Networks and Support Vector Machines in Data Mining
Usng eural etworks and Support Vector Machnes n Data Mnng RICHARD A. WASIOWSKI Computer Scence Department Calforna State Unversty Domnguez Hlls Carson, CA 90747 USA Abstract: - Multvarate data analyss
More informationSupport Vector Machines. CS534 - Machine Learning
Support Vector Machnes CS534 - Machne Learnng Perceptron Revsted: Lnear Separators Bnar classfcaton can be veed as the task of separatng classes n feature space: b > 0 b 0 b < 0 f() sgn( b) Lnear Separators
More informationMeta-heuristics for Multidimensional Knapsack Problems
2012 4th Internatonal Conference on Computer Research and Development IPCSIT vol.39 (2012) (2012) IACSIT Press, Sngapore Meta-heurstcs for Multdmensonal Knapsack Problems Zhbao Man + Computer Scence Department,
More informationOutline. Type of Machine Learning. Examples of Application. Unsupervised Learning
Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton
More informationAssignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009.
Farrukh Jabeen Algorthms 51 Assgnment #2 Due Date: June 15, 29. Assgnment # 2 Chapter 3 Dscrete Fourer Transforms Implement the FFT for the DFT. Descrbed n sectons 3.1 and 3.2. Delverables: 1. Concse descrpton
More informationS1 Note. Basis functions.
S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type
More informationCHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION
48 CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 3.1 INTRODUCTION The raw mcroarray data s bascally an mage wth dfferent colors ndcatng hybrdzaton (Xue
More informationClassifier Selection Based on Data Complexity Measures *
Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.
More informationCluster Analysis of Electrical Behavior
Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School
More informationFuzzy Model Identification Using Support Vector Clustering Method
Fuzzy Model Identfcaton Usng Support Vector Clusterng Method $\úhj OUçar, Yakup Demr, and Cüneyt * ]HOLú Electrcal and Electroncs Engneerng Department, Engneerng Faculty, ÕUDW Unversty, Elazg, Turkey agulucar@eee.org,ydemr@frat.edu.tr
More informationParallelism for Nested Loops with Non-uniform and Flow Dependences
Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr
More informationFace Recognition University at Buffalo CSE666 Lecture Slides Resources:
Face Recognton Unversty at Buffalo CSE666 Lecture Sldes Resources: http://www.face-rec.org/algorthms/ Overvew of face recognton algorthms Correlaton - Pxel based correspondence between two face mages Structural
More informationSubspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;
Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features
More informationSHAPE RECOGNITION METHOD BASED ON THE k-nearest NEIGHBOR RULE
SHAPE RECOGNITION METHOD BASED ON THE k-nearest NEIGHBOR RULE Dorna Purcaru Faculty of Automaton, Computers and Electroncs Unersty of Craoa 13 Al. I. Cuza Street, Craoa RO-1100 ROMANIA E-mal: dpurcaru@electroncs.uc.ro
More informationDetermining the Optimal Bandwidth Based on Multi-criterion Fusion
Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn
More informationNUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS
ARPN Journal of Engneerng and Appled Scences 006-017 Asan Research Publshng Network (ARPN). All rghts reserved. NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS Igor Grgoryev, Svetlana
More informationAdaptive Virtual Support Vector Machine for the Reliability Analysis of High-Dimensional Problems
Proceedngs of the ASME 2 Internatonal Desgn Engneerng Techncal Conferences & Computers and Informaton n Engneerng Conference IDETC/CIE 2 August 29-3, 2, Washngton, D.C., USA DETC2-47538 Adaptve Vrtual
More informationA fast algorithm for color image segmentation
Unersty of Wollongong Research Onlne Faculty of Informatcs - Papers (Arche) Faculty of Engneerng and Informaton Scences 006 A fast algorthm for color mage segmentaton L. Dong Unersty of Wollongong, lju@uow.edu.au
More informationA Robust LS-SVM Regression
PROCEEDIGS OF WORLD ACADEMY OF SCIECE, EGIEERIG AD ECHOLOGY VOLUME 7 AUGUS 5 ISS 37- A Robust LS-SVM Regresson József Valyon, and Gábor Horváth Abstract In comparson to the orgnal SVM, whch nvolves a quadratc
More informationA Deflected Grid-based Algorithm for Clustering Analysis
A Deflected Grd-based Algorthm for Clusterng Analyss NANCY P. LIN, CHUNG-I CHANG, HAO-EN CHUEH, HUNG-JEN CHEN, WEI-HUA HAO Department of Computer Scence and Informaton Engneerng Tamkang Unversty 5 Yng-chuan
More informationBFF1303: ELECTRICAL / ELECTRONICS ENGINEERING. Direct Current Circuits : Methods of Analysis
BFF1303: ELECTRICAL / ELECTRONICS ENGINEERING Drect Current Crcuts : Methods of Analyss Ismal Mohd Kharuddn, Zulkfl Md Yusof Faculty of Manufacturng Engneerng Unerst Malaysa Pahang Drect Current Crcut
More informationDiscriminative classifiers for object classification. Last time
Dscrmnatve classfers for object classfcaton Thursday, Nov 12 Krsten Grauman UT Austn Last tme Supervsed classfcaton Loss and rsk, kbayes rule Skn color detecton example Sldng ndo detecton Classfers, boostng
More informationFace Recognition Based on SVM and 2DPCA
Vol. 4, o. 3, September, 2011 Face Recognton Based on SVM and 2DPCA Tha Hoang Le, Len Bu Faculty of Informaton Technology, HCMC Unversty of Scence Faculty of Informaton Scences and Engneerng, Unversty
More informationLECTURE : MANIFOLD LEARNING
LECTURE : MANIFOLD LEARNING Rta Osadchy Some sldes are due to L.Saul, V. C. Raykar, N. Verma Topcs PCA MDS IsoMap LLE EgenMaps Done! Dmensonalty Reducton Data representaton Inputs are real-valued vectors
More informationA Modified Median Filter for the Removal of Impulse Noise Based on the Support Vector Machines
A Modfed Medan Flter for the Removal of Impulse Nose Based on the Support Vector Machnes H. GOMEZ-MORENO, S. MALDONADO-BASCON, F. LOPEZ-FERRERAS, M. UTRILLA- MANSO AND P. GIL-JIMENEZ Departamento de Teoría
More informationCollaboratively Regularized Nearest Points for Set Based Recognition
Academc Center for Computng and Meda Studes, Kyoto Unversty Collaboratvely Regularzed Nearest Ponts for Set Based Recognton Yang Wu, Mchhko Mnoh, Masayuk Mukunok Kyoto Unversty 9/1/013 BMVC 013 @ Brstol,
More informationClassifying Acoustic Transient Signals Using Artificial Intelligence
Classfyng Acoustc Transent Sgnals Usng Artfcal Intellgence Steve Sutton, Unversty of North Carolna At Wlmngton (suttons@charter.net) Greg Huff, Unversty of North Carolna At Wlmngton (jgh7476@uncwl.edu)
More informationElliptical Rule Extraction from a Trained Radial Basis Function Neural Network
Ellptcal Rule Extracton from a Traned Radal Bass Functon Neural Network Andrey Bondarenko, Arkady Borsov DITF, Rga Techncal Unversty, Meza ¼, Rga, LV-1658, Latva Andrejs.bondarenko@gmal.com, arkadjs.borsovs@cs.rtu.lv
More informationUnsupervised Learning
Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and
More informationCLASSIFICATION OF ULTRASONIC SIGNALS
The 8 th Internatonal Conference of the Slovenan Socety for Non-Destructve Testng»Applcaton of Contemporary Non-Destructve Testng n Engneerng«September -3, 5, Portorož, Slovena, pp. 7-33 CLASSIFICATION
More informationComplex System Reliability Evaluation using Support Vector Machine for Incomplete Data-set
Internatonal Journal of Performablty Engneerng, Vol. 7, No. 1, January 2010, pp.32-42. RAMS Consultants Prnted n Inda Complex System Relablty Evaluaton usng Support Vector Machne for Incomplete Data-set
More informationHuman Face Recognition Using Generalized. Kernel Fisher Discriminant
Human Face Recognton Usng Generalzed Kernel Fsher Dscrmnant ng-yu Sun,2 De-Shuang Huang Ln Guo. Insttute of Intellgent Machnes, Chnese Academy of Scences, P.O.ox 30, Hefe, Anhu, Chna. 2. Department of
More informationOutline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like:
Self-Organzng Maps (SOM) Turgay İBRİKÇİ, PhD. Outlne Introducton Structures of SOM SOM Archtecture Neghborhoods SOM Algorthm Examples Summary 1 2 Unsupervsed Hebban Learnng US Hebban Learnng, Cntd 3 A
More informationBAYESIAN MULTI-SOURCE DOMAIN ADAPTATION
BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,
More informationSolving two-person zero-sum game by Matlab
Appled Mechancs and Materals Onlne: 2011-02-02 ISSN: 1662-7482, Vols. 50-51, pp 262-265 do:10.4028/www.scentfc.net/amm.50-51.262 2011 Trans Tech Publcatons, Swtzerland Solvng two-person zero-sum game by
More information6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour
6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the
More informationCircuit Analysis I (ENGR 2405) Chapter 3 Method of Analysis Nodal(KCL) and Mesh(KVL)
Crcut Analyss I (ENG 405) Chapter Method of Analyss Nodal(KCL) and Mesh(KVL) Nodal Analyss If nstead of focusng on the oltages of the crcut elements, one looks at the oltages at the nodes of the crcut,
More informationDiscrimination of Faulted Transmission Lines Using Multi Class Support Vector Machines
16th NAIONAL POWER SYSEMS CONFERENCE, 15th-17th DECEMBER, 2010 497 Dscrmnaton of Faulted ransmsson Lnes Usng Mult Class Support Vector Machnes D.hukaram, Senor Member IEEE, and Rmjhm Agrawal Abstract hs
More information5 The Primal-Dual Method
5 The Prmal-Dual Method Orgnally desgned as a method for solvng lnear programs, where t reduces weghted optmzaton problems to smpler combnatoral ones, the prmal-dual method (PDM) has receved much attenton
More informationSpam Filtering Based on Support Vector Machines with Taguchi Method for Parameter Selection
E-mal Spam Flterng Based on Support Vector Machnes wth Taguch Method for Parameter Selecton We-Chh Hsu, Tsan-Yng Yu E-mal Spam Flterng Based on Support Vector Machnes wth Taguch Method for Parameter Selecton
More informationThe Codesign Challenge
ECE 4530 Codesgn Challenge Fall 2007 Hardware/Software Codesgn The Codesgn Challenge Objectves In the codesgn challenge, your task s to accelerate a gven software reference mplementaton as fast as possble.
More informationParameters Optimization of SVM Based on Improved FOA and Its Application in Fault Diagnosis
Parameters Optmzaton of SVM Based on Improved FOA and Its Applcaton n Fault Dagnoss Qantu Zhang1*, Lqng Fang1, Sca Su, Yan Lv1 1 Frst Department, Mechancal Engneerng College, Shjazhuang, Hebe Provnce,
More informationRelevance Feedback Document Retrieval using Non-Relevant Documents
Relevance Feedback Document Retreval usng Non-Relevant Documents TAKASHI ONODA, HIROSHI MURATA and SEIJI YAMADA Ths paper reports a new document retreval method usng non-relevant documents. From a large
More informationCS 534: Computer Vision Model Fitting
CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust
More informationDiscriminative Dictionary Learning with Pairwise Constraints
Dscrmnatve Dctonary Learnng wth Parwse Constrants Humn Guo Zhuoln Jang LARRY S. DAVIS UNIVERSITY OF MARYLAND Nov. 6 th, Outlne Introducton/motvaton Dctonary Learnng Dscrmnatve Dctonary Learnng wth Parwse
More informationMachine Learning: Algorithms and Applications
14/05/1 Machne Learnng: Algorthms and Applcatons Florano Zn Free Unversty of Bozen-Bolzano Faculty of Computer Scence Academc Year 011-01 Lecture 10: 14 May 01 Unsupervsed Learnng cont Sldes courtesy of
More informationAn Evolvable Clustering Based Algorithm to Learn Distance Function for Supervised Environment
IJCSI Internatonal Journal of Computer Scence Issues, Vol. 7, Issue 5, September 2010 ISSN (Onlne): 1694-0814 www.ijcsi.org 374 An Evolvable Clusterng Based Algorthm to Learn Dstance Functon for Supervsed
More informationNetwork Intrusion Detection Based on PSO-SVM
TELKOMNIKA Indonesan Journal of Electrcal Engneerng Vol.1, No., February 014, pp. 150 ~ 1508 DOI: http://dx.do.org/10.11591/telkomnka.v1.386 150 Network Intruson Detecton Based on PSO-SVM Changsheng Xang*
More informationKernel Adaptive Filtering Subject to Equality Function Constraints
Kernel Adaptve Flterng Subject to Equalty Functon onstrants Badong hen Zhengda Qn Nannng Zheng José Príncpe Insttute of Artfcal Intellgence and Robotcs X'an Jaotong Unversty X'an 70049 hna Department of
More informationThe Study of Remote Sensing Image Classification Based on Support Vector Machine
Sensors & Transducers 03 by IFSA http://www.sensorsportal.com The Study of Remote Sensng Image Classfcaton Based on Support Vector Machne, ZHANG Jan-Hua Key Research Insttute of Yellow Rver Cvlzaton and
More informationGeneral Vector Machine. Hong Zhao Department of Physics, Xiamen University
General Vector Machne Hong Zhao (zhaoh@xmu.edu.cn) Department of Physcs, Xamen Unversty The support vector machne (SVM) s an mportant class of learnng machnes for functon approach, pattern recognton, and
More information2x x l. Module 3: Element Properties Lecture 4: Lagrange and Serendipity Elements
Module 3: Element Propertes Lecture : Lagrange and Serendpty Elements 5 In last lecture note, the nterpolaton functons are derved on the bass of assumed polynomal from Pascal s trangle for the fled varable.
More informationJapanese Dependency Analysis Based on Improved SVM and KNN
Proceedngs of the 7th WSEAS Internatonal Conference on Smulaton, Modellng and Optmzaton, Bejng, Chna, September 15-17, 2007 140 Japanese Dependency Analyss Based on Improved SVM and KNN ZHOU HUIWEI and
More informationBackpropagation: In Search of Performance Parameters
Bacpropagaton: In Search of Performance Parameters ANIL KUMAR ENUMULAPALLY, LINGGUO BU, and KHOSROW KAIKHAH, Ph.D. Computer Scence Department Texas State Unversty-San Marcos San Marcos, TX-78666 USA ae049@txstate.edu,
More informationAn AAM-based Face Shape Classification Method Used for Facial Expression Recognition
Internatonal Journal of Research n Engneerng and Technology (IJRET) Vol. 2, No. 4, 23 ISSN 2277 4378 An AAM-based Face Shape Classfcaton Method Used for Facal Expresson Recognton Lunng. L, Jaehyun So,
More informationResearch of Neural Network Classifier Based on FCM and PSO for Breast Cancer Classification
Research of Neural Network Classfer Based on FCM and PSO for Breast Cancer Classfcaton Le Zhang 1, Ln Wang 1, Xujewen Wang 2, Keke Lu 2, and Ajth Abraham 3 1 Shandong Provncal Key Laboratory of Network
More informationArtificial Intelligence (AI) methods are concerned with. Artificial Intelligence Techniques for Steam Generator Modelling
Artfcal Intellgence Technques for Steam Generator Modellng Sarah Wrght and Tshldz Marwala Abstract Ths paper nvestgates the use of dfferent Artfcal Intellgence methods to predct the values of several contnuous
More informationHermite Splines in Lie Groups as Products of Geodesics
Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the
More informationModel Selection with Cross-Validations and Bootstraps Application to Time Series Prediction with RBFN Models
Model Selecton wth Cross-Valdatons and Bootstraps Applcaton to Tme Seres Predcton wth RBF Models Amaury Lendasse Vncent Wertz and Mchel Verleysen Unversté catholque de Louvan CESAME av. G. Lemaître 3 B-348
More informationNAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics
Introducton G10 NAG Fortran Lbrary Chapter Introducton G10 Smoothng n Statstcs Contents 1 Scope of the Chapter... 2 2 Background to the Problems... 2 2.1 Smoothng Methods... 2 2.2 Smoothng Splnes and Regresson
More informationGA-Based Learning Algorithms to Identify Fuzzy Rules for Fuzzy Neural Networks
Seventh Internatonal Conference on Intellgent Systems Desgn and Applcatons GA-Based Learnng Algorthms to Identfy Fuzzy Rules for Fuzzy Neural Networks K Almejall, K Dahal, Member IEEE, and A Hossan, Member
More informationTaxonomy of Large Margin Principle Algorithms for Ordinal Regression Problems
Taxonomy of Large Margn Prncple Algorthms for Ordnal Regresson Problems Amnon Shashua Computer Scence Department Stanford Unversty Stanford, CA 94305 emal: shashua@cs.stanford.edu Anat Levn School of Computer
More informationRecommended Items Rating Prediction based on RBF Neural Network Optimized by PSO Algorithm
Recommended Items Ratng Predcton based on RBF Neural Network Optmzed by PSO Algorthm Chengfang Tan, Cayn Wang, Yuln L and Xx Q Abstract In order to mtgate the data sparsty and cold-start problems of recommendaton
More informationOptimizing Document Scoring for Query Retrieval
Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng
More informationKent State University CS 4/ Design and Analysis of Algorithms. Dept. of Math & Computer Science LECT-16. Dynamic Programming
CS 4/560 Desgn and Analyss of Algorthms Kent State Unversty Dept. of Math & Computer Scence LECT-6 Dynamc Programmng 2 Dynamc Programmng Dynamc Programmng, lke the dvde-and-conquer method, solves problems
More informationOptimizing SVR using Local Best PSO for Software Effort Estimation
Journal of Informaton Technology and Computer Scence Volume 1, Number 1, 2016, pp. 28 37 Journal Homepage: www.jtecs.ub.ac.d Optmzng SVR usng Local Best PSO for Software Effort Estmaton Dnda Novtasar 1,
More informationAn Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices
Internatonal Mathematcal Forum, Vol 7, 2012, no 52, 2549-2554 An Applcaton of the Dulmage-Mendelsohn Decomposton to Sparse Null Space Bases of Full Row Rank Matrces Mostafa Khorramzadeh Department of Mathematcal
More informationContent Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers
IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth
More informationMaximum Variance Combined with Adaptive Genetic Algorithm for Infrared Image Segmentation
Internatonal Conference on Logstcs Engneerng, Management and Computer Scence (LEMCS 5) Maxmum Varance Combned wth Adaptve Genetc Algorthm for Infrared Image Segmentaton Huxuan Fu College of Automaton Harbn
More informationLearning a Class-Specific Dictionary for Facial Expression Recognition
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 16, No 4 Sofa 016 Prnt ISSN: 1311-970; Onlne ISSN: 1314-4081 DOI: 10.1515/cat-016-0067 Learnng a Class-Specfc Dctonary for
More informationReview of approximation techniques
CHAPTER 2 Revew of appromaton technques 2. Introducton Optmzaton problems n engneerng desgn are characterzed by the followng assocated features: the objectve functon and constrants are mplct functons evaluated
More informationFace Detection with Deep Learning
Face Detecton wth Deep Learnng Yu Shen Yus122@ucsd.edu A13227146 Kuan-We Chen kuc010@ucsd.edu A99045121 Yzhou Hao y3hao@ucsd.edu A98017773 Mn Hsuan Wu mhwu@ucsd.edu A92424998 Abstract The project here
More informationA Genetic Programming-PCA Hybrid Face Recognition Algorithm
Journal of Sgnal and Informaton Processng, 20, 2, 70-74 do:0.4236/jsp.20.23022 Publshed Onlne August 20 (http://www.scrp.org/journal/jsp) A Genetc Programmng-PCA Hybrd Face Recognton Algorthm Behzad Bozorgtabar,
More informationA Naïve Support Vector Regression Benchmark for the NN3 Forecasting Competition
A Naïve Support Vector Regresson Benchmark for the NN3 Forecastng Competton Sven F. Crone and Swantje Petsch Abstract Support Vector Regresson s one of the promsng contenders n predctng the 111 tme seres
More informationAn Improved Particle Swarm Optimization for Feature Selection
Journal of Bonc Engneerng 8 (20)?????? An Improved Partcle Swarm Optmzaton for Feature Selecton Yuannng Lu,2, Gang Wang,2, Hulng Chen,2, Hao Dong,2, Xaodong Zhu,2, Sujng Wang,2 Abstract. College of Computer
More informationSequential search. Building Java Programs Chapter 13. Sequential search. Sequential search
Sequental search Buldng Java Programs Chapter 13 Searchng and Sortng sequental search: Locates a target value n an array/lst by examnng each element from start to fnsh. How many elements wll t need to
More informationLearning General Gaussian Kernels by Optimizing Kernel Polarization
Chnese Journal of Electroncs Vol.18, No.2, Apr. 2009 Learnng General Gaussan Kernels by Optmzng Kernel Polarzaton WANG Tnghua 1, HUANG Houkuan 1, TIAN Shengfeng 1 and DENG Dayong 2 (1.School of Computer
More information