Bayesian Classifier Combination
|
|
- Susan Vivien Wood
- 6 years ago
- Views:
Transcription
1 Bayesan Classfer Combnaton Zoubn Ghahraman and Hyun-Chul Km Gatsby Computatonal Neuroscence Unt Unversty College London London WC1N 3AR, UK September 8, 2003 Abstract Bayesan model averagng lnearly mxes the probablstc predctons of multple models, each weghted by ts posteror probablty. Ths s the coherent Bayesan way of combnng multple models only under very restrctve assumptons, whch we outlne. We explore a general framework for Bayesan model combnaton (whch dffers from model averagng) n the context of classfcaton. Ths framework explctly models the relatonshp between each model s output and the unknown true label. The framework does not requre that the models be probablstc (they can even be human assessors), that they share pror nformaton or receve the same tranng data, or that they be ndependent n ther errors. Fnally, the Bayesan combner does not need to beleve any of the models s n fact correct. We test several varants of ths classfer combnaton procedure startng from a classc statstcal model proposed by [1] and usng MCMC to add more complex but mportant features to the model. Comparsons on several datasets to smpler methods lke majorty votng show that the Bayesan methods not only perform well but result n nterpretable dagnostcs on the data ponts and the models. 1 Introducton There are many methods avalable for classfcaton. When faced wth a new problem, where one has lttle pror knowledge, t s temptng to try many dfferent classfers n the hope that combnng ther predctons would gve good performance. Ths had lead to the prolferaton of classfer combnaton, a.k.a. ensemble learnng, methods [3]. The Bayesan model averagng (BMA) framework appears to be deally suted to combnng the outputs of multple classfers. However, ths s msleadng. Before we dscuss Bayesan classfer combnaton (BCC), the topc of ths paper, let us revew BMA and outlne why t s not the rght framework for combnng classfers. 1 The work was done whle H-C.K. was a vstng student from POSTECH, South Korea. 1 We have focused on classfcaton, although many of the deas carry forth to other modellng problems; we return to ths n the dscusson. 1
2 Assume there are K dfferent classfers. Bayesan model averagng starts wth a pror over the classfers, p(k) for the kth classfer. Ths s meant to capture the (pror) belef n each classfer. Then we observe some data D, and we compute the margnal lkelhood or model evdence p(d k) for each k (whch can nvolve ntegratng out the parameters of the classfer). Usng Bayes rule we compute the posteror p(k D) = p(k)p(d k)/p(d) and we use these posterors to weght the classfers predctons: K K p(t x, D) = p(t, k x, D) = p(t x, k, D)p(k D) (1) where x denotes a new nput data pont and t the predcted class label assocated wth data pont. The key element of ths well-known procedure s that the predctve dstrbuton of each classfer s lnearly weghted by ts posteror probablty. Whle ths approach s appealng and well-motvated from a Bayesan framework, t suffers from three mportant lmtatons: 1) It s only vald f we beleve that the K classfers capture mutually exclusve and exhaustve possbltes about how the data was generated. In fact, we mght not beleve at all that any of the K classfers reflects the true data generaton. However, we may stll want to be able to combne them to form a predcton. 2) For many classfcaton methods avalable n the machne learnng communty, t s not possble to compute, or even defne, the margnal lkelhood (for example, C4.5, knn, etc.). Moreover, one should n prncple be able to nclude human experts nto any classfer combnaton framework. The human expert would not naturally defne a lkelhood functon from whch margnal lkelhoods can be computed. 3) Not all classfers may have observed the same data or started wth the same pror assumptons. The Bayesan framework descrbed above would have dffcultes dealng wth such cases, snce the posteror s computed by condtonng on the same data set. Here we propose an approach to Bayesan classfer combnaton whch does not assume that any of the classfers s the true one. Moreover, t does not requre that the classfers be probablstc; they can even be human experts. Fnally, the classfers can embody wdely dfferent pror assumptons about the data, and have observed dfferent data sets. There are well-known technques for classfer combnaton, so called ensemble methods([3, 9]). 2, such as baggng, boostng, and daggng. These methods try to make ndvdual classfers dfferent by tranng them wth dfferent tranng sets or weghtng data ponts dfferently. Ths s because t s mportant to make the ndvdual classfers as ndependent as possble for ensemble methods to work well. In ths work, we do not restrct how the ndvdual classfers are traned, but nstead assume they are gven and fxed. Another powerful and general method, called stacked generalsaton can be used to combne lower-level models [10]. Stackng methods for classfer combnaton use another classfer whch has as nputs both the orgnal nputs and the output of the ndvdual classfers. Stackng can be combned wth baggng and daggng [9]. In 2 Note that the term ensemble learnng has also been used n the Bayesan lterature n a dfferent context to refer to approxmate Bayesan model averagng usng varatonal methods. 2
3 ths work, we do not use the nput vectors and we explctly model the errors and correlatons between ndvdual classfers. Therefore, our work deals wth a dfferent problem from those whch are usually handled usng ensemble and stackng methods. It should be possble to extend our method to encompass a fully-bayesan generalsaton of stackng, but we leave ths for future work. The method we propose for Bayesan classfer combnaton n a machne learnng context s drectly derved from the method proposed n [5] for modellng dsagreement between human assessors, whch n turn s an extenson of [2]. Ths method assumes ndvdual classfers are ndependent, whch s often unrealstc and results n lmted performance. We therefore start wth these models and propose three extensons for modellng the correlatons between ndvdual classfers. The lterature of combnng probablty dstrbutons s qute extensve, and revews of other methods ncludng lnear, logarthmc and multvarate normal opnon pools, can be found n [4] and [6]. 2 Independent Models for Bayesan Classfer Combnaton 2.1 Probablstc Model for Classfer Combnaton We descrbe the method proposed n [2] wth the vew of applyng t to classfer combnaton. For the th data pont, we assume the true label t s generated by a multnomal dstrbuton wth parameters p: p(t = j p) = p j. Then, we assume that the output of classfer k s generated by a multnomal dstrbuton wth parameters π (k) c (k) p(c (k) t = j) = π (k) j,c (k) j :. For smplcty we assume that the classfers have dscrete outputs,.e. c (k) {1,..., J} where J s the number of classes. The extenson to ndvdual classfers whch output probablty dstrbutons s obvously mportant and wll be explored n the future. The matrx π (k) captures the confuson matrx for classfer k. If we assume that the classfer outputs are ndependent gven the true label t, we K get p(c, t p, π) = p t π(k) where c denotes the vector of class labels over t,c (k) all classfers. If we further assume that labels across data ponts are ndependent and dentcally dstrbuted, we obtan the lkelhood { } p(c, t p, π) = I =1 p t K π (k) t,c (k). (2) Usually, c (k) s known and the other varables and parameters are unknown. By consderng t as hdden varables, we can apply the EM algorthm to fnd ML estmates for p and π. Ths s the approach taken n [2] and we also provde further detals n a longer verson of ths paper [7]. It should be noted that not only does ths perform classfer combnaton, but t provdes estmates of nterpretable quanttes such as the confuson matrces. 3
4 2.2 Independent BCC Model A Bayesan treatment of the probablstc model n Secton 2.1 was recently proposed n [5] for combnng multple human raters. They also consdered multple ratngs (.e. c (k) 1... c(k) M ) for the same nput vector by the same raters. Snce artfcal classfers are not usually varable n how they respond to the same nput, we do not consder replcates n the ratngs. The Bayesan model needs prors on the parameters; we used herarchcal conjugate prors. A row of the confuson matrx π (k) j = [π (k) j,1, π(k) j,2,, π(k) j,j ], s modeled to have = [α (k) j,1, α(k) j,2,, α(k) j,j ]. The pror s modeled by an exponental dstrbuton wth parameters λ j,l. All a Drchlet dstrbuton wth hyperparameters α (k) j dstrbuton of α (k) j,l rows are assumed ndependent wthn and across classfers; even so t s easy to bas the pror to prefer dagonal confuson matrces. (Detaled expressons are provded n the longer verson of the paper [7].) The pror for the class proportons p s also set to be Drchlet, wth hyperparameters ν. Based on the above pror, we can get the posteror for all random varables gven the observed class labels. Snce we assumed ndependence among classfers (as n [5]), the posteror densty s p(p, π, t, α c),2,...,k c (k) { I =1 =1,2,...,I p t K Fgure 1: The drected graphcal model for IBCC, wth plates over classfers K and data ponts I. π t,c (k) } p(p ν)p(π α)p(α λ). (3) We call ths model the Independent Bayesan Classfer Combnaton (IBCC) model. The graphcal model for IBCC s shown n Fg 1. Inference for the unknown random varλ α (k) p ν ables p, π, t, and α can be done va Gbbs samplng. Snce the condtonal denstes on p and π (k) j are both Drchlet, they can be sampled easly; also, t π (k) t can be sampled snce t s a multnomal dstrbuton. However, the exact condtonals for α (k) are not easly obtaned, j,l so we use rejecton samplng. Hyperparameters ν are set so that class are roughly balanced a pror; λ s set to have bgger values on the dagonal than the off-dagonals. Ths encodes the pror that classfer outputs are better than random. 4
5 3 Dependent Models for Bayesan Classfer Combnaton One of the problems wth the above model s the assumpton that classfers are ndependent, whch s often not true n a real stuaton. Consder several poor classfers that make hghly correlated mstakes and one good classfer. Assumng ndependence results n performance based toward majorty votng, whereas accountng for the dependence would dscount the poor classfers by an amount related to ther correlaton. Modellng dependence therefore appears to be an essental element of Bayesan classfer combnaton. We propose three models to deal wth correlaton among classfer outputs. Frst, we nsert a new hdden varable representng the dffculty of each data pont margnalsng ths out results n a weakly dependent model. Second, we explctly model parwse dependence between classfers usng a Markov Network. Thrd, we combne the above two deas. 3.1 Enhanced BCC Model We enhance the IBCC model by usng dfferent confuson matrces accordng to dffculty of each data pont for classfcaton. Easy data ponts are classfed usng a confuson matrx E whch s fxed to have dagonal elements 1 ɛ and off-dagonal elements ɛ/(j 1) (we ve also tred extensons where E s learned). For hard data ponts, each classfer uses ts own confuson matrx, π (k), as before. Whether a data pont s easy or hard s controlled by ndependent Bernoull latent varables s (=1, f hard) wth mean d, whch s gven a Beta pror. The lkelhood term s as follows. { } I K K p(c, t p, π, s) = p t ( π (k) ) s ( E t,c (k) t,c ) (1 s) (4) (k) =1 We call ths model the Enhanced Bayesan Classfer Combnaton (EBCC) model. The graphcal model for the EBCC model n shown n Fg 2. Inference s agan performed usng Gbbs and rejecton samplng. λ β d s =1 β d α (k) π (k),2,...,k t (k) (k) c c s.t. s =1 ν p E,2,...,K t s =0 s.t. s =0 ν p Fgure 2: The graphcal model for the EBCC model. Note that we have a dfferent graphcal model condtonal on the settng of s for each pont; the left graph s for hard data and the rght graph s for easy data. (The usual DAG formalsm does not represent such dependence of structure on varable settng elegantly.) 5
6 3.2 Dependent BCC Model To model correlatons between classfers more drectly, we extend the IBCC model wth a Markov network. The part related to confuson matrces s replaced wth the followng Markov network. 1 p(c V, W, t ) = Z(V, W, t ) exp{ j<k W j,k δ(c (j), c (k) ) + k V (k) t,c (k) } (5) In ths Markov network, V relates t wth c (k), and W relates c (j) wth c (k), whch models correlatons between classfers; Z s a partton functon (normalser). The same prors p(t p)p(p ν) as n IBCC are used. As prors for elements of V and W, we use zero-mean ndependent Gaussans wth varance σ 2 v and σ 2 w. Samplng for most of the parameters of ths model s agan straghtforward. However, samplng from V, W s more subtle due to the partton functon, so we mplemented t usng a Metropols samplng method. We call ths model the Dependent Bayesan Classfer Combnaton (DBCC) model. Snce t s a mx of drected and undrected condtonal ndependence relatons t s most smply depcted as a factor graph (Fg 3). ν p (j) V (1) V V (K) t Fgure 3: The factor graph for the DBCC model. Each dot represents a factor n the jont probablty and connects varables nvolved n that factor. (1) c (j) c (K) c =1,2,...,I W 1,j W 1,K W j,k 3.3 Enhanced Dependent BCC model The Enhanced Dependence BCC model (EDBCC) combnes the easy/hard latent varable for the EBCC wth the explct model of correlaton between classfers of the DBCC. For easy data, the condtonal probablty of each class s gven by: p easy (c (:) U, t ) = 1 Z e (U, t ) exp{ k U t,c } (6) (k) U relates t wth c (k) (playng a role analogous to the E matrx n EBCC). For easy data ponts, t s assumed that classfers are ndependent, for hard data t s assumed to be as n DBCC. The factor graph for the EDBCC model s shown n (Fg 4). 6
7 ν ν β d s =1 p t (j) (1) V V (K) V c c c (1) (j) (K) s.t. s =1 β d s =0 p t U (1) c (j) c (K) c s.t. s =0 Fgure 4: The factor graph for the EDBCC model. Agan we have a dfferent graph condtonal on the settng of s. The left half shows the factor graph for hard data ponts (s = 1) and the rght half for easy data ponts. W 1,j W 1,K W j,k 4 Expermental Results We compared the Bayesan classfer combnaton methods on several data sets and usng dfferent component classfers. We used Satellte and DNA data sets from the Statlog project([8]) and the UCI dgt data set ([1]) 3. Our goal was not to obtan the best classfer performance for ths we would have pad very careful attenton to the component classfers and chosen sophstcated models suted to the propertes of each data set rather our goal was to compare the usefulness of dfferent BCC methods even when component classfers are poor, correlated or traned on partal data. We compared the four varants of the BCC dea outlned above to two other methods: selectng the best classfer usng valdaton data 4 and majorty votng. In all BCC models the valdaton data was used as known t to ground the estmates of model parameters. In theory ths groundng s not necessary: we can treat the labels n the observed data set as smply another classfer s outputs (perhaps the human who handlabelled the data) and assume that no true labels t are ever observed. Ths varant dd not seem to work as well n ntal experments but needs to be explored further. BCC results are based on comparng the posteror mode of t for data ponts n the test set to the true observed label. We dd two sets of experments. In Experment 1, we combned the outputs of the same type of classfer traned on dsjont tranng sets. 5 In Experment 2, we traned several dfferent classfers on the (same) whole tranng set. 6 For all BCC models ran 3 The DNA data set has a tranng set of 2000, a test set of 1186 wth 3 classes and 50 varables. Satellte has a tranng set of 4435, a test set of 2000 wth 6 classes and 36 varables. UCI dgt data set has a tranng set of 3823, a test set of 1797, 10 classes and 64 varables , 1000, 797 data ponts were selected from the orgnal test set as a valdaton set for DNA data set, Satellte data set, UCI dgt data set, respectvely. The rest of the orgnal test set was used to evaluate the performance. 5 For DNA data set, we had 5 dsjont tranng sets and traned C4.5 for each of them. For Satellte data set, we had 4 dsjont tranng sets and traned C4.5 for each of them. For UCI dgt data set, we had 3 dsjont tranng sets and traned SVM wth 2nd-order polynomal kernel and C = For DNA data set, we traned 5 classfers: C4.5 (C1), SVM wth 2nd-order polynomal kernel and C = (C2), 1-Nearest Neghbor (C3), logstc regresson (C4), and Fsher dscrmnant (C5). For Satellte data set, we traned 4 classfers: C4.5 (C1), SVM wth 2nd-order polynomal kernel and C = (C2), 7
8 Experment 1 Experment 2 Data set Satellte UCI dgt DNA Satellte UCI dgt DNA C C C C N/A N/A C5 N/A N/A N/A N/A Val MV IBCC EBCC DBCC EDBCC Table 1: The performances of ndvdual classfers and varous combnaton schemes n the case of usng the same classfer wth the dsjont tranng sets (Experment 1) and dfferent classfers wth the same whole tranng set (Experment 2) the MCMC sampler for at least samples, averagng every 100th and dscardng the frst The dependent models (DBCC and EDBCC) were generally slower to converge. Detals of the samplng and hyperparameter settngs are provded n the longer verson of the paper. Table 1 shows the performance of each classfer and BCC combnaton strategy for both experments. Val and MV denote selectng the classfer wth smallest valdaton set errors, and majorty votng, respectvely. IBCC and EBCC have smlar performance and EBCC model s always better than or as good as majorty votng. Model selecton by valdaton set s qute bad especally n Experment 1. BCC methods are always better than or as good as model selecton by valdaton. The dependent factor graph models (DBCC and EDBCC) do not always work well. Especally on the DNA data set, they dd not seem to learn reasonable parameters, perhaps because the DNA data set s relatvely small and has based class dstrbuton. For Satellte and UCI dgts, t learned resonable parameters and showed comparable performance to other BCC methods. We examned the V and W matrces nferred by the dependent methods and the dffculty assgned to each pont by the enhanced methods. These have ntutve nterpretatons and may provde useful dagnostcs, one of the strengths of the BCC approach. Due to space lmtatons we do not dsplay these matrces or dscuss them n ths paper; see [7]. logstc regresson (C3), and Fsher dscrmnant (C4). For UCI dgts, we traned 3 classfers: SVM wth lnear kernel (C1), SVM wth 2nd-order polynomal kernel (C2), and SVM wth Gaussan kernel (σ = 0.01) (C3), where all SVMs has C =
9 5 Dscusson We have shown several approaches to classfer combnaton whch explctly model the relaton between true labels and classfer outputs. They worked reasonably well and some of them were always better than or as good as majorty votng or valdaton selecton. The parameters n BCC models can be nterpreted resonably and gve useful nformaton such as confuson matrces, correlatons between classfers, and dffculty of data ponts. We emphassed that Bayesan classfer combnaton s not the same as Bayesan model averagng. Our approach s closely related to supra-bayesan methods for aggregatng opnons [4, 6]. Other models and extensons are certanly possble; we outlne some here. Clearly the model presented here needs to be generalsed to combne classfers that output probablty dstrbutons. In ths case, e.g. nstead of a matrx π (k) we need a model that relates t to class probablty dstrbutons. Condtonal Drchlet dstrbutons seem a natural choce for ths. Smlarly, there s no reason to restrct ths approach to combnng classfers. Combnng dfferent regressons s another mportant problem whch could be handled by an approprate choce of the densty of regressor outputs gven true target. A Bayesan generalsaton of stackng methods s another mportant avenue for research. The combner, n our setup, does not see the nput data. If the combner does see the nput and the outputs of all the other classfers, then t should model the full relaton between true labels, nputs, and other classfer outputs. One practcal lmtaton of the DBCC approach s that the computaton tme for the exact partton functon of the Markov network grows exponentally wth the number of classfers. Effcent approxmatons to the partton functon, many of whch have been recently developed, could be used here. Such approxmate nference could also be a tractable replacements for all the MCMC computatons. References [1] C. L. Blake and C. J. Merz. UCI Repostory of machne learnng databases. Irvne, CA: Unversty of Calforna, Department of Informaton and Computer Scence., [2] A. Dawd and A. Skene. Maxmum lkelhood estmaton of observer error-rates usng the em algorthm. Appled Statstcs, 28:20 28, [3] T. G. Detterch. Ensemble methods n machne learnng. Frst Internatonal Workshop on Multple Classfer Systems, LNCS, pages 1 15, [4] C. Genest and J. V. Zdek. Combnng probablty dstrbutons: A crtque and an annotated bblography. Statstcal Scence, 1: , [5] Y. Hatovsky, A. Smth, and Y. Lu. Modellng dsagreements among and wthn raters assessments from the bayesan pont of vew. In Draft. Presented at the Valenca meetng 2002, [6] R. Jacobs. Methods for combnng experts probablty assessments. Neural Computaton, 7: , [7] H. Km and Z. Ghahraman. Graphcal models for Bayesan classfer combnaton. GCNU Techncal Report (n preparaton),
10 [8] D. Mche, D. Spegelhalter, and C. Taylor. Machne Learnng, Neural and Statstcal Classfcaton. Ells Horwood Lmted, [9] K. Tng and I. H. Wtten. Stackng bagged and dagged models. In Proc. of ICML 97. San Francsco, CA, [10] D. H. Wolpert. Stacked generalzaton. Neural Networks, 5: ,
Machine Learning 9. week
Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below
More informationClassifier Selection Based on Data Complexity Measures *
Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.
More informationLearning the Kernel Parameters in Kernel Minimum Distance Classifier
Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department
More informationFeature Reduction and Selection
Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components
More informationOutline. Type of Machine Learning. Examples of Application. Unsupervised Learning
Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton
More informationA Statistical Model Selection Strategy Applied to Neural Networks
A Statstcal Model Selecton Strategy Appled to Neural Networks Joaquín Pzarro Elsa Guerrero Pedro L. Galndo joaqun.pzarro@uca.es elsa.guerrero@uca.es pedro.galndo@uca.es Dpto Lenguajes y Sstemas Informátcos
More informationSupport Vector Machines
/9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.
More informationUnsupervised Learning
Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and
More informationBAYESIAN MULTI-SOURCE DOMAIN ADAPTATION
BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,
More informationBOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET
1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School
More informationSubspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;
Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features
More informationX- Chart Using ANOM Approach
ISSN 1684-8403 Journal of Statstcs Volume 17, 010, pp. 3-3 Abstract X- Chart Usng ANOM Approach Gullapall Chakravarth 1 and Chaluvad Venkateswara Rao Control lmts for ndvdual measurements (X) chart are
More informationCS 534: Computer Vision Model Fitting
CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust
More informationLecture 5: Multilayer Perceptrons
Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented
More informationThe Research of Support Vector Machine in Agricultural Data Classification
The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou
More informationInvestigating the Performance of Naïve- Bayes Classifiers and K- Nearest Neighbor Classifiers
Journal of Convergence Informaton Technology Volume 5, Number 2, Aprl 2010 Investgatng the Performance of Naïve- Bayes Classfers and K- Nearest Neghbor Classfers Mohammed J. Islam *, Q. M. Jonathan Wu,
More informationMathematics 256 a course in differential equations for engineering students
Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the
More informationThree supervised learning methods on pen digits character recognition dataset
Three supervsed learnng methods on pen dgts character recognton dataset Chrs Flezach Department of Computer Scence and Engneerng Unversty of Calforna, San Dego San Dego, CA 92093 cflezac@cs.ucsd.edu Satoru
More informationA Post Randomization Framework for Privacy-Preserving Bayesian. Network Parameter Learning
A Post Randomzaton Framework for Prvacy-Preservng Bayesan Network Parameter Learnng JIANJIE MA K.SIVAKUMAR School Electrcal Engneerng and Computer Scence, Washngton State Unversty Pullman, WA. 9964-75
More informationLecture 4: Principal components
/3/6 Lecture 4: Prncpal components 3..6 Multvarate lnear regresson MLR s optmal for the estmaton data...but poor for handlng collnear data Covarance matrx s not nvertble (large condton number) Robustness
More informationContent Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers
IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth
More informationCluster Analysis of Electrical Behavior
Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School
More informationEdge Detection in Noisy Images Using the Support Vector Machines
Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona
More informationHermite Splines in Lie Groups as Products of Geodesics
Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the
More informationAnalysis of Continuous Beams in General
Analyss of Contnuous Beams n General Contnuous beams consdered here are prsmatc, rgdly connected to each beam segment and supported at varous ponts along the beam. onts are selected at ponts of support,
More informationTESTING AND IMPROVING LOCAL ADAPTIVE IMPORTANCE SAMPLING IN LJF LOCAL-JT IN MULTIPLY SECTIONED BAYESIAN NETWORKS
TESTING AND IMPROVING LOCAL ADAPTIVE IMPORTANCE SAMPLING IN LJF LOCAL-JT IN MULTIPLY SECTIONED BAYESIAN NETWORKS Dan Wu 1 and Sona Bhatt 2 1 School of Computer Scence Unversty of Wndsor, Wndsor, Ontaro
More informationSmoothing Spline ANOVA for variable screening
Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory
More informationAdaptive Transfer Learning
Adaptve Transfer Learnng Bn Cao, Snno Jaln Pan, Yu Zhang, Dt-Yan Yeung, Qang Yang Hong Kong Unversty of Scence and Technology Clear Water Bay, Kowloon, Hong Kong {caobn,snnopan,zhangyu,dyyeung,qyang}@cse.ust.hk
More information2x x l. Module 3: Element Properties Lecture 4: Lagrange and Serendipity Elements
Module 3: Element Propertes Lecture : Lagrange and Serendpty Elements 5 In last lecture note, the nterpolaton functons are derved on the bass of assumed polynomal from Pascal s trangle for the fled varable.
More informationModeling Waveform Shapes with Random Effects Segmental Hidden Markov Models
Modelng Waveform Shapes wth Random Effects Segmental Hdden Markov Models Seyoung Km, Padhrac Smyth Department of Computer Scence Unversty of Calforna, Irvne CA 9697-345 {sykm,smyth}@cs.uc.edu Abstract
More informationImplementation Naïve Bayes Algorithm for Student Classification Based on Graduation Status
Internatonal Journal of Appled Busness and Informaton Systems ISSN: 2597-8993 Vol 1, No 2, September 2017, pp. 6-12 6 Implementaton Naïve Bayes Algorthm for Student Classfcaton Based on Graduaton Status
More informationUser Authentication Based On Behavioral Mouse Dynamics Biometrics
User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA
More informationHuman Face Recognition Using Generalized. Kernel Fisher Discriminant
Human Face Recognton Usng Generalzed Kernel Fsher Dscrmnant ng-yu Sun,2 De-Shuang Huang Ln Guo. Insttute of Intellgent Machnes, Chnese Academy of Scences, P.O.ox 30, Hefe, Anhu, Chna. 2. Department of
More informationSupport Vector Machines
Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned
More informationNAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics
Introducton G10 NAG Fortran Lbrary Chapter Introducton G10 Smoothng n Statstcs Contents 1 Scope of the Chapter... 2 2 Background to the Problems... 2 2.1 Smoothng Methods... 2 2.2 Smoothng Splnes and Regresson
More informationTerm Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task
Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto
More informationDetermining the Optimal Bandwidth Based on Multi-criterion Fusion
Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn
More informationA User Selection Method in Advertising System
Int. J. Communcatons, etwork and System Scences, 2010, 3, 54-58 do:10.4236/jcns.2010.31007 Publshed Onlne January 2010 (http://www.scrp.org/journal/jcns/). A User Selecton Method n Advertsng System Shy
More informationA Multivariate Analysis of Static Code Attributes for Defect Prediction
Research Paper) A Multvarate Analyss of Statc Code Attrbutes for Defect Predcton Burak Turhan, Ayşe Bener Department of Computer Engneerng, Bogazc Unversty 3434, Bebek, Istanbul, Turkey {turhanb, bener}@boun.edu.tr
More informationA Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems
A Unfed Framework for Semantcs and Feature Based Relevance Feedback n Image Retreval Systems Ye Lu *, Chunhu Hu 2, Xngquan Zhu 3*, HongJang Zhang 2, Qang Yang * School of Computng Scence Smon Fraser Unversty
More informationHelsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr)
Helsnk Unversty Of Technology, Systems Analyss Laboratory Mat-2.08 Independent research projects n appled mathematcs (3 cr) "! #$&% Antt Laukkanen 506 R ajlaukka@cc.hut.f 2 Introducton...3 2 Multattrbute
More informationA Hidden Markov Model Variant for Sequence Classification
Proceedngs of the Twenty-Second Internatonal Jont Conference on Artfcal Intellgence A Hdden Markov Model Varant for Sequence Classfcaton Sam Blasak and Huzefa Rangwala Computer Scence, George Mason Unversty
More informationTN348: Openlab Module - Colocalization
TN348: Openlab Module - Colocalzaton Topc The Colocalzaton module provdes the faclty to vsualze and quantfy colocalzaton between pars of mages. The Colocalzaton wndow contans a prevew of the two mages
More informationClassification / Regression Support Vector Machines
Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM
More informationBiostatistics 615/815
The E-M Algorthm Bostatstcs 615/815 Lecture 17 Last Lecture: The Smplex Method General method for optmzaton Makes few assumptons about functon Crawls towards mnmum Some recommendatons Multple startng ponts
More informationSVM-based Learning for Multiple Model Estimation
SVM-based Learnng for Multple Model Estmaton Vladmr Cherkassky and Yunqan Ma Department of Electrcal and Computer Engneerng Unversty of Mnnesota Mnneapols, MN 55455 {cherkass,myq}@ece.umn.edu Abstract:
More informationData Mining: Model Evaluation
Data Mnng: Model Evaluaton Aprl 16, 2013 1 Issues: Evaluatng Classfcaton Methods Accurac classfer accurac: predctng class label predctor accurac: guessng value of predcted attrbutes Speed tme to construct
More informationAn Entropy-Based Approach to Integrated Information Needs Assessment
Dstrbuton Statement A: Approved for publc release; dstrbuton s unlmted. An Entropy-Based Approach to ntegrated nformaton Needs Assessment June 8, 2004 Wllam J. Farrell Lockheed Martn Advanced Technology
More informationSynthesizer 1.0. User s Guide. A Varying Coefficient Meta. nalytic Tool. Z. Krizan Employing Microsoft Excel 2007
Syntheszer 1.0 A Varyng Coeffcent Meta Meta-Analytc nalytc Tool Employng Mcrosoft Excel 007.38.17.5 User s Gude Z. Krzan 009 Table of Contents 1. Introducton and Acknowledgments 3. Operatonal Functons
More informationFusion Performance Model for Distributed Tracking and Classification
Fuson Performance Model for Dstrbuted rackng and Classfcaton K.C. Chang and Yng Song Dept. of SEOR, School of I&E George Mason Unversty FAIRFAX, VA kchang@gmu.edu Martn Lggns Verdan Systems Dvson, Inc.
More informationFast Sparse Gaussian Processes Learning for Man-Made Structure Classification
Fast Sparse Gaussan Processes Learnng for Man-Made Structure Classfcaton Hang Zhou Insttute for Vson Systems Engneerng, Dept Elec. & Comp. Syst. Eng. PO Box 35, Monash Unversty, Clayton, VIC 3800, Australa
More informationA Binarization Algorithm specialized on Document Images and Photos
A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a
More informationEmpirical Distributions of Parameter Estimates. in Binary Logistic Regression Using Bootstrap
Int. Journal of Math. Analyss, Vol. 8, 4, no. 5, 7-7 HIKARI Ltd, www.m-hkar.com http://dx.do.org/.988/jma.4.494 Emprcal Dstrbutons of Parameter Estmates n Bnary Logstc Regresson Usng Bootstrap Anwar Ftranto*
More informationA classification scheme for applications with ambiguous data
A classfcaton scheme for applcatons wth ambguous data Thomas P. Trappenberg Centre for Cogntve Neuroscence Department of Psychology Unversty of Oxford Oxford OX1 3UD, England Thomas.Trappenberg@psy.ox.ac.uk
More informationS1 Note. Basis functions.
S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type
More informationIncremental Learning with Support Vector Machines and Fuzzy Set Theory
The 25th Workshop on Combnatoral Mathematcs and Computaton Theory Incremental Learnng wth Support Vector Machnes and Fuzzy Set Theory Yu-Mng Chuang 1 and Cha-Hwa Ln 2* 1 Department of Computer Scence and
More informationFEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur
FEATURE EXTRACTION Dr. K.Vjayarekha Assocate Dean School of Electrcal and Electroncs Engneerng SASTRA Unversty, Thanjavur613 41 Jont Intatve of IITs and IISc Funded by MHRD Page 1 of 8 Table of Contents
More informationFeature-Based Matrix Factorization
Feature-Based Matrx Factorzaton arxv:1109.2271v3 [cs.ai] 29 Dec 2011 Tanq Chen, Zhao Zheng, Quxa Lu, Wenan Zhang, Yong Yu {tqchen,zhengzhao,luquxa,wnzhang,yyu}@apex.stu.edu.cn Apex Data & Knowledge Management
More informationAn Ensemble Learning algorithm for Blind Signal Separation Problem
An Ensemble Learnng algorthm for Blnd Sgnal Separaton Problem Yan L 1 and Peng Wen 1 Department of Mathematcs and Computng, Faculty of Engneerng and Surveyng The Unversty of Southern Queensland, Queensland,
More informationPerformance Evaluation of Information Retrieval Systems
Why System Evaluaton? Performance Evaluaton of Informaton Retreval Systems Many sldes n ths secton are adapted from Prof. Joydeep Ghosh (UT ECE) who n turn adapted them from Prof. Dk Lee (Unv. of Scence
More informationAnnouncements. Supervised Learning
Announcements See Chapter 5 of Duda, Hart, and Stork. Tutoral by Burge lnked to on web page. Supervsed Learnng Classfcaton wth labeled eamples. Images vectors n hgh-d space. Supervsed Learnng Labeled eamples
More informationAnalysis of Malaysian Wind Direction Data Using ORIANA
Modern Appled Scence March, 29 Analyss of Malaysan Wnd Drecton Data Usng ORIANA St Fatmah Hassan (Correspondng author) Centre for Foundaton Studes n Scence Unversty of Malaya, 63 Kuala Lumpur, Malaysa
More informationTsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance
Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for
More informationCSCI 5417 Information Retrieval Systems Jim Martin!
CSCI 5417 Informaton Retreval Systems Jm Martn! Lecture 11 9/29/2011 Today 9/29 Classfcaton Naïve Bayes classfcaton Ungram LM 1 Where we are... Bascs of ad hoc retreval Indexng Term weghtng/scorng Cosne
More informationHybridization of Expectation-Maximization and K-Means Algorithms for Better Clustering Performance
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 16, No 2 Sofa 2016 Prnt ISSN: 1311-9702; Onlne ISSN: 1314-4081 DOI: 10.1515/cat-2016-0017 Hybrdzaton of Expectaton-Maxmzaton
More informationImproved Methods for Lithography Model Calibration
Improved Methods for Lthography Model Calbraton Chrs Mack www.lthoguru.com, Austn, Texas Abstract Lthography models, ncludng rgorous frst prncple models and fast approxmate models used for OPC, requre
More information12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification
Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero
More informationWishing you all a Total Quality New Year!
Total Qualty Management and Sx Sgma Post Graduate Program 214-15 Sesson 4 Vnay Kumar Kalakband Assstant Professor Operatons & Systems Area 1 Wshng you all a Total Qualty New Year! Hope you acheve Sx sgma
More informationBayesian Approach for Fatigue Life Prediction from Field Inspection
Bayesan Approach for Fatgue Lfe Predcton from Feld Inspecton Dawn An, and Jooho Cho School of Aerospace & Mechancal Engneerng, Korea Aerospace Unversty skal@nate.com, jhcho@kau.ac.kr Nam H. Km, and Srram
More informationSimulation: Solving Dynamic Models ABE 5646 Week 11 Chapter 2, Spring 2010
Smulaton: Solvng Dynamc Models ABE 5646 Week Chapter 2, Sprng 200 Week Descrpton Readng Materal Mar 5- Mar 9 Evaluatng [Crop] Models Comparng a model wth data - Graphcal, errors - Measures of agreement
More informationProblem Definitions and Evaluation Criteria for Computational Expensive Optimization
Problem efntons and Evaluaton Crtera for Computatonal Expensve Optmzaton B. Lu 1, Q. Chen and Q. Zhang 3, J. J. Lang 4, P. N. Suganthan, B. Y. Qu 6 1 epartment of Computng, Glyndwr Unversty, UK Faclty
More informationParameter estimation for incomplete bivariate longitudinal data in clinical trials
Parameter estmaton for ncomplete bvarate longtudnal data n clncal trals Naum M. Khutoryansky Novo Nordsk Pharmaceutcals, Inc., Prnceton, NJ ABSTRACT Bvarate models are useful when analyzng longtudnal data
More informationKeywords - Wep page classification; bag of words model; topic model; hierarchical classification; Support Vector Machines
(IJCSIS) Internatonal Journal of Computer Scence and Informaton Securty, Herarchcal Web Page Classfcaton Based on a Topc Model and Neghborng Pages Integraton Wongkot Srura Phayung Meesad Choochart Haruechayasak
More informationSelecting Shape Features Using Multi-class Relevance Vector Machine
Selectng Shape Features Usng Mult-class Relevance Vector Machne Hao Zhang Jtendra Malk Electrcal Engneerng and Computer Scences Unversty of Calforna at Berkeley Techncal Report No. UCB/EECS-5-6 http://www.eecs.berkeley.edu/pubs/techrpts/5/eecs-5-6.html
More informationHierarchical Semantic Perceptron Grid based Neural Network CAO Huai-hu, YU Zhen-wei, WANG Yin-yan Abstract Key words 1.
Herarchcal Semantc Perceptron Grd based Neural CAO Hua-hu, YU Zhen-we, WANG Yn-yan (Dept. Computer of Chna Unversty of Mnng and Technology Bejng, Bejng 00083, chna) chhu@cumtb.edu.cn Abstract A herarchcal
More informationOptimizing Document Scoring for Query Retrieval
Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng
More informationMixed Linear System Estimation and Identification
48th IEEE Conference on Decson and Control, Shangha, Chna, December 2009 Mxed Lnear System Estmaton and Identfcaton A. Zymns S. Boyd D. Gornevsky Abstract We consder a mxed lnear system model, wth both
More informationNUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS
ARPN Journal of Engneerng and Appled Scences 006-017 Asan Research Publshng Network (ARPN). All rghts reserved. NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS Igor Grgoryev, Svetlana
More informationCHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION
48 CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 3.1 INTRODUCTION The raw mcroarray data s bascally an mage wth dfferent colors ndcatng hybrdzaton (Xue
More informationCourse Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms
Course Introducton Course Topcs Exams, abs, Proects A quc loo at a few algorthms 1 Advanced Data Structures and Algorthms Descrpton: We are gong to dscuss algorthm complexty analyss, algorthm desgn technques
More informationCompiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz
Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster
More informationAssignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009.
Farrukh Jabeen Algorthms 51 Assgnment #2 Due Date: June 15, 29. Assgnment # 2 Chapter 3 Dscrete Fourer Transforms Implement the FFT for the DFT. Descrbed n sectons 3.1 and 3.2. Delverables: 1. Concse descrpton
More informationIMAGE FUSION BASED ON EXTENSIONS OF INDEPENDENT COMPONENT ANALYSIS
IMAGE FUSION BASED ON EXTENSIONS OF INDEPENDENT COMPONENT ANALYSIS M Chen a, *, Yngchun Fu b, Deren L c, Qanqng Qn c a College of Educaton Technology, Captal Normal Unversty, Bejng 00037,Chna - (merc@hotmal.com)
More informationLearning-Based Top-N Selection Query Evaluation over Relational Databases
Learnng-Based Top-N Selecton Query Evaluaton over Relatonal Databases Lang Zhu *, Wey Meng ** * School of Mathematcs and Computer Scence, Hebe Unversty, Baodng, Hebe 071002, Chna, zhu@mal.hbu.edu.cn **
More informationImproving Web Image Search using Meta Re-rankers
VOLUME-1, ISSUE-V (Aug-Sep 2013) IS NOW AVAILABLE AT: www.dcst.com Improvng Web Image Search usng Meta Re-rankers B.Kavtha 1, N. Suata 2 1 Department of Computer Scence and Engneerng, Chtanya Bharath Insttute
More informationAn Indian Journal FULL PAPER ABSTRACT KEYWORDS. Trade Science Inc.
[Type text] [Type text] [Type text] ISSN : 97-735 Volume Issue 9 BoTechnology An Indan Journal FULL PAPER BTAIJ, (9), [333-3] Matlab mult-dmensonal model-based - 3 Chnese football assocaton super league
More informationContext-Specific Bayesian Clustering for Gene Expression Data
Context-Specfc Bayesan Clusterng for Gene Expresson Data Yoseph Barash School of Computer Scence & Engneerng Hebrew Unversty, Jerusalem, 91904, Israel hoan@cs.huj.ac.l Nr Fredman School of Computer Scence
More informationFuzzy Filtering Algorithms for Image Processing: Performance Evaluation of Various Approaches
Proceedngs of the Internatonal Conference on Cognton and Recognton Fuzzy Flterng Algorthms for Image Processng: Performance Evaluaton of Varous Approaches Rajoo Pandey and Umesh Ghanekar Department of
More informationComparing High-Order Boolean Features
Brgham Young Unversty BYU cholarsarchve All Faculty Publcatons 2005-07-0 Comparng Hgh-Order Boolean Features Adam Drake adam_drake@yahoo.com Dan A. Ventura ventura@cs.byu.edu Follow ths and addtonal works
More informationOutline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1
4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:
More informationA Similarity-Based Prognostics Approach for Remaining Useful Life Estimation of Engineered Systems
2008 INTERNATIONAL CONFERENCE ON PROGNOSTICS AND HEALTH MANAGEMENT A Smlarty-Based Prognostcs Approach for Remanng Useful Lfe Estmaton of Engneered Systems Tany Wang, Janbo Yu, Davd Segel, and Jay Lee
More informationUSING GRAPHING SKILLS
Name: BOLOGY: Date: _ Class: USNG GRAPHNG SKLLS NTRODUCTON: Recorded data can be plotted on a graph. A graph s a pctoral representaton of nformaton recorded n a data table. t s used to show a relatonshp
More informationSome Advanced SPC Tools 1. Cumulative Sum Control (Cusum) Chart For the data shown in Table 9-1, the x chart can be generated.
Some Advanced SP Tools 1. umulatve Sum ontrol (usum) hart For the data shown n Table 9-1, the x chart can be generated. However, the shft taken place at sample #21 s not apparent. 92 For ths set samples,
More informationMachine Learning. Topic 6: Clustering
Machne Learnng Topc 6: lusterng lusterng Groupng data nto (hopefully useful) sets. Thngs on the left Thngs on the rght Applcatons of lusterng Hypothess Generaton lusters mght suggest natural groups. Hypothess
More informationA Hill-climbing Landmarker Generation Algorithm Based on Efficiency and Correlativity Criteria
A Hll-clmbng Landmarker Generaton Algorthm Based on Effcency and Correlatvty Crtera Daren Ler, Irena Koprnska, and Sanjay Chawla School of Informaton Technologes, Unversty of Sydney Madsen Buldng F09,
More informationConsensus-Based Combining Method for Classifier Ensembles
76 The Internatonal Arab Journal of Informaton Technology, Vol. 15, No. 1, January 2018 Consensus-Based Combnng Method for Classfer Ensembles Omar Alzub 1, Jafar Alzub 2, Sara Tedmor 3, Hasan Rashadeh
More informationAdjustment methods for differential measurement errors in multimode surveys
Adjustment methods for dfferental measurement errors n multmode surveys Salah Merad UK Offce for Natonal Statstcs ESSnet MM DCSS, Fnal Meetng Wesbaden, Germany, 4-5 September 2014 Outlne Introducton Stablsng
More informationUSING LINEAR REGRESSION FOR THE AUTOMATION OF SUPERVISED CLASSIFICATION IN MULTITEMPORAL IMAGES
USING LINEAR REGRESSION FOR THE AUTOMATION OF SUPERVISED CLASSIFICATION IN MULTITEMPORAL IMAGES 1 Fetosa, R.Q., 2 Merelles, M.S.P., 3 Blos, P. A. 1,3 Dept. of Electrcal Engneerng ; Catholc Unversty of
More informationA Robust LS-SVM Regression
PROCEEDIGS OF WORLD ACADEMY OF SCIECE, EGIEERIG AD ECHOLOGY VOLUME 7 AUGUS 5 ISS 37- A Robust LS-SVM Regresson József Valyon, and Gábor Horváth Abstract In comparson to the orgnal SVM, whch nvolves a quadratc
More informationDiscriminative Dictionary Learning with Pairwise Constraints
Dscrmnatve Dctonary Learnng wth Parwse Constrants Humn Guo Zhuoln Jang LARRY S. DAVIS UNIVERSITY OF MARYLAND Nov. 6 th, Outlne Introducton/motvaton Dctonary Learnng Dscrmnatve Dctonary Learnng wth Parwse
More information