A Multivariate Analysis of Static Code Attributes for Defect Prediction
|
|
- Emory O’Connor’
- 6 years ago
- Views:
Transcription
1 Research Paper) A Multvarate Analyss of Statc Code Attrbutes for Defect Predcton Burak Turhan, Ayşe Bener Department of Computer Engneerng, Bogazc Unversty 3434, Bebek, Istanbul, Turkey {turhanb, bener}@boun.edu.tr Abstract Defect predcton s mportant n order to reduce test tmes by allocatng valuable test resources effectvely. In ths work, we propose a model usng multvarate approaches n conjuncton wth Bayesan methods for defect predctons. The motvaton behnd usng a multvarate approach s to overcome the ndependence assumpton of unvarate approaches about software attrbutes. Usng Bayesan methods gves practtoners an dea about the defectveness of software modules n a probablstc framework rather than the hard classfcaton methods such as decson trees. Furthermore the software attrbutes used n ths work are chosen among the statc code attrbutes that can easly be extracted from source code, whch prevents human errors or subjectvty. These attrbutes are preprocessed wth feature selecton technques to select the most relevant attrbutes for predcton. Fnally we compared our proposed model wth the best results reported so far on publc datasets and we conclude that usng multvarate approaches can perform better. Keywords: Defect predcton, Software Metrcs, Naïve Bayes. Topcs: Software Qualty, Methods and Tools. 1. Introducton Testng s the most costly and tme consumng part of software development lfecycle, regardless of the development process used. Therefore effectve testng leads to sgnfcant decrease n project costs and schedules. The am of defect predcton s to gve an dea about the testng prortes, so that exhaustve testng s prevented. Usng an automated model may help project managers to allocate testng resources effectvely. These models can predct the degree of defectveness f relevant features of software are suppled to them. These relevant features are acheved by usng software metrcs. Researchers usually prefer focusng on the selecton of a subset of avalable features [10]. Feature subset selecton s manly preferred because of ts nterpretablty, snce the selected features correspond to actual and n some occasons controllable measurements from software. Ths gves the ablty to generate rules about the desred values of metrcs for 'good' software. It s easer to explan such rules to programmers and managers [6]. Ths s also the answer to why most of the studes use decson trees as predctors. Decson trees can be nterpreted as a set of rules and they can be understood by less techncally nvolved people [6]. But decson trees are hard classfcaton methods that can predct a module as ether defectve or non-defectve. Alternatvely, Bayesan approaches provde a probablstc framework and yeld soft classfcaton methods wth posteror probabltes attached to the predctons [1]. Ths s why we employed Bayesan approaches n ths work. On the other hand, feature subset selecton requres an exhaustve search for choosng the optmal subset. Thus, feature selecton algorthms use greedy approaches lke backward or forward selecton [7]. In forward selecton, one starts wth an empty set of features, and a feature s selected only f t ncreases the performance of the predctor, otherwse t s dscarded. Backward selecton s smlar n the sense that one starts wth all features, and a feature s removed f t does not affect the performance of the predctor. These approaches evaluate the features one at a tme and they do not consder the effects of features f taken as pars, trples and n-tuples. Whle a sngle feature may not affect the estmaton performance sgnfcantly, pars, trples or n-tuples of features may [7]. In order to
2 overcome ths problem, ths study employs feature extracton technques and compares the results wth a baselne study, where InfoGan algorthm s used to rank and select a subset of features [10]. Major contrbuton of ths research s to ncorporate multvarate approaches rather than unvarate ones. Unvarate approaches assume the ndependence of features whereas multvarate approaches take the relatons between features nto consderaton. Obvously unvarate models are smpler than multvarate models. Whle t s good practce to start modelng wth smple models, the problem at hand should also be nvestgated by usng more complex models. Then t should be valdated by measurng performance whether usng more complex models s worth the extra complexty ntroduced n the modelng. Ths research performs experments wth both smple and complex models and compares ther performances. In the followng secton, feature extracton methods used n ths research are brefly descrbed. Then, models used for defect predcton are explaned. After descrbng the expermental desgn and the results, conclusons wll be gven. global optmalty, and guarantee of asymptotc convergence are ts major features [16]. In general, Eucldean dstance s used to calculate the smlarty of two nstances. However, the use of the Eucldean dstance to represent par wse dstances makes the model unable to preserve the ntrnsc geometry of the data. Two nearby ponts, n terms of Eucldean dstance, may ndeed be dstant, because ther actual dstance s the path between these ponts along the manfold. The length of the path along the manfold s referred to as the geodesc dstance [16]. A -D spral s an example of a manfold, whch s actually a 1-D lne that s folded and embedded n -D See Fgure 1, adapted from [9]). Applyng Isomap on the spral unfolds t to ts true structure. Isomap smply performs classcal Multdmensonal Scalng [4] on par wse geodesc-dstance matrx.. Feature Extracton Methods In feature extracton, new features are formed by combnng the exstng ones. These new set of features may not be nterpreted easly as before [6]. On the contrary, there are cases where they turn out to be nterpretable [5]. The new features may also lead to better predcton performances by removng rrelevant and non-nformatve features. An advantage of feature extracton methods used n ths study s that they project data to an orthogonal feature space. One has to decde between ease of nterpretablty and better predcton performances n such cases. In ths research authors prefer better performance and therefore they explore feature extracton methodologes. Prncpal Component Analyss PCA) has been used n other defect predcton studes [11], [13], [8], [14],[]. We also use PCA n ths research. PCA reveals the optmum lnear structure of data ponts. But t s unable to fnd nonlnear relatons, f there exsts such relatons n data. In order to nvestgate non-lnear relatons, we use Isomap algorthm as another feature extracton technque..1. Isomap Isomap nherts the advantages of PCA and extends them to learn nonlnear structures that are hdden n hgh dmensonal data. Computatonal effcency, Fgure 1. Geodesc dstance metrc: Ponts X and Y are at dstnct ends of the spral. Usng Eucldean dstance, the true structure of spral,.e. 1-D lne folded and embedded n -D, can not be revealed. Geodesc dstance represents smlar or dfferent) data ponts more accurately than the Eucldean dstance, but the queston s how to estmate t? Here the local lnearty prncple s used and t s assumed that neghborng ponts le on a lnear patch of the manfold, so for nearby ponts the Eucldean dstances correctly estmate the geodesc dstances. For dstant ponts, the geodesc dstances are estmated by addng up neghborng dstances over the manfold usng a shortest-path algorthm. Isomap fnds the true dmensonalty of nonlnear structures. The nterpretaton of projecton axes can be meanngful n some cases [5]. Isomap uses a sngle parameter to defne the neghborhood for data ponts.e. for k-nearest neghbors of a data pont, par wse geodesc dstances are assumed to be equvalent to Eucldean dstances. Ths parameter should be fne tuned, preferably by cross-valdaton, to obtan optmum results. Data sample s transformed to have a
3 lnear structure n the new projecton space; e.g. the spral s unfolded to a lne. 3. Predctor Models Ths secton explans predctor models used for defect predcton. As a baselne, the Nave Bayes classfer s taken snce t s shown to acqure best results obtaned so far [10]. We remove the assumptons of the Nave Bayes classfer one at a tme and construct the lnear and quadratc dscrmnants. The assumpton n Nave Bayes s that the features of data sample are ndependent, thus t employs the unvarate normal dstrbuton. We beleve ths assumpton s not vald for software data and snce there are correlatons between software data features. So we use a multvarate normal dstrbuton to model the correlatons among features. In the next secton unvarate and multvarate normal dstrbutons are brefly explaned Unvarate vs. Multvarate Normal Dstrbuton In unvarate normal dstrbuton, x ~ N, ), x s sad to be normal dstrbuted wth mean μ and standard devaton σ and the probablty dstrbuton functon pdf) s defned as: 1 x p x) exp 1) ) The term nsde the exponental term n Equaton s the normalzed Eucldean dstance, where the dstance of a data sample x to the sample mean μ s measured n terms of standard devatons σ. Ths ensures to scale the dstances of dfferent features n case feature values vary sgnfcantly. Ths measure does not consder the correlatons among features. In the multvarate case, x s a d-dmensonal vector that s normal dstrbuted, x ~ N, ), and the pdf of a multvarate normal dstrbuton s defned as: 1 1 T 1 p x) exp x x ) d 1 ) Where Σ s the covarance matrx and μ s the mean vector. The term nsde the exponental term n Equaton s another dstance functon and called the Mahalanobs dstance [1]. In ths case, the dstance to the mean vector s normalzed by the covarance matrx and the correlatons of features are also consdered. Ths results n less contrbuton of hghly correlated features and features wth hgh varance. Our assumpton s that software data features are correlated and a multvarate model would be more approprate than the unvarate model. Besdes, multvarate normal dstrbuton s analytcally smple, tractable and robust to departures from normalty [1]. As no free lunch theorem states [17], nothng comes for free and usng a multvarate model ncreases the number of parameters to estmate. In the unvarate case, only parameters, μ and Σ are estmated, whle n the multvarate case, d parameters for μ and d.d parameters for Σ need to be estmated. 3.. Multvarate Classfcaton In software defect predcton, one ams to dscrmnate classes C 0 and C 1 where samples n C 0 are non defectve and samples n C 1 are defectve. We combne the multvarate normal dstrbuton and the Bayes rule, use dfferent assumptons, and acheve dfferent dscrmnants wth dfferent complexty levels See Table 1). We prefer dscrmnant pont of vew, snce t s geometrcally nterpretable. A dscrmnant n general s a hyper plane that separates d-dmensonal space nto dsjont subspaces. General structure of a dscrmnant s explaned next. Table 1. Complextes of predctors n a K-class problem wth d features. Predctor QD LD NB # Parameters K x d x d)) + K x d) + K) d x d) + K x d) + K) d) + K x d) + K) Bayes theorem states that the posteror dstrbuton of a sample s proportonal to the pror dstrbuton and the lkelhood of the gven sample. More formally: P x C ) P C ) P C x) 3) P x) Equaton 4 s read as: "The probablty of a gven data nstance x to belong to class C s equal to the multplcaton of the lkelhood that x s comng from the dstrbuton that generates C and the probablty of observng C 's n the whole sample, normalzed by the evdence. Evdence s gven by: P x) P x C ) P ) 4) C and t s a normalzaton constant for all classes, thus t can be safely dscarded. Then Equaton 4 becomes: P C x) P x C ) P C ) 5)
4 In a classfcaton problem we compute the posteror probabltes PC x) for each class and choose the one wth the hghest posteror. Ths s equvalent to defnng a dscrmnant functon g x) for class C and g x) s derved from Equaton 6 by takng the logarthms for convenence. g x) log P x C )) log P C )) 6) In order to acheve a dscrmnant value, one needs to compute the pror and lkelhood terms. Pror probablty PC ) can be estmated from the sample by countng. The crtcal ssue s to choose a sutable dstrbuton for the lkelhood term Px C ). Ths s where the multvarate normal dstrbuton takes place. In ths study lkelhood term s modeled by the multvarate normal dstrbuton. Computng dscrmnant values for each class and assgnng the nstance to the class wth the hghest value s equvalent to usng Bayes Theorem for choosng the class wth the hghest posteror probablty. For the -class case, t s suffcent to construct a sngle dscrmnant by gx) = g 0 x) g 1 x). Usng dscrmnant pont of vew, we wll explan dfferent predctors n the followng secton. In all cases, an nstance x s classfed as C such that arg max g x)) k k 3.3. Quadratc Dscrmnant Assumpton: Each class has dstnct Σ and μ. Dervaton: Combnng Equaton and Equaton T 1 g x) log S ) x m S x m log P C )) 7) and by defnng new varables W, w and w 0, the quadratc dscrmnant s obtaned as T T g x) x W x w x w 8) where S W 9) 1 w S m 10) 1 1 T 1 w 0 m S m log S ) log P C )) 11) and S, m and PC ) are maxmum lkelhood estmates of Σ, μ and PC ) respectvely. Quadratc model consders the correlaton of the features dfferently for each class. In case of K-classes, the number of parameters to estmate s K.d.d) for covarance estmates and K.d) for mean estmates. Also K pror probablty estmatons are needed Lnear Dscrmnant Assumpton: Each class has a common Σ and dstnct μ Dervaton: Assumpton states that classes share a common covarance matrx. The estmator s found by ether usng the whole data sample or by the weghted average of class covarances whch s gven as S P C ) S 1) Placng ths term n Equaton 7 we get 1 T 1 T 1 g x) x S m m S m ) log P C )) 13) whch s now a lnear dscrmnant n the form of T g x) w x w 14) where w w 0 S m 15) 1 1 T 1 0 m S m log P C )) 16) Ths model consders the correlaton of the features but assumes the varances and correlaton of features are the same for both classes. The number of parameters to estmate for covarance matrx s now ndependent of K. For covarance estmates d.d), for mean estmates K.d) and for prors K parameters should be estmated Naïve Bayes Assumpton: Each class has a common Σ wth off dagonal entres equal to 0, and dstnct μ Dervaton: Assumpton states the ndependence of features by usng a dagonal covarance matrx. Then the model reduces to a unvarate model gven n Equaton 17. d t 1 x j m j g x) log P C )) 17) j 1 s j Ths model does not take the correlaton of the features nto account and t measures the devaton from the mean n terms of standard devatons. For Nave Bayes, d) covarance, K.d) mean and K pror parameters should be estmated. 4. Experments and Results Desgn of experments and evaluaton of results n software defect predcton problems have partcular mportance. Most of the experment desgns have mportant flaws such as self tests and nsuffcent
5 performance measures as reported n [10]. Most research reported only the accuracy of predctors as a performance ndcator. Examnng defect predcton datasets, t s easly seen that they are not balanced. In other words, the number of defectve nstances s much less than the number of nondefectve nstances. As ponted out n [10], one can acheve 95% accuracy on a 5% defectve dataset by buldng a dummy classfer that always classfes nstances as nondefectve. A framework of MxN experment desgn, whch means M replcatons of N holdout cross valdaton) experments, s also gven n [10] and addtonal performance measures are reported, such as probablty of detecton pd) and probablty of false alarm pf). Ths research follows the same notaton. 10-fold cross-valdaton approach s used n the experments. That s, datasets are dvded nto 10 bns, 9 bns are used for tranng and 1 bn s used for testng. Repeatng these 10 folds ensures that each bn s used for tranng and testng whle mnmzng samplng bas. Each holdout experment s also repeated 10 tmes and n each repetton the datasets are randomzed to overcome any orderng effect and to acheve relable statstcs. Reported results are the mean values of these 100 experments for each dataset. Quadratc dscrmnant QD), lnear dscrmnant LD) and Nave Bayes NB) are the predctors used n ths research. As performance measures pd, pf and balance bal) are reported. pd s a measure for correctly detectng defectve modules and t s the rato of the number of defectve predcted modules to the number of actual defectve modules. Obvously hgher pd's are desred. As the name suggests, pf s a measure for false alarms and t s nterpreted as the probablty of predctng a module as defectve whle t s not ndeed. pf s desred to have low values. Balance measure s used to choose the optmal pd, pf) pars such that area under the ROC curve s maxmzed and t s defned as the normalzed Eucldean dstance from the desred pont 0,1) to pd, pf) n a ROC curve. Table. Dataset Descrptons Fgure. Experment Desgn. The experments conducted n [10] are replcated and extended n ths study. Framework for experment desgn n [10] s followed and updated as n Fgure. In order to extract features, PCA and Isomap are performed on the log fltered data attrbutes. An advantage of log flterng s that t scales the features so that extreme values are handled. Another advantage of log flterng s that normal dstrbuton better fts to data. In other words, data attrbutes are assumed to be lognormal dstrbuted. 5 to 30 features are extracted for all datasets usng PCA and Isomap. Best subset of features reported n [10] s also used n the experments. Ths subset of features dffers n each dataset. The best performng dmensonaltes acheved by PCA and Isomap are also dfferent for each dataset. These observatons support the dea that there s no global set of features that descrbe the software. So, maxmum possble metrcs of software should be collected and analyzed as long as t s feasble to collect them. Name #Modules DefectRate %) CM PC PC PC PC KC KC MW For evaluaton, 8 dfferent publc datasets obtaned from NASA MDP repostory [1] are used. Sample szes vary from 15 to 5589 modules. Each dataset has 38 features representng statc code attrbutes. As seen n Table defect rates are too low whch consoldates the use of above mentoned performance measures. All mplementatons are done n MATLAB envronment usng standard toolboxes. Results are tabulated n Table 3. Mean results of pd, pf) pars selected by the bal measure after 10x10 holdout experments are gven. For PCA and ISO labeled entres, these results are selected from 5 to 30 features obtaned by PCA and Isomap respectvely. For SUB labeled entres, the best subset of features
6 Table 3. Results Performances Data Predctor pd%) pf%) bal%) CM1 SUB+NB PC1 PCA+NB PC PCA+NB PC3 PCA+LD PC4 PCA+QD KC3 ISO+NB KC4 ISO+LD MW1 ISO+LD Average: obtaned by InfoGan are used as reported n [10]. In Table 3, results ndcated n bold face are statstcally sgnfcant than other methods wth α = 0.05 after applyng a t-test, consderng pd performance measure. Subset selecton s better than feature extracton methods n only 1 out of 8 datasets CM1). In the remanng datasets, best performances are obtaned ether by applyng PCA or Isomap nstead of InfoGan. In PC1, PC, PC3 and PC4, best mean performances are acheved applyng PCA, and n KC3, KC4 and MW1 Isomap yelded better results. It s observed that Isomap gves the best performances on relatvely small datasets. As the module szes ncrease PCA performs better. Except PC3 dataset, our replcated results are smlar to reported mean results n [10]. But varances of replcated experments.e. subsettng) are larger than PCA and Isomap approach especally for pf measure. NB and LD are observed to behave smlarly whereas QD results are dfferent than NB and LD n terms of performance. It s observed for QD, that as the number of features ncrease, performances get worse especally for pf measure and the varances ncrease. Possble reason for ths s the complexty of the model.e. too many parameters to estmate). As for the predctors, Nave Bayes NB) s chosen 4 tmes, lnear dscrmnant LD) s chosen 3 tmes and quadratc dscrmnant QD) s chosen only once. From these results, t can be concluded that clams statng any of these predctors as the 'globally' correct one, should be avoded. As expected, no specfc confguraton of a feature selecton and a predctor s always better than the others. Even though NB s the majorty wnner, t s clearly seen that performances on some datasets are ncreased by usng multvarate methods: QD and LD. Applyng QD gves the best result n PC4 dataset, but t s not statstcally sgnfcant. It can be concluded that QD can be dscarded because of ts complexty. In cases where LD wns, statstcal sgnfcances are observed, so the addtonal complexty ntroduced can be justfed. There may be other predctors performng better than these. Constructng better predctors s an open ended problem and as better results are reported, the problem gets more dffcult due to celng effect.e. t s harder to confrm the hypothess that predctor A performs better than predctor B, when A and B perform maxmum achevable performance or close to t [3]. Overall performance of the approach mproves on the best results reported so far [10]. Prevous research reported mean pd, pf) = 71,5) whch yelds bal = 7 averaged over all datasets. Replcaton of these experments yeld mean pd, pf) = 64, 19) and bal = 71. After expermentng wth all possble combnatons of InfoGan, PCA, Isomap wth NB, LD and QD, an mprovement s observed by pckng the best combnatons for all datasets. Improved results yeld mean pd, pf) = 77, 5) where bal = 76. Whle no change n pf measure s observed, pd measure s mproved by 6%. A fnal comment should be made about the runnng tmes of algorthms. As expected, QD takes more tme than LD and NB. However ths dfference s not too sgnfcant. The domnant factor that affects the runnng tmes are the sample szes. 5. Conclusons and Future Work In ths research software defect predcton s consdered as a data mnng problem. Several experments are conducted, ncludng the replcaton of prevous research on publcly avalable datasets from NASA repostory. Performances of dfferent predctors together wth dfferent feature extracton methods are evaluated. Results are compared wth the best performances reported so far and some mprovements are observed. The prevous research advces that one should not seek for globally best subset of features, rather to focus on buldng predctors that combnes nformaton from multple features. In addton, authors also beleve that research should focus on a balanced combnaton of those. In other words, buldng successful predctors depends on how useful nformaton s suppled to them. Whle makng research on better predctors, research on obtanng useful nformaton from features should also be carred out. A contrbuton of ths research s usng lnear and nonlnear feature extracton methods n order to combne nformaton from multple features. In software defect predcton there s more research on feature subset selecton than feature extracton. Results
7 suggest that t s worth to explore more to deepen our knowledge on feature extracton studes. Another contrbuton of ths research s the modelng of correlatons among features. Improved results are obtaned by usng multvarate statstcal methods. Furthermore, the probabltes of predctons are provded by employng Bayesan approaches, whch can gve project managers and practtoners a better understandng of the defectveness of software modules. Further research should nvestgate the valdaton of the log normal dstrbuton assumpton of software data used n ths research. It s better practce to apply goodness of ft tests, rather than assumng a normal dstrbuton. Other exponental famly dstrbutons should also be nvestgated. Another research area s to nvestgate flters to transform data nto sutable dstrbutons. Acknowledgements Ths research s supported n part by Bogazc Unversty research fund under grant number BAP- 06HA104. Authors would lke to thank Koray Balcı, who has contrbuted to the earler versons of ths manuscrpt. References [1] E. Alpaydn, Introducton to Machne Learnng, The MIT Press, October 004. [] E. Ceylan, F. O. Kutlubay, and A. B. Bener, Software defect dentfcaton usng machne learnng technques, In Proceedngs of the 3nd EUROMICRO Conference on Software Engneerng and Advanced Applcatons, IEEE Computer Socety, Washngton, DC, USA, 006, pp [3] P. R. Cohen. Emprcal Methods for Artfcal Intllgence, The MIT Press, London, England, [4] T. Cox and M. Cox, Multdmensonal Scalng. Chapman & Hall, London, [5] V. de Slva and J. B. Tenenbaum, Global versus local methods n nonlnear dmensonalty reducton, In S. Becker, S. Thrun, and K. Obermayer, edtors, Advances n Neural Informaton Processng Systems, 15, MIT Press, Cambrdge, MA, 003, pp [6] N. E. Fenton and M. Nel, A crtque of software defect predcton models, IEEE Transactons. on Software. Engneerng., 55), 1999, pp [7] Guyon and Elsseff, An ntroducton to varable and feature selecton, Journal of Machne Learnng Research, 3, 003, pp [8] T. M. Khoshgoftaar and J. C. Munson, Predctng software development errors usng software complexty metrcs, IEEE Journal on Selected Areas n Communcatons, 8), Feb. 1990, pp [9] J. A. Lee, A. Lendasse, N. Donckers, and M. Verleysen., A robust nonlnear projecton method, In Proceedngs of ESANN 000, European Symposum on Artfcal Neural Networks, Bruges Belgum), 000, pp [10] T. Menzes, J. Greenwald, and A. Frank, Data mnng statc code attrbutes to learn defect predctors, IEEE Transactons on Software Engneerng, 331), 007, pp. 13. [11] J. Munson and Y. M. Khoshgoftaar, Regresson modellng of software qualty: emprcal nvestgaton, J. Electron. Mater., 196), 1990, pp [1] NASA/WVU IV&V Faclty, Metrcs Data Program, avalable from [13] M. Nel, Multvarate assessment of software products, Softw. Test., Verf. Relab., 14), 199, pp [14] D. E. Neumann, An enhanced neural network technque for software rsk analyss, IEEE Tranactons on. Software Engneerng, 89), 00, pp [15] G. Boettcher, T. Menzes and T. Ostrand, PROMISE Repostory of emprcal software engneerng data West Vrgna Unversty, Department of Computer Scence, 007 [16] J. B. Tenenbaum, V. de Slva, and J. C. Langford, A global geometrc framework for nonlnear dmensonalty reducton, Scence, 90, 000, pp [17] D. H. Wolpert and W. G. Macready, No free lunch theorems for optmzaton IEEE Transactons on Evolutonary Computaton, 11), Aprl 1997, pp. 67 8
Feature Reduction and Selection
Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components
More informationClassifier Selection Based on Data Complexity Measures *
Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.
More informationSupport Vector Machines
/9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.
More informationHelsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr)
Helsnk Unversty Of Technology, Systems Analyss Laboratory Mat-2.08 Independent research projects n appled mathematcs (3 cr) "! #$&% Antt Laukkanen 506 R ajlaukka@cc.hut.f 2 Introducton...3 2 Multattrbute
More informationLearning the Kernel Parameters in Kernel Minimum Distance Classifier
Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department
More informationCS 534: Computer Vision Model Fitting
CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust
More informationLECTURE : MANIFOLD LEARNING
LECTURE : MANIFOLD LEARNING Rta Osadchy Some sldes are due to L.Saul, V. C. Raykar, N. Verma Topcs PCA MDS IsoMap LLE EgenMaps Done! Dmensonalty Reducton Data representaton Inputs are real-valued vectors
More informationData Mining: Model Evaluation
Data Mnng: Model Evaluaton Aprl 16, 2013 1 Issues: Evaluatng Classfcaton Methods Accurac classfer accurac: predctng class label predctor accurac: guessng value of predcted attrbutes Speed tme to construct
More informationInvestigating the Performance of Naïve- Bayes Classifiers and K- Nearest Neighbor Classifiers
Journal of Convergence Informaton Technology Volume 5, Number 2, Aprl 2010 Investgatng the Performance of Naïve- Bayes Classfers and K- Nearest Neghbor Classfers Mohammed J. Islam *, Q. M. Jonathan Wu,
More informationX- Chart Using ANOM Approach
ISSN 1684-8403 Journal of Statstcs Volume 17, 010, pp. 3-3 Abstract X- Chart Usng ANOM Approach Gullapall Chakravarth 1 and Chaluvad Venkateswara Rao Control lmts for ndvdual measurements (X) chart are
More informationDetermining the Optimal Bandwidth Based on Multi-criterion Fusion
Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn
More informationAn Entropy-Based Approach to Integrated Information Needs Assessment
Dstrbuton Statement A: Approved for publc release; dstrbuton s unlmted. An Entropy-Based Approach to ntegrated nformaton Needs Assessment June 8, 2004 Wllam J. Farrell Lockheed Martn Advanced Technology
More informationThe Research of Support Vector Machine in Agricultural Data Classification
The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou
More informationWishing you all a Total Quality New Year!
Total Qualty Management and Sx Sgma Post Graduate Program 214-15 Sesson 4 Vnay Kumar Kalakband Assstant Professor Operatons & Systems Area 1 Wshng you all a Total Qualty New Year! Hope you acheve Sx sgma
More informationCluster Analysis of Electrical Behavior
Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School
More informationSubspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;
Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features
More informationy and the total sum of
Lnear regresson Testng for non-lnearty In analytcal chemstry, lnear regresson s commonly used n the constructon of calbraton functons requred for analytcal technques such as gas chromatography, atomc absorpton
More informationEdge Detection in Noisy Images Using the Support Vector Machines
Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona
More informationSome Advanced SPC Tools 1. Cumulative Sum Control (Cusum) Chart For the data shown in Table 9-1, the x chart can be generated.
Some Advanced SP Tools 1. umulatve Sum ontrol (usum) hart For the data shown n Table 9-1, the x chart can be generated. However, the shft taken place at sample #21 s not apparent. 92 For ths set samples,
More informationImplementation Naïve Bayes Algorithm for Student Classification Based on Graduation Status
Internatonal Journal of Appled Busness and Informaton Systems ISSN: 2597-8993 Vol 1, No 2, September 2017, pp. 6-12 6 Implementaton Naïve Bayes Algorthm for Student Classfcaton Based on Graduaton Status
More informationTerm Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task
Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto
More informationA Statistical Model Selection Strategy Applied to Neural Networks
A Statstcal Model Selecton Strategy Appled to Neural Networks Joaquín Pzarro Elsa Guerrero Pedro L. Galndo joaqun.pzarro@uca.es elsa.guerrero@uca.es pedro.galndo@uca.es Dpto Lenguajes y Sstemas Informátcos
More informationSmoothing Spline ANOVA for variable screening
Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory
More informationBOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET
1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School
More information12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification
Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero
More informationLecture 4: Principal components
/3/6 Lecture 4: Prncpal components 3..6 Multvarate lnear regresson MLR s optmal for the estmaton data...but poor for handlng collnear data Covarance matrx s not nvertble (large condton number) Robustness
More informationA Semi-parametric Regression Model to Estimate Variability of NO 2
Envronment and Polluton; Vol. 2, No. 1; 2013 ISSN 1927-0909 E-ISSN 1927-0917 Publshed by Canadan Center of Scence and Educaton A Sem-parametrc Regresson Model to Estmate Varablty of NO 2 Meczysław Szyszkowcz
More information6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour
6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the
More informationDetection of an Object by using Principal Component Analysis
Detecton of an Object by usng Prncpal Component Analyss 1. G. Nagaven, 2. Dr. T. Sreenvasulu Reddy 1. M.Tech, Department of EEE, SVUCE, Trupath, Inda. 2. Assoc. Professor, Department of ECE, SVUCE, Trupath,
More informationS1 Note. Basis functions.
S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type
More informationImprovement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration
Improvement of Spatal Resoluton Usng BlockMatchng Based Moton Estmaton and Frame Integraton Danya Suga and Takayuk Hamamoto Graduate School of Engneerng, Tokyo Unversty of Scence, 6-3-1, Nuku, Katsuska-ku,
More informationOutline. Type of Machine Learning. Examples of Application. Unsupervised Learning
Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton
More informationUser Authentication Based On Behavioral Mouse Dynamics Biometrics
User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA
More informationContent Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers
IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth
More informationClassifying Acoustic Transient Signals Using Artificial Intelligence
Classfyng Acoustc Transent Sgnals Usng Artfcal Intellgence Steve Sutton, Unversty of North Carolna At Wlmngton (suttons@charter.net) Greg Huff, Unversty of North Carolna At Wlmngton (jgh7476@uncwl.edu)
More informationBAYESIAN MULTI-SOURCE DOMAIN ADAPTATION
BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,
More informationProblem Set 3 Solutions
Introducton to Algorthms October 4, 2002 Massachusetts Insttute of Technology 6046J/18410J Professors Erk Demane and Shaf Goldwasser Handout 14 Problem Set 3 Solutons (Exercses were not to be turned n,
More informationProblem Definitions and Evaluation Criteria for Computational Expensive Optimization
Problem efntons and Evaluaton Crtera for Computatonal Expensve Optmzaton B. Lu 1, Q. Chen and Q. Zhang 3, J. J. Lang 4, P. N. Suganthan, B. Y. Qu 6 1 epartment of Computng, Glyndwr Unversty, UK Faclty
More informationParallelism for Nested Loops with Non-uniform and Flow Dependences
Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr
More informationOutline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1
4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:
More informationMeta-heuristics for Multidimensional Knapsack Problems
2012 4th Internatonal Conference on Computer Research and Development IPCSIT vol.39 (2012) (2012) IACSIT Press, Sngapore Meta-heurstcs for Multdmensonal Knapsack Problems Zhbao Man + Computer Scence Department,
More informationA Binarization Algorithm specialized on Document Images and Photos
A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a
More informationThree supervised learning methods on pen digits character recognition dataset
Three supervsed learnng methods on pen dgts character recognton dataset Chrs Flezach Department of Computer Scence and Engneerng Unversty of Calforna, San Dego San Dego, CA 92093 cflezac@cs.ucsd.edu Satoru
More informationOptimizing Document Scoring for Query Retrieval
Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng
More informationClassification / Regression Support Vector Machines
Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM
More informationAn Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices
Internatonal Mathematcal Forum, Vol 7, 2012, no 52, 2549-2554 An Applcaton of the Dulmage-Mendelsohn Decomposton to Sparse Null Space Bases of Full Row Rank Matrces Mostafa Khorramzadeh Department of Mathematcal
More informationUnsupervised Learning
Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and
More informationMULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION
MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION Paulo Quntlano 1 & Antono Santa-Rosa 1 Federal Polce Department, Brasla, Brazl. E-mals: quntlano.pqs@dpf.gov.br and
More informationIncremental Learning with Support Vector Machines and Fuzzy Set Theory
The 25th Workshop on Combnatoral Mathematcs and Computaton Theory Incremental Learnng wth Support Vector Machnes and Fuzzy Set Theory Yu-Mng Chuang 1 and Cha-Hwa Ln 2* 1 Department of Computer Scence and
More informationFEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur
FEATURE EXTRACTION Dr. K.Vjayarekha Assocate Dean School of Electrcal and Electroncs Engneerng SASTRA Unversty, Thanjavur613 41 Jont Intatve of IITs and IISc Funded by MHRD Page 1 of 8 Table of Contents
More informationProper Choice of Data Used for the Estimation of Datum Transformation Parameters
Proper Choce of Data Used for the Estmaton of Datum Transformaton Parameters Hakan S. KUTOGLU, Turkey Key words: Coordnate systems; transformaton; estmaton, relablty. SUMMARY Advances n technologes and
More informationMathematics 256 a course in differential equations for engineering students
Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the
More informationLecture 5: Multilayer Perceptrons
Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented
More informationA Fast Content-Based Multimedia Retrieval Technique Using Compressed Data
A Fast Content-Based Multmeda Retreval Technque Usng Compressed Data Borko Furht and Pornvt Saksobhavvat NSF Multmeda Laboratory Florda Atlantc Unversty, Boca Raton, Florda 3343 ABSTRACT In ths paper,
More informationType-2 Fuzzy Non-uniform Rational B-spline Model with Type-2 Fuzzy Data
Malaysan Journal of Mathematcal Scences 11(S) Aprl : 35 46 (2017) Specal Issue: The 2nd Internatonal Conference and Workshop on Mathematcal Analyss (ICWOMA 2016) MALAYSIAN JOURNAL OF MATHEMATICAL SCIENCES
More informationMachine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)
Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes
More informationAnnouncements. Supervised Learning
Announcements See Chapter 5 of Duda, Hart, and Stork. Tutoral by Burge lnked to on web page. Supervsed Learnng Classfcaton wth labeled eamples. Images vectors n hgh-d space. Supervsed Learnng Labeled eamples
More informationA Modified Median Filter for the Removal of Impulse Noise Based on the Support Vector Machines
A Modfed Medan Flter for the Removal of Impulse Nose Based on the Support Vector Machnes H. GOMEZ-MORENO, S. MALDONADO-BASCON, F. LOPEZ-FERRERAS, M. UTRILLA- MANSO AND P. GIL-JIMENEZ Departamento de Teoría
More informationLearning-Based Top-N Selection Query Evaluation over Relational Databases
Learnng-Based Top-N Selecton Query Evaluaton over Relatonal Databases Lang Zhu *, Wey Meng ** * School of Mathematcs and Computer Scence, Hebe Unversty, Baodng, Hebe 071002, Chna, zhu@mal.hbu.edu.cn **
More informationCompiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz
Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster
More informationHuman Face Recognition Using Generalized. Kernel Fisher Discriminant
Human Face Recognton Usng Generalzed Kernel Fsher Dscrmnant ng-yu Sun,2 De-Shuang Huang Ln Guo. Insttute of Intellgent Machnes, Chnese Academy of Scences, P.O.ox 30, Hefe, Anhu, Chna. 2. Department of
More informationFuzzy Filtering Algorithms for Image Processing: Performance Evaluation of Various Approaches
Proceedngs of the Internatonal Conference on Cognton and Recognton Fuzzy Flterng Algorthms for Image Processng: Performance Evaluaton of Varous Approaches Rajoo Pandey and Umesh Ghanekar Department of
More informationA Fast Visual Tracking Algorithm Based on Circle Pixels Matching
A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng
More informationFace Recognition Based on SVM and 2DPCA
Vol. 4, o. 3, September, 2011 Face Recognton Based on SVM and 2DPCA Tha Hoang Le, Len Bu Faculty of Informaton Technology, HCMC Unversty of Scence Faculty of Informaton Scences and Engneerng, Unversty
More informationJournal of Chemical and Pharmaceutical Research, 2014, 6(6): Research Article. A selective ensemble classification method on microarray data
Avalable onlne www.ocpr.com Journal of Chemcal and Pharmaceutcal Research, 2014, 6(6):2860-2866 Research Artcle ISSN : 0975-7384 CODEN(USA) : JCPRC5 A selectve ensemble classfcaton method on mcroarray
More informationFuzzy Modeling of the Complexity vs. Accuracy Trade-off in a Sequential Two-Stage Multi-Classifier System
Fuzzy Modelng of the Complexty vs. Accuracy Trade-off n a Sequental Two-Stage Mult-Classfer System MARK LAST 1 Department of Informaton Systems Engneerng Ben-Guron Unversty of the Negev Beer-Sheva 84105
More informationClassification of Face Images Based on Gender using Dimensionality Reduction Techniques and SVM
Classfcaton of Face Images Based on Gender usng Dmensonalty Reducton Technques and SVM Fahm Mannan 260 266 294 School of Computer Scence McGll Unversty Abstract Ths report presents gender classfcaton based
More informationNAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics
Introducton G10 NAG Fortran Lbrary Chapter Introducton G10 Smoothng n Statstcs Contents 1 Scope of the Chapter... 2 2 Background to the Problems... 2 2.1 Smoothng Methods... 2 2.2 Smoothng Splnes and Regresson
More informationSimulation: Solving Dynamic Models ABE 5646 Week 11 Chapter 2, Spring 2010
Smulaton: Solvng Dynamc Models ABE 5646 Week Chapter 2, Sprng 200 Week Descrpton Readng Materal Mar 5- Mar 9 Evaluatng [Crop] Models Comparng a model wth data - Graphcal, errors - Measures of agreement
More informationTsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance
Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for
More informationDeep Classification in Large-scale Text Hierarchies
Deep Classfcaton n Large-scale Text Herarches Gu-Rong Xue Dkan Xng Qang Yang 2 Yong Yu Dept. of Computer Scence and Engneerng Shangha Jao-Tong Unversty {grxue, dkxng, yyu}@apex.sjtu.edu.cn 2 Hong Kong
More informationEECS 730 Introduction to Bioinformatics Sequence Alignment. Luke Huan Electrical Engineering and Computer Science
EECS 730 Introducton to Bonformatcs Sequence Algnment Luke Huan Electrcal Engneerng and Computer Scence http://people.eecs.ku.edu/~huan/ HMM Π s a set of states Transton Probabltes a kl Pr( l 1 k Probablty
More informationReducing Frame Rate for Object Tracking
Reducng Frame Rate for Object Trackng Pavel Korshunov 1 and We Tsang Oo 2 1 Natonal Unversty of Sngapore, Sngapore 11977, pavelkor@comp.nus.edu.sg 2 Natonal Unversty of Sngapore, Sngapore 11977, oowt@comp.nus.edu.sg
More informationEXTENDED BIC CRITERION FOR MODEL SELECTION
IDIAP RESEARCH REPORT EXTEDED BIC CRITERIO FOR ODEL SELECTIO Itshak Lapdot Andrew orrs IDIAP-RR-0-4 Dalle olle Insttute for Perceptual Artfcal Intellgence P.O.Box 59 artgny Valas Swtzerland phone +4 7
More informationIntelligent Information Acquisition for Improved Clustering
Intellgent Informaton Acquston for Improved Clusterng Duy Vu Unversty of Texas at Austn duyvu@cs.utexas.edu Mkhal Blenko Mcrosoft Research mblenko@mcrosoft.com Prem Melvlle IBM T.J. Watson Research Center
More informationBiostatistics 615/815
The E-M Algorthm Bostatstcs 615/815 Lecture 17 Last Lecture: The Smplex Method General method for optmzaton Makes few assumptons about functon Crawls towards mnmum Some recommendatons Multple startng ponts
More informationUnsupervised Learning and Clustering
Unsupervsed Learnng and Clusterng Why consder unlabeled samples?. Collectng and labelng large set of samples s costly Gettng recorded speech s free, labelng s tme consumng 2. Classfer could be desgned
More informationMachine Learning 9. week
Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below
More informationEmpirical Distributions of Parameter Estimates. in Binary Logistic Regression Using Bootstrap
Int. Journal of Math. Analyss, Vol. 8, 4, no. 5, 7-7 HIKARI Ltd, www.m-hkar.com http://dx.do.org/.988/jma.4.494 Emprcal Dstrbutons of Parameter Estmates n Bnary Logstc Regresson Usng Bootstrap Anwar Ftranto*
More informationCSCI 5417 Information Retrieval Systems Jim Martin!
CSCI 5417 Informaton Retreval Systems Jm Martn! Lecture 11 9/29/2011 Today 9/29 Classfcaton Naïve Bayes classfcaton Ungram LM 1 Where we are... Bascs of ad hoc retreval Indexng Term weghtng/scorng Cosne
More informationAn Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation
17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 An Iteratve Soluton Approach to Process Plant Layout usng Mxed
More informationPositive Semi-definite Programming Localization in Wireless Sensor Networks
Postve Sem-defnte Programmng Localzaton n Wreless Sensor etworks Shengdong Xe 1,, Jn Wang, Aqun Hu 1, Yunl Gu, Jang Xu, 1 School of Informaton Scence and Engneerng, Southeast Unversty, 10096, anjng Computer
More informationA Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems
A Unfed Framework for Semantcs and Feature Based Relevance Feedback n Image Retreval Systems Ye Lu *, Chunhu Hu 2, Xngquan Zhu 3*, HongJang Zhang 2, Qang Yang * School of Computng Scence Smon Fraser Unversty
More informationAssignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009.
Farrukh Jabeen Algorthms 51 Assgnment #2 Due Date: June 15, 29. Assgnment # 2 Chapter 3 Dscrete Fourer Transforms Implement the FFT for the DFT. Descrbed n sectons 3.1 and 3.2. Delverables: 1. Concse descrpton
More informationHermite Splines in Lie Groups as Products of Geodesics
Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the
More informationConcurrent Apriori Data Mining Algorithms
Concurrent Apror Data Mnng Algorthms Vassl Halatchev Department of Electrcal Engneerng and Computer Scence York Unversty, Toronto October 8, 2015 Outlne Why t s mportant Introducton to Assocaton Rule Mnng
More informationCorrelative features for the classification of textural images
Correlatve features for the classfcaton of textural mages M A Turkova 1 and A V Gadel 1, 1 Samara Natonal Research Unversty, Moskovskoe Shosse 34, Samara, Russa, 443086 Image Processng Systems Insttute
More informationSpam Filtering Based on Support Vector Machines with Taguchi Method for Parameter Selection
E-mal Spam Flterng Based on Support Vector Machnes wth Taguch Method for Parameter Selecton We-Chh Hsu, Tsan-Yng Yu E-mal Spam Flterng Based on Support Vector Machnes wth Taguch Method for Parameter Selecton
More informationRecognizing Faces. Outline
Recognzng Faces Drk Colbry Outlne Introducton and Motvaton Defnng a feature vector Prncpal Component Analyss Lnear Dscrmnate Analyss !"" #$""% http://www.nfotech.oulu.f/annual/2004 + &'()*) '+)* 2 ! &
More informationFusion Performance Model for Distributed Tracking and Classification
Fuson Performance Model for Dstrbuted rackng and Classfcaton K.C. Chang and Yng Song Dept. of SEOR, School of I&E George Mason Unversty FAIRFAX, VA kchang@gmu.edu Martn Lggns Verdan Systems Dvson, Inc.
More informationRelevance Assignment and Fusion of Multiple Learning Methods Applied to Remote Sensing Image Analysis
Assgnment and Fuson of Multple Learnng Methods Appled to Remote Sensng Image Analyss Peter Bajcsy, We-Wen Feng and Praveen Kumar Natonal Center for Supercomputng Applcaton (NCSA), Unversty of Illnos at
More informationAdaptive Transfer Learning
Adaptve Transfer Learnng Bn Cao, Snno Jaln Pan, Yu Zhang, Dt-Yan Yeung, Qang Yang Hong Kong Unversty of Scence and Technology Clear Water Bay, Kowloon, Hong Kong {caobn,snnopan,zhangyu,dyyeung,qyang}@cse.ust.hk
More informationThe Discriminate Analysis and Dimension Reduction Methods of High Dimension
Open Journal of Socal Scences, 015, 3, 7-13 Publshed Onlne March 015 n ScRes. http://www.scrp.org/journal/jss http://dx.do.org/10.436/jss.015.3300 The Dscrmnate Analyss and Dmenson Reducton Methods of
More informationSupervised Nonlinear Dimensionality Reduction for Visualization and Classification
IEEE Transactons on Systems, Man, and Cybernetcs Part B: Cybernetcs 1 Supervsed Nonlnear Dmensonalty Reducton for Vsualzaton and Classfcaton Xn Geng, De-Chuan Zhan, and Zh-Hua Zhou, Member, IEEE Abstract
More informationA Robust Method for Estimating the Fundamental Matrix
Proc. VIIth Dgtal Image Computng: Technques and Applcatons, Sun C., Talbot H., Ourseln S. and Adraansen T. (Eds.), 0- Dec. 003, Sydney A Robust Method for Estmatng the Fundamental Matrx C.L. Feng and Y.S.
More informationSVM-based Learning for Multiple Model Estimation
SVM-based Learnng for Multple Model Estmaton Vladmr Cherkassky and Yunqan Ma Department of Electrcal and Computer Engneerng Unversty of Mnnesota Mnneapols, MN 55455 {cherkass,myq}@ece.umn.edu Abstract:
More informationReview of approximation techniques
CHAPTER 2 Revew of appromaton technques 2. Introducton Optmzaton problems n engneerng desgn are characterzed by the followng assocated features: the objectve functon and constrants are mplct functons evaluated
More informationLaplacian Eigenmap for Image Retrieval
Laplacan Egenmap for Image Retreval Xaofe He Partha Nyog Department of Computer Scence The Unversty of Chcago, 1100 E 58 th Street, Chcago, IL 60637 ABSTRACT Dmensonalty reducton has been receved much
More informationThe Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique
//00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy
More informationSkew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach
Angle Estmaton and Correcton of Hand Wrtten, Textual and Large areas of Non-Textual Document Images: A Novel Approach D.R.Ramesh Babu Pyush M Kumat Mahesh D Dhannawat PES Insttute of Technology Research
More information