Relational Lasso An Improved Method Using the Relations among Features

Size: px
Start display at page:

Download "Relational Lasso An Improved Method Using the Relations among Features"

Transcription

1 Relatonal Lasso An Improved Method Usng the Relatons among Features Kotaro Ktagawa Kumko Tanaka-Ish Graduate School of Informaton Scence and Technology, The Unversty of Tokyo Abstract Relatonal lasso s a method that ncorporates feature relatons wthn machne learnng. By usng automatcally obtaned nosy relatons among features, relatonal lasso learns an addtonal penalty parameter per feature, whch s then ncorporated n terms of a regularzer wthn the target optmzaton functon. Relatonal lasso has been tested on three dfferent tasks: text categorzaton, polarty estmaton, and parsng, where t was compared wth conventonal lasso and adaptve lasso (Zou, 2006) when usng a mult-class logstc regresson optmzaton method. Relatonal lasso outperformed these other lasso methods n the tests. 1 Introducton As machne learnng methods scale up and now deal wth mllons of features, we deally want to add all possble features wthout havng to manually verfy or consder the effectveness of each feature wth respect to the performance. In other words, we need an automatc way to explot any usable nformaton that can be obtaned from features. However, wth current machne learnng methods, addng nosy features could lower performance, so the user stll has to decde whch features are worth addng. Regularzaton methods have recently receved greater nterest because of ths need. Regularzaton s expressed as a constrant term wthn an optmzaton functon, where the term s gven as a functon regardng the mportance weght of each feature. Regularzaton provdes a means of mportance control embedded wthn the target optmzaton problem. Among the varous regularzers, lasso, proposed by (Tbshran, 1996) s the most wdely used because of ts mathematcal comprehensveness. Ths paper descrbes a method, whch we call relatonal lasso, that mproves upon the conventonal lasso method. We show that relatonal lasso mproves the overall performance of classfcaton compared wth that of other lasso methods. Ths study was motvated through a lmtaton we observed n conventonal lasso: that features could be nter-related, but such dependences are not ncorporated wthn the current regularzaton. Therefore, the conventonal method tends to favor correlated features, whch can lead to the mportance of non-correlated features beng neglected. Relatonal lasso overcomes ths lmtaton of conventonal lasso by ntroducng an addtonal penalty parameter for each feature; ths parameter s estmated automatcally gven the nosy relatons among features, where the relatons are also automatcally generated. Whle the proposed method does not add to the computatonal complexty of the conventonal regularzaton method, t mproves the qualty of classfcaton. We explan the method and show emprcal results from tests based on three dfferent text classfcaton and parsng tasks. Regularzaton was orgnally proposed as a way to avod over-fttng by favorng some small number of features. Our method presented n ths artcle explots ths approach further and attempts to estmate the mportance of each feature wthn the relatons t has wth other features. As explaned n more detal n the followng secton, attempts along the same lne have been made to ncorporate underlyng relatons among features, such as by fused lasso through the orderng among features (Tbshran et al., 2005), or by group lasso through groups among features (Yuan and Ln, 2006). However, fused lasso assumes problems wth features that can be ordered n some meanngful way, and group lasso requres under Proceedngs of the 5th Internatonal Jont Conference on Natural Language Processng, pages , Chang Ma, Thaland, November 8 13, c 2011 AFNLP

2 lyng group nformaton to be confgured before the method s appled. The closely related work to ours s adaptve lasso (Zou, 2006), whch ntroduces an addtonal parameter per feature. However, snce adaptve lasso does not explctly process relatons among features, the estmaton of ths addtonal parameter s completely dfferent from our method. Moreover, as wll be emprcally shown, our method outperforms adaptve lasso, whch n fact performed worse than even the conventonal method. Related work on regularzaton technques has shown the potental of regularzaton not only to prevent over-fttng, but also to serve as a knd of feature selecton. Relatonal lasso provdes another step along ths lne. Here, we show that our method outperforms other lasso methods n dfferent classfcaton tasks. Moreover, t works well even when nosy features are added, somethng that degrades the performance of other lasso methods. 2 Related Work As explaned, feature selecton s the key to our method s effectveness, and here we summarze the related work done along ths lne. A substantal number of studes have been done on feature selecton technques, and these technques can be classfed nto three categores accordng to (Guyon and Elsseeff, 2003): wrapper methods, flter methods and embedded methods. The wrapper s the most naïve way of selectng a subset of features through predctve accuracy (Kohav and John, 1997). The user searches the possble feature space greedly usng an nducton algorthm and selects the best subset wth the best predctve accuracy. Wrappers wth greedy algorthms can be computatonally expensve and the stepwse selecton s often trapped nto a local optmal soluton. Snce the possble number of subsets for a set s exponental to the number of features n the orgnal set, n practce the user typcally defnes the subset arbtrarly dependng on the category of features, as was tested by (Scott and Matwn, 1999). Ths makes mpossble any fne adjustment as to whch ndvdual features to use. Flter methods, on the other hand, select good features accordng to some crtera, thus provdng the means for selectng ndvdual features. These methods have been extensvely studed, and a good overvew s avalable n (Mannng, 1999). The representatves of the evaluaton functon for choosng good/bad features are ch-squares and mutual nformaton, and features havng hgher scores for these functons are consdered good features. Whle flterng methods are effectve and therefore often used, these methods are ndependent of the learnng method that they are used wth. Moreover, the performance s not guaranteed to mprove even though feature selecton s used. The last category s embedded methods, where the feature selecton s embedded wthn the overall classfcaton problem. The decson tree s an example of an embedded method, and machne learnng technques usng prunng steps have been studed (Perkns et al., 2003). Lasso, an acronym for least absolute shrnkage and selecton operator, usng the L 1 norm (Tbshran, 1996) s a computatonally effcent method for smultaneously achevng the estmaton and feature selecton. Although lasso helps acheve an effectve model, the L 1 norm could cause based estmaton among features (Knght and Fu, 2000) by not beng able to dstngush between truly sgnfcant and nosy features. To cope wth ths problem, (Zou and Haste, 2005) proposed the elastc net method, whch s expressed as the conjunct of L 1 and L 2 norms of the feature weghts. They show that ths method works when the number of features substantally exceeds the number of learnng data, and also when there s strong correlaton between some features. Another proposal s fused lasso, whch ncorporates the order of features (as found n ther numbers, such as found n the case when each mage pxel value forms a feature) (Tbshran et al., 2005). Here, the target applcaton s proten mass spectroscopy and gene expresson data, and the method s only applcable to a target where the order among features s explct, as n the case of gene or mage pxels. (Yuan and Ln, 2006) proposed grouped lasso, whch ncorporates underlyng groups among features. The fused and grouped lasso methods requre confguraton of the structure among features. Recently, a new approach called weghted lasso s proposed whch calculates the L 1 norm on features, each of whch s weghted. As one method, (Zou, 2006) proposed a two-step approach called 1081

3 adaptve lasso. Ths paper proposes an alternatve weghted lasso method, n whch the estmaton of the weghts s processed dfferently from that of adaptve lasso. Although the procedure of adaptve lasso s closest to relatonal lasso, the learnng of adaptve lasso does not explctly handle the relatons among features. All these methods are attempts to ncorporate the relatons, or structure, among features such as dependence, orderng and groups nto machne learnng through the framework of regularzaton. Although such dependence s not always gven or tractable, we beleve ths nformaton can be learned from some automatcally generated nosy relaton among features. 3 L1-Regularzaton of Mult-Class Logstc Regresson Before gong on to the man ponts of relatonal lasso, let us summarze the regularzaton framework that we adopt. Regularzaton s the general method used n classfcaton. The target functon has two terms, one for fttng and another for regularzaton. Ths second term penalzes the weghts acqured by each feature, typcally by ncorporatng the addton of ther norms nto a target functon for the classfer. Ths prevents the target functon becomng too over-ftted by favorng some specfc sets of features. In ths sense, the regularzaton term can be consdered as servng for feature selecton. Of the varous ways to defne the target functon, n ths paper we focus on the mult-class logstc regresson model and L1-regularzaton; namely, the lasso method. The fttng functon adopted n ths paper s a mult-class logstc regresson model, denoted as LR n the followng. LR s used to model the relatonshp between the nput vectors x = R n and labels y Y. The condtonal probablty for a label y gven x s defned as p(y x; w) = 1 Z(x) exp ( w T ϕ(x, y) ) Z(x) = y exp ( w T ϕ(x, y ) ), where ϕ(x, y) R m s the feature vector and w R m s the weght vector. When the tranng examples {(x, y )}( = 1,, l) are gven, mnmzaton of the loss functon L(w) = log p(y x ; w), s equvalent to the maxmum lkelhood estmaton. Regularzaton makes t possble to obtan a good model for LR, wthout restrctng the number of features, by mposng approprate restrctons on weght w. Of the dfferent ways of regularzaton, n ths paper we adopt (Tbshran, 1996) s method of mposng an L 1 norm on parameters because of ts mathematcal smplcty, especally when appled wth LR. Ths method s called lasso and facltates both estmaton and automatc varable selecton. When applyng ths lasso to logstc regresson, the MAP estmaton of weghts for each feature s gven by the followng formula, the target functon to be optmzed: w = arg mn L(w) + λ w w, (1) where λ s the parameter defnng the strength of the regularzaton term s nfluence on the optmzaton. Although our proposal apples n general to varous types of target functon, n ths paper we examne ts effectveness wthn ths partcular target functon. Ths target functon was chosen because the target functon of LR-lasso s mathematcally comprehensve, so mult-class logstc regresson and lasso are wdely appled. Further nvestgaton to determne whether our method works well for other target functons wll be part of our future work. 4 Relatonal Lasso The Proposed Method The lmtaton of conventonal lasso s that relatons among features cannot be ncorporated. For example, hghly correlated features whch lead towards a hgher performance could all acqure relatvely large weghts. Ths would lead to favorng a sngle aspect that counts for the classfcaton and neglectng other mnor but stll mportant aspects whch would enable better classfcaton. Ths happens when a hgh correlaton s found among features. Therefore, when one feature s favored, the other correlated features must be heavly penalzed, so that features whch count 1082

4 for classfcaton from a dfferent aspect are more favored. To express ths, we adopt the weghted lasso approach, so an addtonal penalzng parameter α for each feature s ntroduced n the second term of formula (1): w = arg mn L(w) + λ w α w. (2) The soluton found here, subject to the L 1 regularzer, s equvalent to the soluton obtaned from the constraned optmzaton problem: mnmze w s.t. L(w), α w γ. The parameter γ corresponds to λ of formula (2). Each parameter α determnes the penalty for w and drectly affects the mportance of the th feature. Prevous work on adaptve lasso (Zou, 2006) also ntroduces an addtonal parameter, such as α, n addton to the weght parameter. Adaptve lasso focuses on the presence of oracle propertes 1, and works n two stages of optmzaton. Frst weght w s learned wth conventonal lasso. Second, L 1 norm s re-weghted wth the parameters α as α = 1/ ŵ δ beng set from ntal lasso estmator ŵ usng a parameter δ, and the optmzaton problem s processed usng α. Therefore, ther way of learnng ths addtonal parameter does not explctly concern the explotaton of addtonal nformaton dfferent from the orgnal weght w. On the other hand, our α s estmated gven a nosy relaton among features, thus t plays a dfferent role from w. In ths sense, the way to handle ths addtonal parameter for relatonal lasso s completely dfferent from adaptve lasso. In other words, the orgnalty of our method les n usng α to express the relatons between underlyng features. The relatons between features are denoted as R, whch s provded to the proposed algorthm and used to estmate α. R denotes a parwse dependence relaton between features. That s, f there are m features n total, R (1,, m) 1 Oracle property (Fan and L, 2001) s satsfed f the optmzaton problem can correctly select the nonzero weghts wth probablty convergng to one and the estmators of the nonzero weghts are asymptotcally normal wth the same means and covarance that they would have f the zero coeffcents were known n advance. (1,, m). Unlke prevous work such as fused (Tbshran et al., 2005) and grouped lasso (Yuan and Ln, 2006), R n our work s a nosy relaton whch can be automatcally obtaned by scannng through features. Although there are varous possbltes for obtanng R, one way s through the ncluson relaton among features. Gven a par of features p and q, the feature q ncludes p, f n every data of the learnng data, when the value of feature q s nonzero, the value of feature p s always non-zero. For example, for the case of the adjectve economc and ts stem econom, the latter ncludes the former whle also beng a stem for other terms such as economy and economst. In the fnal classfcaton, t s unknown whch of econom and economc counts. For part-of-speech taggng, econom would not provde much nformaton snce t does not have a complete form, but for topc estmaton, econom mght provde suffcent nformaton by representng the terms economc, economy and economst. In both cases, when the two words appear as features, they share a tght relaton and when one s gven hgh mportance, the other wll as well n conventonal lasso. In relatonal lasso, f one representatve s selected, then other smlar features n the same group wll acqure less mportance by havng a larger penalty. The overall procedure s shown n Procedure 1. The procedure obtans three knds of learnng data nput, a relaton among features, and parameter values. Before optmzng the target functon, denoted n the second lne from the bottom, the procedure calculates α dependng on the gven relaton R among features. Ths procedure s expressed n terms of a whle-structure, whch enables the adjustment of penalty parameters for hghly correlated features. The processed feature s held n set F to avod any duplcate processng of features. In the whle-structure, features are selected one at a tme n the order of larger values of L(w) w, wth w beng the zero vector 2. The number of the selected feature s denoted as k. Then, for all ks whch are related to k n R, the α s enlarged by a 2 There are other possbltes for ths order of processng features, such as randomzng the order. In the Graftng method (Perkns et al., 2003), the processed feature s selected by calculatng L(w) every tme n the whle-structure, w whch s also possble wth relatonal lasso. Ths however remans as future work. 1083

5 Procedure 1 Relatonal Lasso Input: (x, y )( = 1,, n) Parameters λ, a 1, a 2 Relaton among m features R (1,, m) (1,, m) α = 1, w = 0, F = {} v = L(w) w whle F < m do k = arg max v k k F for all k F and (k, k) R do α k = α k + a 1 end for for all k F and (k, k ) R do α k = α k + a 2 end for F = F k end whle w = arg mn L(w) + λ k α k w k w return w certan constant a 1 and a 2, dependng on whether the feature k s ncluded or k s ncluded. When the whle procedure ends, F ncludes all the features. Fnally, w s estmated n terms of LR-lasso, where the second term s weghted further wth the thus estmated α. When some specfc w k becomes zero, ths means that the weght s consdered as not selected for the classfcaton task. One further mprovement that mght be possble for the above procedure s to repeat the whlestructure and the estmaton of w, so that α and w perform co-tranng; ths also remans for our future work. Moreover, the procedure presented here remans an ad hoc modfcaton of conventonal lasso based on our motvaton. A more proper mathematcal reformulaton of ths method wll be part of our future work. 5 Evaluaton 5.1 Expermental Settngs We wll consder the followng two feature sets. Standard features Standard and addtonal features Here, an addtonal feature set s ntroduced so that t can be nosy wth respect to the classfcaton. Therefore, the nterest les n whether the performance s better when we have the addtonal features than t s when we have only the standard features. We consder three methods: Conventonal lasso (Tbshran, 1996) Adaptve lasso (Zou, 2006) Relatonal lasso (proposed method) We are nterested n whether relatonal lasso performs better than the other methods. In practce we can select the best λ parameter used for lasso methods, usng cross-valdaton, although we examned multple λs, whch determne the strength of the regularzers nfluence as shown n formulas (1) and (2) of Sectons 3 and 4. The other parameters ntroduced n Secton 4 are set as follows. Adaptve lasso has the parameter δ = 1, whch s set as the common choce n (Krämer et al., 2009) and the parameter α s defned from ntal estmator ŵ as follows: α = max { 1 wˆ, 1 For relatonal lasso, parameters a 1 and a 2 were each set to 1. For L 1 regularzed LR, a coordnate descent method s mplemented by modfyng LIBLINEAR 3. Coordnate descent methods have been wdely appled elsewhere because of ther sutablty for applcaton to hgher-order problems (Yuam et al., 2010). For each par of a feature and a method, we consdered the followng three problems of text classfcaton, polarty estmaton, and statstcal parsng. The next three sectons explan the standard and addtonal feature sets, the relaton among features R, and the evaluaton scores. Task 1: Text Classfcaton Twenty Newsgroups (20NG) 4 were used as the dataset for text classfcaton. Ths collecton contans 18,846 Englsh documents parttoned across 20 dfferent news groups. The data was sorted by date, wth the frst 60% used as a tranng set and the remanng 40% used as a test set. A smple bag of words was used 3 cjln/ lblnear/ 4 provded by Jason Renne, csal.mt.edu/jrenne/20newsgroups/ }. 1084

6 Table 1: Features used for dependency parsng ungram for w n w 1, w, w j, w j+1, w j+2, w j+2 pos(w), lex(w) for w n w 1, w, w j, pos(w left ), lex(w left ) for w n w 1, w, pos(w rght ), lex(w rght ), pos(w head ),lex(w head ) bgram for (v, w) n (w, w j ), (w 1, w j ) pos(v)pos(w),pos(v)lex(w),lex(v)pos(w),lex(v)lex(w) add bgram for (v, w) n (w, w j+1 ), (w j, w j+1 ) pos(v)pos(w),pos(v)lex(w),lex(v)pos(w),lex(v)lex(w) add preposton for w n w j+1, w j+2, w j+3 lex(w )lex(w j )pos(w), pos(w )lex(w j )lex(w) (f w j s a preposton) as the standard feature sets, whereas all stems of all words were used as addtonal features. R was defned as the relaton between each word and ts stems. A mult-class classfcaton task s typcally evaluated by macro and mcro F1 values, so we also provded these values. Task 2: Polarty Estmaton Polarty dataset v2.0 5 was used as the second data set. The content of each data was a move revew n text, tagged wth the sentment of postve or negatve. The data conssted of 1,000 postve and 1,000 negatve revews. Snce the data set was small, the average accuracy was obtaned through 10-fold cross valdaton. Feature sets were bascally the same as for Task 1, where the standard was a bag of words, the addtonal set conssted of word stems, and relaton R was the relaton among words and ther stems. The evaluaton was based on the accuracy of the bnary classfcaton of postve/negatve. Task 3: Parsng We also tested the methods on a parsng task, whch was a task drastcally dfferent from tasks 1 and 2. We used CoNLL-X formatted sentences from the Wall Street Journal secton of the Penn Tree-bank. Sectons 2-21 were used as tranng data (39,832 sentences), and secton 23 was used as test data (2,416 sentences). The parsng algorthm we tested s the standard shft-reduce parsng proposed by (Nvre, 2003), where the parsng proceeds by successve determnaton of the relaton between two words (denoted as w and w j ). Such a determnaton s consdered a 4-class classfcaton problem that s modeled and learned by LR, augmented by the three lasso methods beng evaluated. 5 provded by Bo Pang, edu/people/pabo/move-revew-data/ The standard features used are lsted n Table 1. Here, pos(w) ndcates the part of speech of the word w, where w ndcates the th word of a gven sentence, and w left ndcates the already parsed dependent word of w placed to ts farthest left sde. The addtonal feature set ncluded all dependent words nvolvng w and w j, and all bgrams concernng words used as features n the standard set. In our dependency parsng task, we measured the word accuracy whch was defned as the rato of words assgned correct heads dvded by the total of all words. mcro F value 70% 75% 80% 85% 90% 95% relatonal conventonal adaptve number of non zero features Fgure 1: Task 1: Mcro F1 values for the number of non-zero features macro F value 70% 75% 80% 85% 90% 95% relaton conventonal adaptve number of non zero features Fgure 2: Task 1: Macro F1 values for the number of non-zero features 1085

7 5.2 Results Fgures 1 and 2 show the results for Task 1, Fgure 3 those for Task 2, and Fgure 4 those for Task 3. Horzontal axes show the number of features and vertcal axes show accuracy. Each graph has three lnes, ndcatng the conventonal, adaptve and relatonal lasso methods appled to the standard and addtonal features all together. Each lne has fve ponts, each correspondng to a dfferent value of λ. The horzontal coordnate was determned by countng how many features remaned non-zero for each value of λ. Overall, all fgures, except for Fgure 3 show that relatonal lasso outperformed the adaptve and conventonal lasso methods. Ths was to be expected, snce relatonal lasso has the relaton R as nput, unlke the conventonal lasso method. Ths confrms that nformaton from the underlyng R does mprove lasso performance. Curously, the performance of adaptve lasso for some fgures was lower than that of the conventonal method. The reason for ths wll be gven later n ths secton. As Fgure 3 and Fgure 4 show, the performance was compettve among the three lasso methods, when the number of features were small. However, wth a large number of features, relatonal lasso generally outperforms the other lassos. accuracy 80% 82% 84% 86% 88% 90% 92% relatonal conventonal adaptve number of non zero features Fgure 3: Task 2: Accuracy for the number of nonzero features It s dffcult to compare the performance snce dfferent λ values lead to dfferent levels of performance, so the maxmum performance obtaned by changng the λ value s shown n Table 2. Columns are for dfferent lasso methods wth standard and addtonal feature sets, whereas rows represent dfferent tasks. Note that the best values of λ dffer dependng on the pars of methods and features. accuracy 82% 84% 86% 88% 90% 92% relatonal conventonal adaptve number of non zero features [10^6] Fgure 4: Task 3: Word accuracy for the number of non-zero features Overall, the last column presents the hghest performance n each row, thus suggestng the effectveness of relatonal lasso. For Task 2, when features of standard and addtonal sets were used, the performance of the conventonal method decreased compared to that when only the standard set was used. Ths could happen f the addtonal feature set s nosy and the regularzer cannot explot the useful nformaton from the addtonal set of features. On the other hand, the performance of relatonal and adaptve lasso for the same task was mproved by extractng the useful nformaton; that s, the performance was hgher than when usng only the standard features. Ths shows that the use of underlyng nformaton among features enhances the overall performance. For Tasks 1 and 3, addng features led to better performance than when usng only the standard set. Note, though, that the performance ncrease was greatest for relatonal lasso. Thus, relatonal lasso s the best among the three lasso methods at explotng nformaton, and thus performs better n terms of accuracy. In ths table, too, we see that for Task 2, the performance of adaptve lasso s below that of the conventonal lasso. We consder the reason for ths to be as follows. Snce the optmzaton for adaptve lasso s done n two stages, some of the features are dropped wthn the frst stage. In the second stage, these features wll never re-acqure any mportance. In other words, the feature selecton must be done at the very end n order to preserve the possblty of some features to re-acqure the mportance through learnng. Before endng, we must note the mpact of relatonal lasso on the speed of overall processng. 1086

8 Table 2: Maxmum Performance among Varous Values of λ for Three Lasso Methods Lasso Methods Conventonal Adaptve Relatonal Feature Sets Std Std+Add Std Std+Add Std Std+Add Task 1 (mcro) 79.67% 81.03% 76.78% 77.16% 78.87% 81.81% Task 1 (macro) 79.43% 80.69% 76.53% 77.67% 78.52% 81.72% Task % 84.95% 85.56% 86.1% 84.82% 86.45% Task % 88.81% 74.64% 87.71% 75.97% 89.23% The pre-processng to obtan α s very fast, snce t only scans the number of features once. The bottleneck of the procedure les n the estmaton of w snce ths requres convergence through a repettve procedure. Therefore, the computatonal complexty of relatonal lasso wll not change even wth α wthn the regularzer, and the overall speed of relatonal lasso s almost the same as that of the conventonal method. In contrast, adaptve lasso requres twce as much tme snce the bottleneck part s done twce. In ths sense, our method outperforms adaptve lasso n speed and s not sgnfcantly slower than conventonal, at least for the settngs we have examned n ths secton. 6 Concluson Relatonal lasso utlzes relatons among features to better explot nformaton through regularzaton, especally through lasso methods. The conventonal lasso method s not desgned to ncorporate relatons among features, and ths leads to based weghtng of a group of features havng smlar behavor. Relatonal lasso controls such relatons underlyng features by ntroducng a penalty parameter for each feature. The penalty ncreases when a feature s related to some other feature havng less of a penalty. Ths parameter score s ncorporated as the regularzaton term of the target machne learnng functon for the optmzaton objectve. We compared relatonal lasso to the conventonal method and the adaptve lasso proposed by (Zou, 2006), whch also uses an addtonal parameter per feature. We evaluated the methods based on how well they performed three tasks of text categorzaton, polarty estmaton, and parsng. Relatonal lasso outperformed the other lasso methods n these tasks. Moreover, the performance of the conventonal lasso methods deterorated when nosy features were added, whle relatonal lasso successfully extracted useful nformaton from these features and ts performance mproved. As part of our future work, we plan to nvestgate whether our method works for other tasks such as taggng, and wth other target functons. Moreover, there are many drectons we can take to further mprove the method, such as through cotranng. Last, t wll be nterestng to see how our method can be mathematcally reformulated. References Janqng Fan, Runze L Varable selecton va nonconcave penalzed lkelhood and ts oracle propertes. Journal of the Amercan Statstcal Assocaton, vol.96, pp Isabelle Guyon, André Elsseeff An ntroducton to varable and feature selecton. The Journal of Machne Learnng Research, vol.3, pp G. V. Kass An exploratory technque for nvestgatng large quanttes of categorcal data. Journal of the Royal Statstcal Socety. Seres C, vol.29, pp Keth Knght, Wenjang Fu Asymptotcs for lassotype estmators. The Annals of Statstcs, vol.28, pp Ron Kohav, George H. John Wrappers for feature subset selecton, Artfcal ntellgence, vol.97, pp Ncole Krämer, Julane Schäfer, Anne-Laure Boulestex Regularzed estmaton of large-scale gene assocaton networks usng graphcal Gaussan models. BMC Bonformatcs, vo.10. Chrstopher Mannng, Hnrch Schuetze Foundatons of Statstcal Natural Language Processng. MIT Press. Joakm Nvre An effcent algorthm for projectve dependency parsng. Proceedngs of the 8th Internatonal Workshop on Parsng Technologes, pp Smon Perkns, Kervn Lacker, and James Theler Graftng: Fast, Incremental Feature Selecton by Gradent Descent n Functon Space. The Journal of Machne Learnng Research, vol.3, pp Yvan Saeys, Iñak Inza, and Pedro Larrañaga A revew of feature selecton technques n bonformatcs. Bonformatcs, vol.23, num.19, pp Sam Scott, Stan Matwn Feature engneerng for text classfcaton. Proceedngs of ICML-99, 16th Internatonal Conference on Machne Learnng, pp

9 Robert Tbshran Regresson shrnkage and selecton va the lasso. Journal of the Royal Statstcal Socety, Seres B, vol.58, pp Robert Tbshran, Mchael Saunders, Saharon Rosset, J Zhu, and Keth Knght Sparsty and smoothness va the fused lasso. Journal of the Royal Statstcal Socety, Seres B, vol.67, pp Guo-Xun Yuan, Ka-We Chang, Cho-Ju Hseh, Chh-Jen Ln A comparson of optmzaton methods and software for large-scale l1-regularzed lnear classfcaton. The Journal of Machne Learnng Research, vol.11, pp Mng Yuan, Y Ln Model selecton and estmaton n regresson wth grouped varables. Journal of the Royal Statstcal Socety, Seres B, vol.68, pp Yng Yang, Jan O. Pedersen A comparatve study on feature selecton n text categorzaton. Proceedngs of the Fourteenth Internatonal Conference on Machne Learnng, pp Hu Zou The adaptve lasso and ts oracle propertes. Journal of the Amercan Statstcal Assocaton, vol.101, pp Hu Zou, Trevor Haste Regularzaton and varable selecton va the elastc net. Journal of the Royal Statstcal Socety, Seres B, vol.67, pp

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration Improvement of Spatal Resoluton Usng BlockMatchng Based Moton Estmaton and Frame Integraton Danya Suga and Takayuk Hamamoto Graduate School of Engneerng, Tokyo Unversty of Scence, 6-3-1, Nuku, Katsuska-ku,

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

TN348: Openlab Module - Colocalization

TN348: Openlab Module - Colocalization TN348: Openlab Module - Colocalzaton Topc The Colocalzaton module provdes the faclty to vsualze and quantfy colocalzaton between pars of mages. The Colocalzaton wndow contans a prevew of the two mages

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 A mathematcal programmng approach to the analyss, desgn and

More information

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng

More information

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 An Iteratve Soluton Approach to Process Plant Layout usng Mxed

More information

Discriminative Dictionary Learning with Pairwise Constraints

Discriminative Dictionary Learning with Pairwise Constraints Dscrmnatve Dctonary Learnng wth Parwse Constrants Humn Guo Zhuoln Jang LARRY S. DAVIS UNIVERSITY OF MARYLAND Nov. 6 th, Outlne Introducton/motvaton Dctonary Learnng Dscrmnatve Dctonary Learnng wth Parwse

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

Mathematics 256 a course in differential equations for engineering students

Mathematics 256 a course in differential equations for engineering students Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the

More information

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization Problem efntons and Evaluaton Crtera for Computatonal Expensve Optmzaton B. Lu 1, Q. Chen and Q. Zhang 3, J. J. Lang 4, P. N. Suganthan, B. Y. Qu 6 1 epartment of Computng, Glyndwr Unversty, UK Faclty

More information

Hermite Splines in Lie Groups as Products of Geodesics

Hermite Splines in Lie Groups as Products of Geodesics Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS46: Mnng Massve Datasets Jure Leskovec, Stanford Unversty http://cs46.stanford.edu /19/013 Jure Leskovec, Stanford CS46: Mnng Massve Datasets, http://cs46.stanford.edu Perceptron: y = sgn( x Ho to fnd

More information

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur FEATURE EXTRACTION Dr. K.Vjayarekha Assocate Dean School of Electrcal and Electroncs Engneerng SASTRA Unversty, Thanjavur613 41 Jont Intatve of IITs and IISc Funded by MHRD Page 1 of 8 Table of Contents

More information

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems A Unfed Framework for Semantcs and Feature Based Relevance Feedback n Image Retreval Systems Ye Lu *, Chunhu Hu 2, Xngquan Zhu 3*, HongJang Zhang 2, Qang Yang * School of Computng Scence Smon Fraser Unversty

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law) Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes

More information

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

Biostatistics 615/815

Biostatistics 615/815 The E-M Algorthm Bostatstcs 615/815 Lecture 17 Last Lecture: The Smplex Method General method for optmzaton Makes few assumptons about functon Crawls towards mnmum Some recommendatons Multple startng ponts

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS ARPN Journal of Engneerng and Appled Scences 006-017 Asan Research Publshng Network (ARPN). All rghts reserved. NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS Igor Grgoryev, Svetlana

More information

User Authentication Based On Behavioral Mouse Dynamics Biometrics

User Authentication Based On Behavioral Mouse Dynamics Biometrics User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

Edge Detection in Noisy Images Using the Support Vector Machines

Edge Detection in Noisy Images Using the Support Vector Machines Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona

More information

An Image Fusion Approach Based on Segmentation Region

An Image Fusion Approach Based on Segmentation Region Rong Wang, L-Qun Gao, Shu Yang, Yu-Hua Cha, and Yan-Chun Lu An Image Fuson Approach Based On Segmentaton Regon An Image Fuson Approach Based on Segmentaton Regon Rong Wang, L-Qun Gao, Shu Yang 3, Yu-Hua

More information

Machine Learning: Algorithms and Applications

Machine Learning: Algorithms and Applications 14/05/1 Machne Learnng: Algorthms and Applcatons Florano Zn Free Unversty of Bozen-Bolzano Faculty of Computer Scence Academc Year 011-01 Lecture 10: 14 May 01 Unsupervsed Learnng cont Sldes courtesy of

More information

APPLICATION OF MULTIVARIATE LOSS FUNCTION FOR ASSESSMENT OF THE QUALITY OF TECHNOLOGICAL PROCESS MANAGEMENT

APPLICATION OF MULTIVARIATE LOSS FUNCTION FOR ASSESSMENT OF THE QUALITY OF TECHNOLOGICAL PROCESS MANAGEMENT 3. - 5. 5., Brno, Czech Republc, EU APPLICATION OF MULTIVARIATE LOSS FUNCTION FOR ASSESSMENT OF THE QUALITY OF TECHNOLOGICAL PROCESS MANAGEMENT Abstract Josef TOŠENOVSKÝ ) Lenka MONSPORTOVÁ ) Flp TOŠENOVSKÝ

More information

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms Course Introducton Course Topcs Exams, abs, Proects A quc loo at a few algorthms 1 Advanced Data Structures and Algorthms Descrpton: We are gong to dscuss algorthm complexty analyss, algorthm desgn technques

More information

SVM-based Learning for Multiple Model Estimation

SVM-based Learning for Multiple Model Estimation SVM-based Learnng for Multple Model Estmaton Vladmr Cherkassky and Yunqan Ma Department of Electrcal and Computer Engneerng Unversty of Mnnesota Mnneapols, MN 55455 {cherkass,myq}@ece.umn.edu Abstract:

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for

More information

UB at GeoCLEF Department of Geography Abstract

UB at GeoCLEF Department of Geography   Abstract UB at GeoCLEF 2006 Mguel E. Ruz (1), Stuart Shapro (2), June Abbas (1), Slva B. Southwck (1) and Davd Mark (3) State Unversty of New York at Buffalo (1) Department of Lbrary and Informaton Studes (2) Department

More information

Meta-heuristics for Multidimensional Knapsack Problems

Meta-heuristics for Multidimensional Knapsack Problems 2012 4th Internatonal Conference on Computer Research and Development IPCSIT vol.39 (2012) (2012) IACSIT Press, Sngapore Meta-heurstcs for Multdmensonal Knapsack Problems Zhbao Man + Computer Scence Department,

More information

Wishing you all a Total Quality New Year!

Wishing you all a Total Quality New Year! Total Qualty Management and Sx Sgma Post Graduate Program 214-15 Sesson 4 Vnay Kumar Kalakband Assstant Professor Operatons & Systems Area 1 Wshng you all a Total Qualty New Year! Hope you acheve Sx sgma

More information

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour 6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the

More information

3D vector computer graphics

3D vector computer graphics 3D vector computer graphcs Paolo Varagnolo: freelance engneer Padova Aprl 2016 Prvate Practce ----------------------------------- 1. Introducton Vector 3D model representaton n computer graphcs requres

More information

The Codesign Challenge

The Codesign Challenge ECE 4530 Codesgn Challenge Fall 2007 Hardware/Software Codesgn The Codesgn Challenge Objectves In the codesgn challenge, your task s to accelerate a gven software reference mplementaton as fast as possble.

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

Fuzzy Filtering Algorithms for Image Processing: Performance Evaluation of Various Approaches

Fuzzy Filtering Algorithms for Image Processing: Performance Evaluation of Various Approaches Proceedngs of the Internatonal Conference on Cognton and Recognton Fuzzy Flterng Algorthms for Image Processng: Performance Evaluaton of Varous Approaches Rajoo Pandey and Umesh Ghanekar Department of

More information

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints Australan Journal of Basc and Appled Scences, 2(4): 1204-1208, 2008 ISSN 1991-8178 Sum of Lnear and Fractonal Multobjectve Programmng Problem under Fuzzy Rules Constrants 1 2 Sanjay Jan and Kalash Lachhwan

More information

An Entropy-Based Approach to Integrated Information Needs Assessment

An Entropy-Based Approach to Integrated Information Needs Assessment Dstrbuton Statement A: Approved for publc release; dstrbuton s unlmted. An Entropy-Based Approach to ntegrated nformaton Needs Assessment June 8, 2004 Wllam J. Farrell Lockheed Martn Advanced Technology

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

Classifying Acoustic Transient Signals Using Artificial Intelligence

Classifying Acoustic Transient Signals Using Artificial Intelligence Classfyng Acoustc Transent Sgnals Usng Artfcal Intellgence Steve Sutton, Unversty of North Carolna At Wlmngton (suttons@charter.net) Greg Huff, Unversty of North Carolna At Wlmngton (jgh7476@uncwl.edu)

More information

Active Contours/Snakes

Active Contours/Snakes Actve Contours/Snakes Erkut Erdem Acknowledgement: The sldes are adapted from the sldes prepared by K. Grauman of Unversty of Texas at Austn Fttng: Edges vs. boundares Edges useful sgnal to ndcate occludng

More information

Incremental Learning with Support Vector Machines and Fuzzy Set Theory

Incremental Learning with Support Vector Machines and Fuzzy Set Theory The 25th Workshop on Combnatoral Mathematcs and Computaton Theory Incremental Learnng wth Support Vector Machnes and Fuzzy Set Theory Yu-Mng Chuang 1 and Cha-Hwa Ln 2* 1 Department of Computer Scence and

More information

Solving two-person zero-sum game by Matlab

Solving two-person zero-sum game by Matlab Appled Mechancs and Materals Onlne: 2011-02-02 ISSN: 1662-7482, Vols. 50-51, pp 262-265 do:10.4028/www.scentfc.net/amm.50-51.262 2011 Trans Tech Publcatons, Swtzerland Solvng two-person zero-sum game by

More information

Support Vector Machines. CS534 - Machine Learning

Support Vector Machines. CS534 - Machine Learning Support Vector Machnes CS534 - Machne Learnng Perceptron Revsted: Lnear Separators Bnar classfcaton can be veed as the task of separatng classes n feature space: b > 0 b 0 b < 0 f() sgn( b) Lnear Separators

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

Three supervised learning methods on pen digits character recognition dataset

Three supervised learning methods on pen digits character recognition dataset Three supervsed learnng methods on pen dgts character recognton dataset Chrs Flezach Department of Computer Scence and Engneerng Unversty of Calforna, San Dego San Dego, CA 92093 cflezac@cs.ucsd.edu Satoru

More information

Backpropagation: In Search of Performance Parameters

Backpropagation: In Search of Performance Parameters Bacpropagaton: In Search of Performance Parameters ANIL KUMAR ENUMULAPALLY, LINGGUO BU, and KHOSROW KAIKHAH, Ph.D. Computer Scence Department Texas State Unversty-San Marcos San Marcos, TX-78666 USA ae049@txstate.edu,

More information

Virtual Machine Migration based on Trust Measurement of Computer Node

Virtual Machine Migration based on Trust Measurement of Computer Node Appled Mechancs and Materals Onlne: 2014-04-04 ISSN: 1662-7482, Vols. 536-537, pp 678-682 do:10.4028/www.scentfc.net/amm.536-537.678 2014 Trans Tech Publcatons, Swtzerland Vrtual Machne Mgraton based on

More information

Empirical Distributions of Parameter Estimates. in Binary Logistic Regression Using Bootstrap

Empirical Distributions of Parameter Estimates. in Binary Logistic Regression Using Bootstrap Int. Journal of Math. Analyss, Vol. 8, 4, no. 5, 7-7 HIKARI Ltd, www.m-hkar.com http://dx.do.org/.988/jma.4.494 Emprcal Dstrbutons of Parameter Estmates n Bnary Logstc Regresson Usng Bootstrap Anwar Ftranto*

More information

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET 1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School

More information

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

Face Detection with Deep Learning

Face Detection with Deep Learning Face Detecton wth Deep Learnng Yu Shen Yus122@ucsd.edu A13227146 Kuan-We Chen kuc010@ucsd.edu A99045121 Yzhou Hao y3hao@ucsd.edu A98017773 Mn Hsuan Wu mhwu@ucsd.edu A92424998 Abstract The project here

More information

A New Approach For the Ranking of Fuzzy Sets With Different Heights

A New Approach For the Ranking of Fuzzy Sets With Different Heights New pproach For the ankng of Fuzzy Sets Wth Dfferent Heghts Pushpnder Sngh School of Mathematcs Computer pplcatons Thapar Unversty, Patala-7 00 Inda pushpndersnl@gmalcom STCT ankng of fuzzy sets plays

More information

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics Introducton G10 NAG Fortran Lbrary Chapter Introducton G10 Smoothng n Statstcs Contents 1 Scope of the Chapter... 2 2 Background to the Problems... 2 2.1 Smoothng Methods... 2 2.2 Smoothng Splnes and Regresson

More information

Proper Choice of Data Used for the Estimation of Datum Transformation Parameters

Proper Choice of Data Used for the Estimation of Datum Transformation Parameters Proper Choce of Data Used for the Estmaton of Datum Transformaton Parameters Hakan S. KUTOGLU, Turkey Key words: Coordnate systems; transformaton; estmaton, relablty. SUMMARY Advances n technologes and

More information

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,

More information

SURFACE PROFILE EVALUATION BY FRACTAL DIMENSION AND STATISTIC TOOLS USING MATLAB

SURFACE PROFILE EVALUATION BY FRACTAL DIMENSION AND STATISTIC TOOLS USING MATLAB SURFACE PROFILE EVALUATION BY FRACTAL DIMENSION AND STATISTIC TOOLS USING MATLAB V. Hotař, A. Hotař Techncal Unversty of Lberec, Department of Glass Producng Machnes and Robotcs, Department of Materal

More information

Deep Classification in Large-scale Text Hierarchies

Deep Classification in Large-scale Text Hierarchies Deep Classfcaton n Large-scale Text Herarches Gu-Rong Xue Dkan Xng Qang Yang 2 Yong Yu Dept. of Computer Scence and Engneerng Shangha Jao-Tong Unversty {grxue, dkxng, yyu}@apex.sjtu.edu.cn 2 Hong Kong

More information

Adaptive Transfer Learning

Adaptive Transfer Learning Adaptve Transfer Learnng Bn Cao, Snno Jaln Pan, Yu Zhang, Dt-Yan Yeung, Qang Yang Hong Kong Unversty of Scence and Technology Clear Water Bay, Kowloon, Hong Kong {caobn,snnopan,zhangyu,dyyeung,qyang}@cse.ust.hk

More information

Learning a Class-Specific Dictionary for Facial Expression Recognition

Learning a Class-Specific Dictionary for Facial Expression Recognition BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 16, No 4 Sofa 016 Prnt ISSN: 1311-970; Onlne ISSN: 1314-4081 DOI: 10.1515/cat-016-0067 Learnng a Class-Specfc Dctonary for

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

Keywords - Wep page classification; bag of words model; topic model; hierarchical classification; Support Vector Machines

Keywords - Wep page classification; bag of words model; topic model; hierarchical classification; Support Vector Machines (IJCSIS) Internatonal Journal of Computer Scence and Informaton Securty, Herarchcal Web Page Classfcaton Based on a Topc Model and Neghborng Pages Integraton Wongkot Srura Phayung Meesad Choochart Haruechayasak

More information

Intra-Parametric Analysis of a Fuzzy MOLP

Intra-Parametric Analysis of a Fuzzy MOLP Intra-Parametrc Analyss of a Fuzzy MOLP a MIAO-LING WANG a Department of Industral Engneerng and Management a Mnghsn Insttute of Technology and Hsnchu Tawan, ROC b HSIAO-FAN WANG b Insttute of Industral

More information

Classification / Regression Support Vector Machines

Classification / Regression Support Vector Machines Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM

More information

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster

More information

X- Chart Using ANOM Approach

X- Chart Using ANOM Approach ISSN 1684-8403 Journal of Statstcs Volume 17, 010, pp. 3-3 Abstract X- Chart Usng ANOM Approach Gullapall Chakravarth 1 and Chaluvad Venkateswara Rao Control lmts for ndvdual measurements (X) chart are

More information

Feature Subset Selection Based on Ant Colony Optimization and. Support Vector Machine

Feature Subset Selection Based on Ant Colony Optimization and. Support Vector Machine Proceedngs of the 7th WSEAS Int. Conf. on Sgnal Processng, Computatonal Geometry & Artfcal Vson, Athens, Greece, August 24-26, 27 182 Feature Subset Selecton Based on Ant Colony Optmzaton and Support Vector

More information

Collaboratively Regularized Nearest Points for Set Based Recognition

Collaboratively Regularized Nearest Points for Set Based Recognition Academc Center for Computng and Meda Studes, Kyoto Unversty Collaboratvely Regularzed Nearest Ponts for Set Based Recognton Yang Wu, Mchhko Mnoh, Masayuk Mukunok Kyoto Unversty 9/1/013 BMVC 013 @ Brstol,

More information

Analysis of Continuous Beams in General

Analysis of Continuous Beams in General Analyss of Contnuous Beams n General Contnuous beams consdered here are prsmatc, rgdly connected to each beam segment and supported at varous ponts along the beam. onts are selected at ponts of support,

More information

CSCI 5417 Information Retrieval Systems Jim Martin!

CSCI 5417 Information Retrieval Systems Jim Martin! CSCI 5417 Informaton Retreval Systems Jm Martn! Lecture 11 9/29/2011 Today 9/29 Classfcaton Naïve Bayes classfcaton Ungram LM 1 Where we are... Bascs of ad hoc retreval Indexng Term weghtng/scorng Cosne

More information

Spam Filtering Based on Support Vector Machines with Taguchi Method for Parameter Selection

Spam Filtering Based on Support Vector Machines with Taguchi Method for Parameter Selection E-mal Spam Flterng Based on Support Vector Machnes wth Taguch Method for Parameter Selecton We-Chh Hsu, Tsan-Yng Yu E-mal Spam Flterng Based on Support Vector Machnes wth Taguch Method for Parameter Selecton

More information

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data A Fast Content-Based Multmeda Retreval Technque Usng Compressed Data Borko Furht and Pornvt Saksobhavvat NSF Multmeda Laboratory Florda Atlantc Unversty, Boca Raton, Florda 3343 ABSTRACT In ths paper,

More information

Comparing High-Order Boolean Features

Comparing High-Order Boolean Features Brgham Young Unversty BYU cholarsarchve All Faculty Publcatons 2005-07-0 Comparng Hgh-Order Boolean Features Adam Drake adam_drake@yahoo.com Dan A. Ventura ventura@cs.byu.edu Follow ths and addtonal works

More information

CAN COMPUTERS LEARN FASTER? Seyda Ertekin Computer Science & Engineering The Pennsylvania State University

CAN COMPUTERS LEARN FASTER? Seyda Ertekin Computer Science & Engineering The Pennsylvania State University CAN COMPUTERS LEARN FASTER? Seyda Ertekn Computer Scence & Engneerng The Pennsylvana State Unversty sertekn@cse.psu.edu ABSTRACT Ever snce computers were nvented, manknd wondered whether they mght be made

More information

Rules for Using Multi-Attribute Utility Theory for Estimating a User s Interests

Rules for Using Multi-Attribute Utility Theory for Estimating a User s Interests Rules for Usng Mult-Attrbute Utlty Theory for Estmatng a User s Interests Ralph Schäfer 1 DFKI GmbH, Stuhlsatzenhausweg 3, 66123 Saarbrücken Ralph.Schaefer@dfk.de Abstract. In ths paper, we show that Mult-Attrbute

More information

A Robust LS-SVM Regression

A Robust LS-SVM Regression PROCEEDIGS OF WORLD ACADEMY OF SCIECE, EGIEERIG AD ECHOLOGY VOLUME 7 AUGUS 5 ISS 37- A Robust LS-SVM Regresson József Valyon, and Gábor Horváth Abstract In comparson to the orgnal SVM, whch nvolves a quadratc

More information

Query Clustering Using a Hybrid Query Similarity Measure

Query Clustering Using a Hybrid Query Similarity Measure Query clusterng usng a hybrd query smlarty measure Fu. L., Goh, D.H., & Foo, S. (2004). WSEAS Transacton on Computers, 3(3), 700-705. Query Clusterng Usng a Hybrd Query Smlarty Measure Ln Fu, Don Hoe-Lan

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

Conditional Speculative Decimal Addition*

Conditional Speculative Decimal Addition* Condtonal Speculatve Decmal Addton Alvaro Vazquez and Elsardo Antelo Dep. of Electronc and Computer Engneerng Unv. of Santago de Compostela, Span Ths work was supported n part by Xunta de Galca under grant

More information

Fitting: Deformable contours April 26 th, 2018

Fitting: Deformable contours April 26 th, 2018 4/6/08 Fttng: Deformable contours Aprl 6 th, 08 Yong Jae Lee UC Davs Recap so far: Groupng and Fttng Goal: move from array of pxel values (or flter outputs) to a collecton of regons, objects, and shapes.

More information

BioTechnology. An Indian Journal FULL PAPER. Trade Science Inc.

BioTechnology. An Indian Journal FULL PAPER. Trade Science Inc. [Type text] [Type text] [Type text] ISSN : 0974-74 Volume 0 Issue BoTechnology 04 An Indan Journal FULL PAPER BTAIJ 0() 04 [684-689] Revew on Chna s sports ndustry fnancng market based on market -orented

More information

Performance Evaluation of Information Retrieval Systems

Performance Evaluation of Information Retrieval Systems Why System Evaluaton? Performance Evaluaton of Informaton Retreval Systems Many sldes n ths secton are adapted from Prof. Joydeep Ghosh (UT ECE) who n turn adapted them from Prof. Dk Lee (Unv. of Scence

More information

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 48 CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 3.1 INTRODUCTION The raw mcroarray data s bascally an mage wth dfferent colors ndcatng hybrdzaton (Xue

More information

Overlapping Clustering with Sparseness Constraints

Overlapping Clustering with Sparseness Constraints 2012 IEEE 12th Internatonal Conference on Data Mnng Workshops Overlappng Clusterng wth Sparseness Constrants Habng Lu OMIS, Santa Clara Unversty hlu@scu.edu Yuan Hong MSIS, Rutgers Unversty yhong@cmc.rutgers.edu

More information

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces Range mages For many structured lght scanners, the range data forms a hghly regular pattern known as a range mage. he samplng pattern s determned by the specfc scanner. Range mage regstraton 1 Examples

More information