Solving MultiClass Support Vector Machines with LaRank

Size: px
Start display at page:

Download "Solving MultiClass Support Vector Machines with LaRank"

Transcription

1 Solvng MultClass Support Vector Machnes wth LaRank Antone Bordes Léon Bottou Patrck Gallnar Jason Weston ( ) NEC Laboratores Amerca, Inc., 4 Independence Way, Prnceton, NJ08540, USA. ( ) LIP6, Unversté de Pars 6, 104 Avenue du Pdt Kennedy, Pars, France Abstract Optmzaton algorthms for large margn multclass recognzers are often too costly to handle ambtous problems wth structured outputs and exponental numbers of classes. Optmzaton algorthms that rely on the full gradent are not effectve because, unlke the soluton, the gradent s not sparse and s very large. The LaRank algorthm sdesteps ths dffculty by relyng on a randomzed exploraton nspred by the perceptron algorthm. We show that ths approach s compettve wth gradent based optmzers on smple multclass problems. Furthermore, a sngle LaRank pass over the tranng examples delvers test error rates that are nearly as good as those of the fnal soluton. 1. Introducton Much has been wrtten about the recognton of multple classes usng large margn kernel machnes such as support vector machnes (SVMs). The most wdely used approaches combne multple bnary classfers separately traned usng ether the one-versus-all or one-versus-one scheme (e.g. Hsu & Ln, 2002). Alternatve proposals (Weston & Watkns, 1998; Crammer & Snger, 2001) reformulate the large margn problem to drectly address the multclass problem. These algorthms are more expensve because they must smultaneously handle all the support vectors assocated wth dfferent nter-class boundares. Rgorous experments (Hsu & Ln, 2002; Rfkn & Klautau, 2004) suggest that ths hgher cost does not translate nto hgher generalzaton performance. Appearng n Proceedngs of the 24 th Internatonal Conference on Machne Learnng, Corvalls, OR, Copyrght 2007 by the author(s)/owner(s). The pcture changes when one consders learnng systems that predct structured outputs (e.g. Bakır et al., 2007). Instead of predctng a class label y for each pattern x, structured output systems produce complex dscrete outputs such as a sequences, trees, or graphs. Snce these potental outputs can be enumerated (n theory), these systems can be vewed as multclass problems wth a number of classes growng exponentally wth the characterstc sze of the output. Dealng wth so many classes n a large margn classfer would be nfeasble wthout smart factorzatons that leverage the specfc structure of the outputs (Taskar et al., 2005; Tsochantards et al., 2005). Ths s best acheved usng a drect multclass formulaton because the factorzaton of the output space mples that all the classes are handled smultaneously. It s therefore mportant to reduce the computatonal cost of multclass SVMs wth a potentally large number of classes. MCSVM Crammer and Snger (2001) propose a multclass formulaton that we call partal rankng. The dual cost s a functon of a n k matrx of Lagrange coeffcents where n s the number of examples and k the number of classes. Each teraton of the MCSVM algorthm maxmzes the restrcton of the dual cost to a sngle row of the coeffcent matrx. Successve rows are selected usng the gradent of the cost functon. Unlke the coeffcents matrx, the gradent s not sparse. Ths approach s not feasble when the number of classes k grows exponentally, because the gradent becomes too large. SVMstruct Tsochantards et al. (2005) essentally use the same partal rankng formulaton for the SVMstruct system. The clever cuttng plane algorthm ensures convergence but only requres to store and compute a small part of the gradent. Ths crucal dfference makes SVMstruct sutable for structured output problems wth a large number of classes.

2 Solvng MultClass Support Vector Machnes wth LaRank Kernel Perceptrons Onlne algorthms nspred by the perceptron (Collns, 2002; Crammer & Snger, 2003) can be nterpreted as the successve soluton of optmzaton subproblems restrcted to coeffcents assocated wth the current tranng example. There s no need to represent the gradent. The random orderng of the tranng examples drves the successve optmzatons. Perceptrons provde surprsngly strong theoretcal guarantees (Graepel et al., 2000). They run very quckly but provde nferor generalzaton performances n practce. LaRank Ths paper proposes LaRank, a stochastc learnng algorthm that combnes partal gradent nformaton wth the randomzaton arsng from the sequence of tranng examples. LaRank uses gradents as sparngly as SVMstruct and yet runs consderably faster. In fact, LaRank reaches an equvalent accuracy faster than algorthms that use the full gradent nformaton. LaRank generalzes better than perceptron-based algorthms. In fact, LaRank provdes the performance of SVMstruct or MCSVM because t solves the same optmzaton problem. LaRank acheves nearly optmal test error rates after a sngle pass over the randomly reordered tranng set. Therefore, LaRank offers the practcalty of an onlne algorthm. Ths paper frst revews and dscusses the multclass formulaton of Crammer and Snger. Then t presents the LaRank algorthm, dscusses ts convergence, and reports expermental results on well known multclass problems. 2. Multclass Support Vector Machnes Ths secton descrbes the partal rankng formulaton of multclass SVMs (Crammer & Snger, 2001). The presentaton frst follows (Tsochantards et al., 2005) then ntroduces a new parametrzaton of the dual program Partal Rankng We want to learn a functon f that maps patterns x to dscrete class labels y Y. We ntroduce a dscrmnant functon S(x, y) R that measures the correctness of the assocaton between pattern x and class label y. The optmal class label s then f(x) = arg maxs(x, y). (1) y Y We assume that the dscrmnant functon has the form S(x, y) = w, Φ(x, y) where Φ(x, y) maps the par (x, y) nto a sutable feature space endowed wth the dot product,. As usual wth kernel machnes, the feature mappng functon Φ s mplctly defned by the specfcaton of a jont kernel functon K(x, y, x, ȳ) = Φ(x, y), Φ( x, ȳ). (2) Consder tranng patterns x 1...x n and ther class labels y 1... y n Y. For each pattern x, we want to make sure that the score S(x, y ) of the correct assocaton s greater than the scores S(x, y), y y, of the ncorrect assocatons. Ths amounts to enforcng a partal order relatonshp on the elements of Y. Ths partal rankng can be expressed by constrants = 1...n y y w, δφ (y, y) 1 where δφ (y, ȳ) stands for Φ(x, y) Φ(x, ȳ). Followng the standard SVM dervaton, we ntroduce slack varables ξ to account for the potental volaton of the constrants and optmze a combnaton of the norm of w and of the sze of the slack varables. subject to mn w { ξ 0 1 n 2 w, w + C ξ (3) y y 2.2. Dual Programs =1 w, δφ (y, y) 1 ξ The usual dervaton leads to solvng the followng equvalent dual problem (Crammer & Snger, 2001; Tsochantards et al., 2005): max α,y y α y 1 2 subject to 8 < : y α αȳ j,y y j,ȳ y j y y α y 0 α y C y y δφ(y, y), δφj(yj, ȳ) Ths problem has n(k 1) varables α y, y y correspondng to the constrants of (3). Once we have the soluton, the dscrmnant functon s S(x, y) = δφ (y, ȳ), Φ(x, y),ȳ y αȳ Ths dual problem can be consderably smplfed by reparametrzng t wth nk varables β y defned as α y β y = f y y αȳ otherwse (5) ȳ y Note that only the β y can be postve. Substtutng n (4), and takng nto account the relaton y βy = 0, (4)

3 Solvng MultClass Support Vector Machnes wth LaRank leads to a much smpler expresson for the dual problem (the δφ (... ) have dsappeared.) max β subject to β y 1 β y 2 βȳ j Φ(x, y),φ(xj, ȳ),j,y,ȳ 8 y y β >< y 0 β y C >: β y = 0 The dscrmnant functon then becomes S(x, y) =,ȳ 2.3. Kernel Factorzaton y βȳ Φ(x, ȳ), Φ(x, y) In practce, smart factorzatons of the jont kernel (2) are crucal to reduce the memory requred to store or cache the kernel values. Ths paper focuses on smple multclass problems whose kernel functon (2) takes the form Φ(x, y), Φ( x, ȳ) = k(x, x) δ(y, ȳ) where k(x, x) s a kernel defned on the patterns, and where δ(y, ȳ) s 1 f y = ȳ and 0 otherwse. The dual problem (6) then becomes (6) max β y 1 β y β 2 βy j k(x, xj) 8,j y < y β y Cδ(y, y) (7) subject to β y : = 0 and the dscrmnant functon becomes S(x, y) = y β y k(x, x) When there are only two classes, ths reduces to the standard SVM soluton (wthout equalty constrant.) Structured output learnng systems (Tsochantards et al., 2005) call for much more sophstcated factorzatons of the jont kernel. For the sake of smplcty, we descrbe the LaRank algorthm n the context of the multclass problem (7) whch s the focus of ths paper. Dealng wth the general problem (6), or handlng the varable margns suggested by Tsochantards et al. (2005), only requres mnor changes. 3. Optmzaton Algorthm Durng the executon of the optmzaton algorthm, we call support vectors all pars (x, y) whose assocated coeffcent β y s non zero; we call support patterns all patterns x that appear n a support vector. The LaRank algorthm stores the followng data: The set S of the current support vectors. The coeffcents β y assocated wth the support vectors (x, y) S. Ths descrbes the soluton snce all the other β coeffcents are zero. The dervatves g (y) of the dual objectve functon wth respect to the coeffcents β y assocated wth the support vectors (x, y) S. g (y) = δ(y, y ) j βy j k(x, x j ) = δ(y, y ) S(x, y) (8) Note that we do not store or even compute the remanng coeffcents of the gradent. In general, these mssng dervatves are not zero because the gradent s not sparse. A nave mplementaton could smply precompute all the kernel values k(x, x j ). Ths would be a waste of processng tme because the locaton of the optmum depends only on the fracton of the kernel matrx that nvolves support patterns. Our code computes kernel values on demand and caches them n sets of the form E(y, j) = { k(x, x j ) such that (x, y) S }. Although ths cache stores several copes of the same kernel values, cachng ndvdual kernel values has a hgher overhead Elementary Step Problem (7) lends tself to a smple teratve algorthm whose elementary steps are nspred by the well known sequental mnmal optmzaton (SMO) algorthm (Platt, 1999). Algorthm 1 SmoStep(, y +, y ): 1: Retreve or compute g (y +). 2: Retreve or compute g (y ). 3: Let λ u = g (y + ) g (y ) 2 k(x,x ) 4: Let λ = max 0, mn( λ u, C δ(y +, y ) β y + ) 5: Update β y + β y + + λ and β y β y λ 6: Update S accordng to whether β y + and β y are zero. 7: Update gradents: j s.t. (x j, y +) S, g j(y +) g j(y +) + λk(x, x j) j s.t. (x j, y ) S, g j(y ) g j(y ) λk(x, x j) Each teraton starts wth the selecton of one pattern x and two classes y + and y. The elementary step modfes the coeffcents β y+ and β y by opposte amounts, β y+ β y+ + λ (9) β y β y λ where λ 0 maxmzes the dual objectve functon (7) subject to the constrants. Ths optmal value s

4 Solvng MultClass Support Vector Machnes wth LaRank easly computed by frst calculatng the unconstraned optmum λ u = g (y + ) g (y ) (10) 2 k(x, x ) and then enforcng the constrants λ = max { 0, mn( λ u, C δ(y +, y ) β y+ ) } (11) Fnally the stored dervatves g j (y) are updated to reflect the coeffcent update. Ths s summarzed n algorthm Step Selecton Strateges Popular SVM solvers based on SMO select successve steps by choosng the par of coeffcents that defnes the feasble search drecton wth the hghest gradent. We cannot use ths strategy because we have chosen to store only a small fracton of the gradent. Stochastc algorthms nspred by the perceptron perform qute well by successvely updatng coeffcents determned by randomly pckng tranng patterns. For nstance, n a multclass context, Taskar (2004, secton 6.1) terates over the randomly ordered patterns: for each pattern x, he computes the scores S(x, y) for all classes and runs SmoStep on the two most volatng classes, that s, the classes that defne the feasble search drecton wth the hghest gradent. In the context of bnary classfcaton, Bordes and Bottou (2005) observe that such perceptron-nspred updates lead to a slow optmzaton of the dual because the coeffcents correspondng to the few support vectors are not updated often enough. They suggest to alternatvely update the coeffcent correspondng to a fresh random example and the coeffcent correspondng to an example randomly chosen among the current support vectors. The related LaSVM algorthm (Bordes et al., 2005) alternates steps explotng a fresh random tranng example and steps explotng current support vectors selected usng the gradent. We now extend ths dea to the multclass formulaton. Snce the multclass problem has both support vectors and support patterns, we defne three ways to select a trple (, y +, y ) for the elementary SmoStep. Algorthm 2 ProcessNew(x ): 1: f x s a support pattern then ext. 2: y + y. 3: y arg mn y Y g (y) 4: Perform SmoStep(, y +, y ) ProcessNew (algorthm 2) operates on a pattern x that s not a support pattern. It chooses the classes y + and y that defne the feasble drecton wth the hghest gradent. Snce all the Algorthm 3 ProcessOld: 1: Randomly pck a support pattern x. 2: y + arg max y Y g (y) subject to β y < C δ(y,y) 3: y arg mn y Y g (y) 4: Perform SmoStep(, y +, y ) Algorthm 4 Optmze: 1: Randomly pck a support pattern x. 2: Let Y = { y Y such that (x, y) S } 3: y + arg max y Y g (y) subject to β y < C δ(y,y) 4: y arg mn y Y g (y) 5: Perform SmoStep(, y +, y ) β y are zero, y + s always y. Choosng of y conssts of fndng arg max y S(x, y) snce equaton (8) holds. ProcessOld (algorthm 3) randomly pcks a support pattern x. It chooses the classes y + and y that defne the feasble drecton wth the hghest gradent. The determnaton of y + mostly nvolves labels y such that β y < 0, for whch the correspondng dervatves g (y) are known. The determnaton of y agan conssts of computng arg max y S(x, y). Optmze (algorthm 4) resembles ProcessOld but pcks the classes y + and y among those that correspond to exstng support vectors (x, y + ) and (x, y ). Usng the gradent s fast because the relevant dervatves are already known and ther number s moderate. The ProcessNew operaton s closely related to the perceptron algorthm. It can be nterpreted as a stochastc gradent update for the mnmzaton of the generalzed margn loss (LeCun et al., 2007, 2.2.3), wth a step sze adjusted accordng to the curvature of the dual (Hldreth, 1957). Crammer and Snger (2003) use a very smlar approach for the MIRA algorthm Adaptve Schedule Prevous works (Bordes & Bottou, 2005; Bordes et al., 2005) smply alternate two step selecton strateges accordng to a fxed schedule. They also report results suggestng that the optmal schedule s n fact datadependent. We would lke to select at each teraton an operaton that causes a large ncrease of the dual n a small amount of tme. For each operaton type, LaRank mantans a runnng estmate of the average rato of the dual ncrease over the duraton. Runnng tmes are measured; dual ncreases are derved from the value of λ computed durng the elementary step. Each teraton of the LaRank algorthm (algorthm 5)

5 Solvng MultClass Support Vector Machnes wth LaRank Algorthm 5 LaRank: 1: S. 2: r Optmze, r ProcessOld, r ProcessNew 1. 3: loop 4: Randomly reorder the tranng examples. 5: k 1. 6: whle k n do 7: Pck operaton s wth odds proportonal to r s. 8: f s = Optmze then 9: Perform Optmze. 10: else f s = ProcessOld then 11: Perform ProcessOld. 12: else 13: Perform ProcessNew(x k ). 14: k k : end f 16: r s µ 17: end whle 18: end loop dual ncrease duraton + (1 µ) r s. randomly selects whch operaton to perform wth a probablty proportonal to these estmates. Our mplementaton uses µ = In order to facltate tmng, we treat sequences of ten Optmze as a sngle atomc operaton Correctness and Complexty Let ν 2 = max {k(x, x )} and let κ, τ, η be small postve tolerances. We assume that the algorthm mplementaton enforces the followng propertes: SmoStep exts when g (y + ) g (y ) τ. Optmze and ProcessOld chooses y + among the y that satsfy β y C δ(y, y ) κ. LaRank makes sure that every operaton has probablty greater than η to be selected at each teraton (see algorthm 5). We refer to ths as the (κ, τ, η)-algorthm. Theorem Wth probablty 1, the (κ, τ, η)-algorthm reaches a κτ-approxmate soluton of problem (7), wth no more than max{ 2ν2 nc τ, 2nC 2 κτ } successful SmoSteps. Proof Sketch The convergence s a consequence from theorem 18 from (Bordes et al., 2005). To apply ths theorem, we must prove that the drectons defned by (9) form a wtness famly for the polytope defned by the constrants of problem (7). Ths s the case because ths polytope s a product of n polytopes for whch we can apply proposton 7 from (Bordes et al., 2005). The number of teratons s then bounded usng a technque smlar to that of (Tsochantards et al., 2005). The complete proof wll be gven n an extended verson of ths paper. The bound on the number of teratons s also a bound on the number of support vectors. It s lnear n the number of examples and does not depend on the possbly large number of classes Stoppng Algorthm 5 does not specfy a crteron for stoppng ts outer loop. Excellent results are obtaned by performng just one or two outer loop teratons (epochs). We use the name LaRank 1 to ndcate that we perform a sngle epoch, that s to say, a sngle pass over the randomly ordered tranng examples. Other stoppng crtera nclude explotng the dualty gap (Schölkopf & Smola, 2002, ) and montorng the performance measured on a valdaton set. We use the name LaRankGap to ndcate that we terate algorthm 5 untl the dfference between the prmal cost (3) and the dual cost (7) becomes smaller than C. However, computng the dualty gap can become qute expensve. 4. Experments Ths secton report experments carred out on varous multclass pattern recognton problems. Although our approach s partly motvated by structured output problems, ths work focuses on well understood multclass tasks n order best characterze the algorthm behavor Expermental Setup Experments were carred out on four datasets brefly descrbed n table 1. The LETTER and USPS datasets are avalable from the UCI repostory. 1 The MNIST dataset 2 s a well known handwrtten dgt recognton benchmark. The INE dataset contans scentfc artcles from 18 journals and proceedngs of the IEEE. We use a flat TF/IDF feature space (see Denoyer & Gallnar, 2006 for further detals). Table 1 also lsts our choces for the parameter C and for the kernels k(x, x). These choces were made on the bass of past experence. We use the same parameters for all algorthms because we mostly compare algorthms that optmze the same crteron. The kernel cache sze was 500MB for all experments Comparng Optmzers Table 2 (top half) compares three optmzaton algorthms for the same dual cost (7). MCSVM (Crammer & Snger, 2001) uses the full gradent and therefore cannot be easly extended to handle structured output problems. We have used the MCSVM mplementaton dstrbuted by the authors

6 Solvng MultClass Support Vector Machnes wth LaRank Table 1. Datasets used for the experments. Tran Ex. Test Ex. Classes Features C k(x, x) LETTER e x x 2 USPS e x x 2 MNIST e x x 2 INE x x Table 2. Compared test error rates and tranng tmes. LETTER USPS MNIST INE MCSVM Test error (%) (stores the full gradent) Dual Tranng tme (sec.) Kernels ( 10 6 ) SVMstruct Test error (%) (stores partal gradent) Dual Tranng tme (sec.) Kernels ( 10 6 ) n/a LaRankGap Test error (%) (stores partal gradent) Dual Tranng tme (sec.) Kernels ( 10 6 ) LaRank 1 Test error (%) (onlne) Dual Tranng tme (sec.) Kernels ( 10 6 ) Not applcable because SVMstruct bypasses the cache when usng lnear kernels. LETTER USPS MNIST INE Fgure 1. Evoluton of the test error as a functon of the number of kernel calculatons

7 Solvng MultClass Support Vector Machnes wth LaRank SVMstruct (Tsochantards et al., 2005) targets structured output problems and therefore uses only a small fracton of the gradent. We have used the mplementaton dstrbuted by the authors. The authors warn that ths mplementaton has not been thoroughly optmzed. LaRankGap terates algorthm 5 untl the dualty gap becomes smaller than parameter C. Ths algorthm only stores a small fracton of the gradent, comparable to that used by SVMstruct. We have mplemented LaRank usng an nterpreted scrptng language wth a specalzed C functon for algorthm 1 (SmoStep). Both SVMstruct and LaRankGap use small subsets of the gradent coeffcents. Although these subsets have smlar sze, LaRankGap avods the tranng tme penalty experenced by SVMstruct. Both SVMstruct and LaRank make heavy use of kernel values nvolvng two support patterns. In contrast, MCSVM updates the complete gradent vector after each step and therefore uses the kernel matrx rows correspondng to support patterns. On our relatvely small problems, ths stronger memory requrement s more than compensated by the lower overhead of MCSVM s smpler cache structure Comparng Onlne Learnng Algorthms Table 2 (bottom half) also reports the results obtaned wth a sngle LaRank epoch (LaRank 1). Ths sngle pass over the tranng examples s suffcent to nearly reach the optmal performance. Ths result s understandable because () onlne perceptrons offer strong theoretcal guarantees after a sngle pass over the tranng examples, and () LaRank drves the optmzaton process by replcatng the randomzaton that happens n the perceptron. For each dataset, fgure 1 shows the evoluton of the test error wth respect to the number of kernel calculatons. The pont marked LaRank 1 corresponds to runnng a sngle LaRank epoch. The pont marked LaRankGap corresponds to usng the dualty gap stoppng crteron as explaned n secton 4.2. Fgure 1 also reports results obtaned wth two popular onlne algorthms: The ponts marked AvgPerceptron 1 and AvgPerceptron 10 respectvely correspond to performng one and ten epochs of the average perceptron algorthm (Freund & Schapre, 1998; Collns, 2002). Multple epochs of the averaged perceptron are very effectve when the necessary kernel values ft n the cache (frst row). Tranng tme ncreases consderably when ths s not the case (second row.) Fgure 2. Impact of the LaRank operatons (USPS dataset). The pont marked MIRA corresponds to the onlne multclass algorthm proposed by Crammer and Snger (2003). We have used the mplementaton provded by the authors as part of the MCSVM package. Ths algorthm computes more kernel values than AvgPerceptron 1 because ts soluton contans more support patterns. Its performance seems senstve to the choce of kernel: Crammer and Snger (2003) report substantally better results usng the same code but dfferent kernels. These results ndcate that performng sngle LaRank epoch s an attractve onlne learnng algorthm. Although LaRank 1 usually runs slower than AvgPerceptron 1 or MIRA, t provdes better and more predctable generalzaton performance Comparng Optmzaton Strateges Fgure 2 shows the error rates and kernel calculatons acheved when one restrcts the set of operatons chosen by algorthm 5. These results were obtaned after a sngle pass on the USPS dataset. As expected, usng only the ProcessNew operaton performs lke MIRA. The average perceptron requres sgnfcantly less kernel calculatons because ts soluton s much more sparse. However, t looses ths ntal sparsty when one performs several epochs (see fgure 1.) Enablng ProcessOld and Optmze sgnfcantly reduces the test error. The best test error s acheved when all operatons are enabled. The number of kernel calculatons s also reduced because ProcessOld and Optmze often elmnate support patterns Comparng ArgMax Calculatons The prevous experments measure the computatonal cost usng tranng tme and number of kernel calculatons. Certan structured output problems use costly

8 Solvng MultClass Support Vector Machnes wth LaRank Table 3. Numbers of arg max (n thousands). LETTER USPS MNIST INE AvgPerceptron AvgPerceptron LaRank LaRankGap SVMstruct algorthms to fnd the class wth the best score (1). The cost of ths arg max calculaton s partly related to the requred number of new kernel values. The average perceptron (and MIRA) performs one such arg max calculaton for each example t processes. In contrast, LaRank performs one arg max calculaton when processng a new example wth ProcessNew, and also when runnng ProcessOld. Table 3 compares the number of arg max calculatons for varous algorthms and datasets. 3 The SVMstruct optmzer performs very well wth ths metrc. The AvgPerceptron and LaRank are very compettve on a sngle epoch and become more costly when performng many epochs. One epoch s suffcent to reach good performance wth LaRank. Ths s not the case for the AvgPerceptron. 5. Concluson We have presented a large margn multclass algorthm that uses gradents as sparngly as SVMstruct wthout experencng the same tranng tme penalty. LaRank can be consdered an onlne algorthm because t nearly reaches ts optmal performance n a sngle pass over the tranng examples. Under these condtons, LaRank acheves test error rates that are compettve wth those of the full optmzaton, and sgnfcantly better than those acheved by perceptrons. Acknowledgments Ncolas Usuner helped provng the theorem bound. Part of ths work was funded by NSF grant CCR Antone Bordes was also supported by the DGA and by the Network of Excellence IST PASCAL. References Bakır, G., Hofmann, T., Schölkopf, B., Smola, A. J., Taskar, B., & Vshwanathan, S. V. N. (Eds.). (2007). Predctng structured outputs. MIT Press. n press. Bordes, A., & Bottou, L. (2005). The Huller: a smple and effcent onlne SVM. Machne Learnng: ECML 2005 (pp ). Sprnger Verlag. LNAI The letter results n table 3 are outlers because the letter kernel runs as fast as the kernel cache. Snce LaRank depends on tmngs, t often runs ProcessOld when a smple Optmze would have be suffcent. Bordes, A., Ertekn, S., Weston, J., & Bottou, L. (2005). Fast kernel classfers wth onlne and actve learnng. Journal of Machne Learnng Research, 6, Collns, M. (2002). Dscrmnatve tranng methods for hdden markov models: theory and experments wth perceptron algorthms. EMNLP 02: Proceedngs of the ACL-02 conference on Emprcal methods n natural language processng (pp. 1 8). Morrstown, NJ: Assocaton for Computatonal Lngustcs. Crammer, K., & Snger, Y. (2001). On the algorthmc mplementaton of multclass kernel-based vector machnes. Journal of Machne Learnng Research, 2, Crammer, K., & Snger, Y. (2003). Ultraconservatve onlne algorthms for multclass problems. Journal of Machne Learnng Research, 3, Denoyer, L., & Gallnar, P. (2006). The ML document mnng challenge. Advances n ML Informaton Retreval and Evaluaton, 5th Internatonal Workshop of the Intatve for the Evaluaton of ML Retreval, INE Schloß Dagsthul, Germany. Freund, Y., & Schapre, R. E. (1998). Large margn classfcaton usng the perceptron algorthm. Machne Learnng: Proceedngs of the Ffteenth Internatonal Conference. San Francsco, CA: Morgan Kaufmann. Graepel, T., Herbrch, R., & Wllamson, R. C. (2000). From margn to sparsty. In Advances n neural nformaton processng systems, vol. 13, MIT Press. Hldreth, C. (1957). A quadratc programmng procedure. Naval Research Logstcs Quarterly, 4, Erratum, bd. p361. Hsu, C.-W., & Ln, C.-J. (2002). A comparson of methods for mult-class support vector machnes. IEEE Transactons on Neural Networks, 13, LeCun, Y., Chopra, S., Hadsell, R., HuangFu, J., & Ranzato, M. (2007). A tutoral on energy-based learnng. In (Bakır et al., 2007), n press. Platt, J. (1999). Fast tranng of support vector machnes usng sequental mnmal optmzaton. Advances n Kernel Methods Support Vector Learnng (pp ). MIT Press. Rfkn, R. M., & Klautau, A. (2004). In defense of one-vsall classfcaton. Journal of Machne Learnng Research, 5, Schölkopf, B., & Smola, A. J. (2002). Learnng wth kernels. MIT Press. Taskar, B. (2004). Learnng structured predcton models: A large margn approach. Doctoral dssertaton, Stanford Unversty. Taskar, B., Chatalbashev, V., Koller, D., & Guestrn, C. (2005). Learnng structured predcton models: a large margn approach. Internatonal Conference on Machne Learnng (ICML) (pp ). Tsochantards, I., Joachms, T., Hofmann, T., & Altun, Y. (2005). Large margn methods for structured and nterdependent output varables. Journal of Machne Learnng Research, 6, Weston, J., & Watkns, C. (1998). Mult-class support vector machnes (Techncal Report CSD-TR-98-04). Department of Computer Scence, Royal Holloway, Unversty of London, Egham, UK.

Classification / Regression Support Vector Machines

Classification / Regression Support Vector Machines Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM

More information

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS46: Mnng Massve Datasets Jure Leskovec, Stanford Unversty http://cs46.stanford.edu /19/013 Jure Leskovec, Stanford CS46: Mnng Massve Datasets, http://cs46.stanford.edu Perceptron: y = sgn( x Ho to fnd

More information

Support Vector Machines. CS534 - Machine Learning

Support Vector Machines. CS534 - Machine Learning Support Vector Machnes CS534 - Machne Learnng Perceptron Revsted: Lnear Separators Bnar classfcaton can be veed as the task of separatng classes n feature space: b > 0 b 0 b < 0 f() sgn( b) Lnear Separators

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 48 CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 3.1 INTRODUCTION The raw mcroarray data s bascally an mage wth dfferent colors ndcatng hybrdzaton (Xue

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law) Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes

More information

5 The Primal-Dual Method

5 The Primal-Dual Method 5 The Prmal-Dual Method Orgnally desgned as a method for solvng lnear programs, where t reduces weghted optmzaton problems to smpler combnatoral ones, the prmal-dual method (PDM) has receved much attenton

More information

Relevance Feedback Document Retrieval using Non-Relevant Documents

Relevance Feedback Document Retrieval using Non-Relevant Documents Relevance Feedback Document Retreval usng Non-Relevant Documents TAKASHI ONODA, HIROSHI MURATA and SEIJI YAMADA Ths paper reports a new document retreval method usng non-relevant documents. From a large

More information

Solving two-person zero-sum game by Matlab

Solving two-person zero-sum game by Matlab Appled Mechancs and Materals Onlne: 2011-02-02 ISSN: 1662-7482, Vols. 50-51, pp 262-265 do:10.4028/www.scentfc.net/amm.50-51.262 2011 Trans Tech Publcatons, Swtzerland Solvng two-person zero-sum game by

More information

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour 6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the

More information

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for

More information

The Codesign Challenge

The Codesign Challenge ECE 4530 Codesgn Challenge Fall 2007 Hardware/Software Codesgn The Codesgn Challenge Objectves In the codesgn challenge, your task s to accelerate a gven software reference mplementaton as fast as possble.

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

GSLM Operations Research II Fall 13/14

GSLM Operations Research II Fall 13/14 GSLM 58 Operatons Research II Fall /4 6. Separable Programmng Consder a general NLP mn f(x) s.t. g j (x) b j j =. m. Defnton 6.. The NLP s a separable program f ts objectve functon and all constrants are

More information

Taxonomy of Large Margin Principle Algorithms for Ordinal Regression Problems

Taxonomy of Large Margin Principle Algorithms for Ordinal Regression Problems Taxonomy of Large Margn Prncple Algorthms for Ordnal Regresson Problems Amnon Shashua Computer Scence Department Stanford Unversty Stanford, CA 94305 emal: shashua@cs.stanford.edu Anat Levn School of Computer

More information

Abstract Ths paper ponts out an mportant source of necency n Smola and Scholkopf's Sequental Mnmal Optmzaton (SMO) algorthm for SVM regresson that s c

Abstract Ths paper ponts out an mportant source of necency n Smola and Scholkopf's Sequental Mnmal Optmzaton (SMO) algorthm for SVM regresson that s c Improvements to SMO Algorthm for SVM Regresson 1 S.K. Shevade S.S. Keerth C. Bhattacharyya & K.R.K. Murthy shrsh@csa.sc.ernet.n mpessk@guppy.mpe.nus.edu.sg cbchru@csa.sc.ernet.n murthy@csa.sc.ernet.n 1

More information

2x x l. Module 3: Element Properties Lecture 4: Lagrange and Serendipity Elements

2x x l. Module 3: Element Properties Lecture 4: Lagrange and Serendipity Elements Module 3: Element Propertes Lecture : Lagrange and Serendpty Elements 5 In last lecture note, the nterpolaton functons are derved on the bass of assumed polynomal from Pascal s trangle for the fled varable.

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1)

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1) Secton 1.2 Subsets and the Boolean operatons on sets If every element of the set A s an element of the set B, we say that A s a subset of B, or that A s contaned n B, or that B contans A, and we wrte A

More information

Efficient Distributed Linear Classification Algorithms via the Alternating Direction Method of Multipliers

Efficient Distributed Linear Classification Algorithms via the Alternating Direction Method of Multipliers Effcent Dstrbuted Lnear Classfcaton Algorthms va the Alternatng Drecton Method of Multplers Caoxe Zhang Honglak Lee Kang G. Shn Department of EECS Unversty of Mchgan Ann Arbor, MI 48109, USA caoxezh@umch.edu

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

Collaboratively Regularized Nearest Points for Set Based Recognition

Collaboratively Regularized Nearest Points for Set Based Recognition Academc Center for Computng and Meda Studes, Kyoto Unversty Collaboratvely Regularzed Nearest Ponts for Set Based Recognton Yang Wu, Mchhko Mnoh, Masayuk Mukunok Kyoto Unversty 9/1/013 BMVC 013 @ Brstol,

More information

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization Problem efntons and Evaluaton Crtera for Computatonal Expensve Optmzaton B. Lu 1, Q. Chen and Q. Zhang 3, J. J. Lang 4, P. N. Suganthan, B. Y. Qu 6 1 epartment of Computng, Glyndwr Unversty, UK Faclty

More information

X- Chart Using ANOM Approach

X- Chart Using ANOM Approach ISSN 1684-8403 Journal of Statstcs Volume 17, 010, pp. 3-3 Abstract X- Chart Usng ANOM Approach Gullapall Chakravarth 1 and Chaluvad Venkateswara Rao Control lmts for ndvdual measurements (X) chart are

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

Optimization Methods: Integer Programming Integer Linear Programming 1. Module 7 Lecture Notes 1. Integer Linear Programming

Optimization Methods: Integer Programming Integer Linear Programming 1. Module 7 Lecture Notes 1. Integer Linear Programming Optzaton Methods: Integer Prograng Integer Lnear Prograng Module Lecture Notes Integer Lnear Prograng Introducton In all the prevous lectures n lnear prograng dscussed so far, the desgn varables consdered

More information

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET 1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

Random Kernel Perceptron on ATTiny2313 Microcontroller

Random Kernel Perceptron on ATTiny2313 Microcontroller Random Kernel Perceptron on ATTny233 Mcrocontroller Nemanja Djurc Department of Computer and Informaton Scences, Temple Unversty Phladelpha, PA 922, USA nemanja.djurc@temple.edu Slobodan Vucetc Department

More information

SVM-based Learning for Multiple Model Estimation

SVM-based Learning for Multiple Model Estimation SVM-based Learnng for Multple Model Estmaton Vladmr Cherkassky and Yunqan Ma Department of Electrcal and Computer Engneerng Unversty of Mnnesota Mnneapols, MN 55455 {cherkass,myq}@ece.umn.edu Abstract:

More information

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices Internatonal Mathematcal Forum, Vol 7, 2012, no 52, 2549-2554 An Applcaton of the Dulmage-Mendelsohn Decomposton to Sparse Null Space Bases of Full Row Rank Matrces Mostafa Khorramzadeh Department of Mathematcal

More information

Meta-heuristics for Multidimensional Knapsack Problems

Meta-heuristics for Multidimensional Knapsack Problems 2012 4th Internatonal Conference on Computer Research and Development IPCSIT vol.39 (2012) (2012) IACSIT Press, Sngapore Meta-heurstcs for Multdmensonal Knapsack Problems Zhbao Man + Computer Scence Department,

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

Discriminative Dictionary Learning with Pairwise Constraints

Discriminative Dictionary Learning with Pairwise Constraints Dscrmnatve Dctonary Learnng wth Parwse Constrants Humn Guo Zhuoln Jang LARRY S. DAVIS UNIVERSITY OF MARYLAND Nov. 6 th, Outlne Introducton/motvaton Dctonary Learnng Dscrmnatve Dctonary Learnng wth Parwse

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 An Iteratve Soluton Approach to Process Plant Layout usng Mxed

More information

Biostatistics 615/815

Biostatistics 615/815 The E-M Algorthm Bostatstcs 615/815 Lecture 17 Last Lecture: The Smplex Method General method for optmzaton Makes few assumptons about functon Crawls towards mnmum Some recommendatons Multple startng ponts

More information

User Authentication Based On Behavioral Mouse Dynamics Biometrics

User Authentication Based On Behavioral Mouse Dynamics Biometrics User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA

More information

Mathematics 256 a course in differential equations for engineering students

Mathematics 256 a course in differential equations for engineering students Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the

More information

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints Australan Journal of Basc and Appled Scences, 2(4): 1204-1208, 2008 ISSN 1991-8178 Sum of Lnear and Fractonal Multobjectve Programmng Problem under Fuzzy Rules Constrants 1 2 Sanjay Jan and Kalash Lachhwan

More information

LECTURE NOTES Duality Theory, Sensitivity Analysis, and Parametric Programming

LECTURE NOTES Duality Theory, Sensitivity Analysis, and Parametric Programming CEE 60 Davd Rosenberg p. LECTURE NOTES Dualty Theory, Senstvty Analyss, and Parametrc Programmng Learnng Objectves. Revew the prmal LP model formulaton 2. Formulate the Dual Problem of an LP problem (TUES)

More information

Parallel matrix-vector multiplication

Parallel matrix-vector multiplication Appendx A Parallel matrx-vector multplcaton The reduced transton matrx of the three-dmensonal cage model for gel electrophoress, descrbed n secton 3.2, becomes excessvely large for polymer lengths more

More information

GenSVM: A Generalized Multiclass Support Vector Machine

GenSVM: A Generalized Multiclass Support Vector Machine Journal of Machne Learnng Research 17 (2016) 1-42 Submtted 12/14; Revsed 11/16; Publshed 12/16 GenSVM: A Generalzed Multclass Support Vector Machne Gerrt J.J. van den Burg Patrck J.F. Groenen Econometrc

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS ARPN Journal of Engneerng and Appled Scences 006-017 Asan Research Publshng Network (ARPN). All rghts reserved. NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS Igor Grgoryev, Svetlana

More information

Private Information Retrieval (PIR)

Private Information Retrieval (PIR) 2 Levente Buttyán Problem formulaton Alce wants to obtan nformaton from a database, but she does not want the database to learn whch nformaton she wanted e.g., Alce s an nvestor queryng a stock-market

More information

A New Approach For the Ranking of Fuzzy Sets With Different Heights

A New Approach For the Ranking of Fuzzy Sets With Different Heights New pproach For the ankng of Fuzzy Sets Wth Dfferent Heghts Pushpnder Sngh School of Mathematcs Computer pplcatons Thapar Unversty, Patala-7 00 Inda pushpndersnl@gmalcom STCT ankng of fuzzy sets plays

More information

Load Balancing for Hex-Cell Interconnection Network

Load Balancing for Hex-Cell Interconnection Network Int. J. Communcatons, Network and System Scences,,, - Publshed Onlne Aprl n ScRes. http://www.scrp.org/journal/jcns http://dx.do.org/./jcns.. Load Balancng for Hex-Cell Interconnecton Network Saher Manaseer,

More information

A Notable Swarm Approach to Evolve Neural Network for Classification in Data Mining

A Notable Swarm Approach to Evolve Neural Network for Classification in Data Mining A Notable Swarm Approach to Evolve Neural Network for Classfcaton n Data Mnng Satchdananda Dehur 1, Bjan Bhar Mshra 2 and Sung-Bae Cho 1 1 Soft Computng Laboratory, Department of Computer Scence, Yonse

More information

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

Wishing you all a Total Quality New Year!

Wishing you all a Total Quality New Year! Total Qualty Management and Sx Sgma Post Graduate Program 214-15 Sesson 4 Vnay Kumar Kalakband Assstant Professor Operatons & Systems Area 1 Wshng you all a Total Qualty New Year! Hope you acheve Sx sgma

More information

Intelligent Information Acquisition for Improved Clustering

Intelligent Information Acquisition for Improved Clustering Intellgent Informaton Acquston for Improved Clusterng Duy Vu Unversty of Texas at Austn duyvu@cs.utexas.edu Mkhal Blenko Mcrosoft Research mblenko@mcrosoft.com Prem Melvlle IBM T.J. Watson Research Center

More information

Wavefront Reconstructor

Wavefront Reconstructor A Dstrbuted Smplex B-Splne Based Wavefront Reconstructor Coen de Vsser and Mchel Verhaegen 14-12-201212 2012 Delft Unversty of Technology Contents Introducton Wavefront reconstructon usng Smplex B-Splnes

More information

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr)

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr) Helsnk Unversty Of Technology, Systems Analyss Laboratory Mat-2.08 Independent research projects n appled mathematcs (3 cr) "! #$&% Antt Laukkanen 506 R ajlaukka@cc.hut.f 2 Introducton...3 2 Multattrbute

More information

Efficient Text Classification by Weighted Proximal SVM *

Efficient Text Classification by Weighted Proximal SVM * Effcent ext Classfcaton by Weghted Proxmal SVM * Dong Zhuang 1, Benyu Zhang, Qang Yang 3, Jun Yan 4, Zheng Chen, Yng Chen 1 1 Computer Scence and Engneerng, Bejng Insttute of echnology, Bejng 100081, Chna

More information

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 A mathematcal programmng approach to the analyss, desgn and

More information

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms

Course Introduction. Algorithm 8/31/2017. COSC 320 Advanced Data Structures and Algorithms. COSC 320 Advanced Data Structures and Algorithms Course Introducton Course Topcs Exams, abs, Proects A quc loo at a few algorthms 1 Advanced Data Structures and Algorthms Descrpton: We are gong to dscuss algorthm complexty analyss, algorthm desgn technques

More information

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,

More information

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

Learning to Project in Multi-Objective Binary Linear Programming

Learning to Project in Multi-Objective Binary Linear Programming Learnng to Project n Mult-Objectve Bnary Lnear Programmng Alvaro Serra-Altamranda Department of Industral and Management System Engneerng, Unversty of South Florda, Tampa, FL, 33620 USA, amserra@mal.usf.edu,

More information

Edge Detection in Noisy Images Using the Support Vector Machines

Edge Detection in Noisy Images Using the Support Vector Machines Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona

More information

Learning-Based Top-N Selection Query Evaluation over Relational Databases

Learning-Based Top-N Selection Query Evaluation over Relational Databases Learnng-Based Top-N Selecton Query Evaluaton over Relatonal Databases Lang Zhu *, Wey Meng ** * School of Mathematcs and Computer Scence, Hebe Unversty, Baodng, Hebe 071002, Chna, zhu@mal.hbu.edu.cn **

More information

CMPS 10 Introduction to Computer Science Lecture Notes

CMPS 10 Introduction to Computer Science Lecture Notes CPS 0 Introducton to Computer Scence Lecture Notes Chapter : Algorthm Desgn How should we present algorthms? Natural languages lke Englsh, Spansh, or French whch are rch n nterpretaton and meanng are not

More information

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

Sequential search. Building Java Programs Chapter 13. Sequential search. Sequential search

Sequential search. Building Java Programs Chapter 13. Sequential search. Sequential search Sequental search Buldng Java Programs Chapter 13 Searchng and Sortng sequental search: Locates a target value n an array/lst by examnng each element from start to fnsh. How many elements wll t need to

More information

Hermite Splines in Lie Groups as Products of Geodesics

Hermite Splines in Lie Groups as Products of Geodesics Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the

More information

AMath 483/583 Lecture 21 May 13, Notes: Notes: Jacobi iteration. Notes: Jacobi with OpenMP coarse grain

AMath 483/583 Lecture 21 May 13, Notes: Notes: Jacobi iteration. Notes: Jacobi with OpenMP coarse grain AMath 483/583 Lecture 21 May 13, 2011 Today: OpenMP and MPI versons of Jacob teraton Gauss-Sedel and SOR teratve methods Next week: More MPI Debuggng and totalvew GPU computng Read: Class notes and references

More information

Alternating Direction Method of Multipliers Implementation Using Apache Spark

Alternating Direction Method of Multipliers Implementation Using Apache Spark Alternatng Drecton Method of Multplers Implementaton Usng Apache Spark Deterch Lawson June 4, 2014 1 Introducton Many applcaton areas n optmzaton have benefted from recent trends towards massve datasets.

More information

Fuzzy Modeling of the Complexity vs. Accuracy Trade-off in a Sequential Two-Stage Multi-Classifier System

Fuzzy Modeling of the Complexity vs. Accuracy Trade-off in a Sequential Two-Stage Multi-Classifier System Fuzzy Modelng of the Complexty vs. Accuracy Trade-off n a Sequental Two-Stage Mult-Classfer System MARK LAST 1 Department of Informaton Systems Engneerng Ben-Guron Unversty of the Negev Beer-Sheva 84105

More information

Active Contours/Snakes

Active Contours/Snakes Actve Contours/Snakes Erkut Erdem Acknowledgement: The sldes are adapted from the sldes prepared by K. Grauman of Unversty of Texas at Austn Fttng: Edges vs. boundares Edges useful sgnal to ndcate occludng

More information

Backpropagation: In Search of Performance Parameters

Backpropagation: In Search of Performance Parameters Bacpropagaton: In Search of Performance Parameters ANIL KUMAR ENUMULAPALLY, LINGGUO BU, and KHOSROW KAIKHAH, Ph.D. Computer Scence Department Texas State Unversty-San Marcos San Marcos, TX-78666 USA ae049@txstate.edu,

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

High-Boost Mesh Filtering for 3-D Shape Enhancement

High-Boost Mesh Filtering for 3-D Shape Enhancement Hgh-Boost Mesh Flterng for 3-D Shape Enhancement Hrokazu Yagou Λ Alexander Belyaev y Damng We z Λ y z ; ; Shape Modelng Laboratory, Unversty of Azu, Azu-Wakamatsu 965-8580 Japan y Computer Graphcs Group,

More information

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data A Fast Content-Based Multmeda Retreval Technque Usng Compressed Data Borko Furht and Pornvt Saksobhavvat NSF Multmeda Laboratory Florda Atlantc Unversty, Boca Raton, Florda 3343 ABSTRACT In ths paper,

More information

A Saturation Binary Neural Network for Crossbar Switching Problem

A Saturation Binary Neural Network for Crossbar Switching Problem A Saturaton Bnary Neural Network for Crossbar Swtchng Problem Cu Zhang 1, L-Qng Zhao 2, and Rong-Long Wang 2 1 Department of Autocontrol, Laonng Insttute of Scence and Technology, Benx, Chna bxlkyzhangcu@163.com

More information

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces Range mages For many structured lght scanners, the range data forms a hghly regular pattern known as a range mage. he samplng pattern s determned by the specfc scanner. Range mage regstraton 1 Examples

More information

Fixing Max-Product: Convergent Message Passing Algorithms for MAP LP-Relaxations

Fixing Max-Product: Convergent Message Passing Algorithms for MAP LP-Relaxations Fxng Max-Product: Convergent Message Passng Algorthms for MAP LP-Relaxatons Amr Globerson Tomm Jaakkola Computer Scence and Artfcal Intellgence Laboratory Massachusetts Insttute of Technology Cambrdge,

More information

Exercises (Part 4) Introduction to R UCLA/CCPR. John Fox, February 2005

Exercises (Part 4) Introduction to R UCLA/CCPR. John Fox, February 2005 Exercses (Part 4) Introducton to R UCLA/CCPR John Fox, February 2005 1. A challengng problem: Iterated weghted least squares (IWLS) s a standard method of fttng generalzed lnear models to data. As descrbed

More information

APPLICATION OF MULTIVARIATE LOSS FUNCTION FOR ASSESSMENT OF THE QUALITY OF TECHNOLOGICAL PROCESS MANAGEMENT

APPLICATION OF MULTIVARIATE LOSS FUNCTION FOR ASSESSMENT OF THE QUALITY OF TECHNOLOGICAL PROCESS MANAGEMENT 3. - 5. 5., Brno, Czech Republc, EU APPLICATION OF MULTIVARIATE LOSS FUNCTION FOR ASSESSMENT OF THE QUALITY OF TECHNOLOGICAL PROCESS MANAGEMENT Abstract Josef TOŠENOVSKÝ ) Lenka MONSPORTOVÁ ) Flp TOŠENOVSKÝ

More information

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems A Unfed Framework for Semantcs and Feature Based Relevance Feedback n Image Retreval Systems Ye Lu *, Chunhu Hu 2, Xngquan Zhu 3*, HongJang Zhang 2, Qang Yang * School of Computng Scence Smon Fraser Unversty

More information

Reducing Frame Rate for Object Tracking

Reducing Frame Rate for Object Tracking Reducng Frame Rate for Object Trackng Pavel Korshunov 1 and We Tsang Oo 2 1 Natonal Unversty of Sngapore, Sngapore 11977, pavelkor@comp.nus.edu.sg 2 Natonal Unversty of Sngapore, Sngapore 11977, oowt@comp.nus.edu.sg

More information

Multi-objective Optimization Using Adaptive Explicit Non-Dominated Region Sampling

Multi-objective Optimization Using Adaptive Explicit Non-Dominated Region Sampling 11 th World Congress on Structural and Multdscplnary Optmsaton 07 th -12 th, June 2015, Sydney Australa Mult-objectve Optmzaton Usng Adaptve Explct Non-Domnated Regon Samplng Anrban Basudhar Lvermore Software

More information

A Robust LS-SVM Regression

A Robust LS-SVM Regression PROCEEDIGS OF WORLD ACADEMY OF SCIECE, EGIEERIG AD ECHOLOGY VOLUME 7 AUGUS 5 ISS 37- A Robust LS-SVM Regresson József Valyon, and Gábor Horváth Abstract In comparson to the orgnal SVM, whch nvolves a quadratc

More information

Human Face Recognition Using Generalized. Kernel Fisher Discriminant

Human Face Recognition Using Generalized. Kernel Fisher Discriminant Human Face Recognton Usng Generalzed Kernel Fsher Dscrmnant ng-yu Sun,2 De-Shuang Huang Ln Guo. Insttute of Intellgent Machnes, Chnese Academy of Scences, P.O.ox 30, Hefe, Anhu, Chna. 2. Department of

More information

EVALUATION OF THE PERFORMANCES OF ARTIFICIAL BEE COLONY AND INVASIVE WEED OPTIMIZATION ALGORITHMS ON THE MODIFIED BENCHMARK FUNCTIONS

EVALUATION OF THE PERFORMANCES OF ARTIFICIAL BEE COLONY AND INVASIVE WEED OPTIMIZATION ALGORITHMS ON THE MODIFIED BENCHMARK FUNCTIONS Academc Research Internatonal ISS-L: 3-9553, ISS: 3-9944 Vol., o. 3, May 0 EVALUATIO OF THE PERFORMACES OF ARTIFICIAL BEE COLOY AD IVASIVE WEED OPTIMIZATIO ALGORITHMS O THE MODIFIED BECHMARK FUCTIOS Dlay

More information

Tighter Perceptron with Improved Dual Use of Cached Data for Model Representation and Validation

Tighter Perceptron with Improved Dual Use of Cached Data for Model Representation and Validation Proceedngs of Internatonal Jont Conference on Neural Networks, Atlanta, Georga, USA, June 49, 29 Tghter Perceptron wth Improved Dual Use of Cached Data for Model Representaton and Valdaton Zhuang Wang

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

Learning a Class-Specific Dictionary for Facial Expression Recognition

Learning a Class-Specific Dictionary for Facial Expression Recognition BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 16, No 4 Sofa 016 Prnt ISSN: 1311-970; Onlne ISSN: 1314-4081 DOI: 10.1515/cat-016-0067 Learnng a Class-Specfc Dctonary for

More information