Support Vector Machines for Business Applications

Size: px
Start display at page:

Download "Support Vector Machines for Business Applications"

Transcription

1 Support Vector Machnes for Busness Applcatons Bran C. Lovell and Chrstan J Walder The Unversty of Queensland and Max Planck Insttute, Tübngen {lovell, walder}@tee.uq.edu.au Introducton Recent years have seen an explosve growth n computng power and data storage wthn busness organsatons. From a busness perspectve, ths means that most companes now have massve archves of customer and product data and more often than not these archves are far too large for human analyss. An obvous queston has therefore arsen, How can one turn these mmense corporate data archves to commercal advantage? To ths end, a number of common applcatons have arsen, from predctng whch products a customer s most lkely to purchase, to desgnng the perfect product based on responses to questonnares. The theory and development of these processes has grown nto a dscplne of ts own, known as Data Mnng, whch draws heavly on the related felds of Machne Learnng, Pattern Recognton, and Mathematcal Statstcs. The Data Mnng dscplne s stll developng however, and a great deal of suboptmal and ad hoc analyss s beng done. Ths s partly due to the complexty of the problems, but s also due to the vast number of avalable technques. Even the most fundamental task n Data Mnng that of nductve nference, or makng predctons based on examples, can be tackled by a great many dfferent technques. Some of these technques are very dffcult to talor to a specfc problem and requre hghly sklled human desgn; others are more generc n applcaton and can be treated more lke the proverbal black box. One partcularly generc and powerful method, known as the Support Vector Machne (SVM) has proven to be both easy to apply and capable of producng results that range from good to excellent n comparson to other methods. Whle applcaton of the method s relatvely straghtforward, the practtoner can stll beneft greatly from a basc understandng of the underlyng machnery. Unfortunately most avalable tutorals on SVMs do requre a very sold mathematcal background, so we have wrtten ths chapter to make SVMs accessble to a wder communty. Ths chapter comprses a basc background on the problem of nducton, followed by the man sectons. In the frst secton we ntroduce the concepts and equatons on whch the SVM s based n an ntutve manner and to dentfy the relatonshp between the SVM and some of the other popular analyss methods. In the second secton we survey some nterestng applcatons of SVMs on practcal real world problems. Fnally, the thrd secton provdes a set of gudelnes and rules of thumb for applyng the tool, wth a pedagogcal example that s desgned to demonstrate everythng that the SVM newcomer requres n order to mmedately apply the tool to a specfc problem doman. The chapter s ntended as a bref ntroducton to the feld that ntroduces the deas, methodologes, as well as a hand-on ntroducton to freely avalable software, allowng the reader to rapdly determne the effectveness of SVMs for ther specfc doman. Background SVMs are most commonly appled to the problem of nductve nference, or makng predctons based on prevously seen examples. To llustrate what s meant by ths, let us consder the data presented n Tables and 2. We see here an example of the problem of nductve nference, more specfcally that of supervsed learnng. In supervsed learnng we are gven a set of nput data along wth ther correspondng labels. The nput data comprses a number of examples about whch several attrbutes are known (n ths case, age, ncome etc). The label ndcates whch class a partcular example belongs to. In the example above, the label tells us whether or not a gven person has broadband nternet connecton to ther home. Ths s called a bnary classfcaton problem because there are only two possble classes. In the second table, we are gven the attrbutes for a dfferent set of consumers, for whom the true class labels are unknown. Our goal

2 s to nfer from the frst table the most lkely labels for the people n the second table, that s, whether or not they have a broadband nternet connecton to ther home. Age Income Years of Educaton Gender Broadband Home Internet Connecton? 30 $56,000 / yr 6 male Yes 50 $60,000 / yr 2 female Yes 6 $2,000 / yr male No 35 $30,000 / yr 2 male No Table : tranng or labelled set The dataset n Table contans demographc nformaton for four randomly selected people. These people were surveyed to determne whether or not they had a broadband home nternet connecton. Age Income Years of Educaton Gender Broadband Home Internet Connecton? 40 $48,000 / yr 7 male unknown 29 $60,000 / yr 8 female unknown Table 2: unlabelled set The dataset n Table 2 contans demographc nformaton for people who may or may not be good canddates for broadband nternet connecton advertsng. The queston arsng s, Whch of these people s lkely to have broadband nternet connecton at home? In the feld of data mnng, we often refer to these sets by the terms test set, tranng set, valdaton set, and so on, but there s some confuson n the lterature about the exact defntons of these terms. For ths reason we avod ths nomenclature, wth the excepton of the term tranng set. For our purposes, the tranng set shall be all that s gven to us n order to nfer some general correspondence between the nput data and labels. We wll refer to the set of data for whch we would lke to predct the labels as the unlabelled set. A schematc dagram for the above process s provded n Fgure. In the case of the SVM classfer (and most other learnng algorthms for that matter), there are a number of parameters whch must be chosen by the user. These parameters control varous aspects of the algorthm, and n order to yeld the best possble performance, t s necessary to make the rght choces. The process of choosng parameters that yeld good performance s often referred to as model selecton. In order to understand ths process, we have to consder what t s that we are amng for n terms of classfer performance. From the pont of vew of the practtoner, the hope s that the algorthm wll be able to make true predctons about unseen cases. Here the true values we are tryng to predct are the class labels of the unlabelled data. From ths perspectve t s natural to measure the performance of a classfer by the probablty of ts msclassfyng an unseen example. It s here that thngs become somewhat less straghtforward, however, due to the followng dlemma. In order to estmate the probablty of a msclassfcaton, we need to know the true underlyng probablty dstrbutons of the data that we are dealng wth. If we actually knew ths, however, we wouldn t have needed to perform nductve nference n the frst place! Indeed knowledge of the true probablty dstrbutons allows us to calculate the theoretcally best possble decson rule correspondng to the socalled Bayesan classfer (Duda et al 200). In recent years, a great deal of research effort has gone nto developng sophstcated theores that make statements about the probablty of a partcular classfer makng errors on new unlabelled cases these statements are typcally referred to as generalzaton bounds. It turns out however, that the research has a long way to go, and n practce one s usually forced to determne the parameters of the learnng algorthm by much more pragmatc means. Perhaps the most straghtforward of these methods nvolves estmatng the probablty of msclassfcaton usng a set of real data for whch the class labels are known to do ths one smply compares the labels predcted by the learnng algorthm to the true known labels. The estmate of

3 msclassfcaton probablty s then gven by the number of examples for whch the algorthm made an error (that s, predcted a label other than the true known label) dvded by the number of examples whch were tested n ths manner. Labelled Tranng Data Learnng Algorthm (SVM) Unlabelled Data Decson Rule Predcted Labels Fgure : The nductve nference process n schematc form. Based on a partcular tranng set of examples wth labels, the learnng algorthm constructs a decson rule whch can then be used to predct the labels of new unlabelled examples. Some care needs to be taken however, n how ths procedure s conducted. A common ptfall for the nexperenced analyst nvolves makng ths estmate of msclassfcaton probablty usng the tranng set from whch the decson rule tself was nferred. The problem wth ths approach s easly seen from the followng smple decson rule example. Imagne a decson rule that makes label predctons by way of the followng procedure (sometmes referred to as the notebook classfer): The notebook classfer decson rule: We wsh to predct the label of the example X. If X s present n the tranng set, make the predcton that ts label s the same as the correspondng label n the tranng set. Otherwse, toss a con to determne the label. For ths method, whle the estmated probablty of msclassfcaton on the tranng set wll be zero, t s clear that for most real world problems the algorthm wll perform no better than tossng a con! The notebook classfer s a commonly used example to llustrate the phenomenon of overfttng whch refers to stuatons where the decson rule fts the tranng set well, but does not generalze well to prevously unseen cases. What we are really amng for s a decson rule that generalzes as well as possble, even f ths means that t cannot perform as well on the tranng set. Cross-valdaton: So t seems that we need a more sophstcated means of estmatng the generalzaton performance of our nferred decson rules, f we are to successfully gude the model selecton process. Fortunately there s a more effectve means of estmatng the generalzaton performance based on the tranng set. Ths procedure, whch s referred to as cross-valdaton or more specfcally n-fold cross valdaton, proceeds n the followng manner (Duda et al 200):. Splt the tranng set nto n equally szed and dsjont subsets (parttons), numbered to n. 2. Construct a decson functon usng a conglomerate of all the data from subsets 2 to n. 3. Use ths decson functon to predct the labels of the examples n subset number. 4. Compare the predcted labels to the known labels n subset number. 5. Repeat steps through 4 a further (n-) tmes, each tme testng on a dfferent subset, and always excludng that subset from tranng.

4 Havng done ths, we can once agan dvde the number of msclassfcatons by the total number of tranng examples to get an estmate of the true generalzaton performance. The pont s that snce we have avoded checkng the performance of the classfer on examples that the algorthm had already seen, we have calculated a far more meanngful measure of classfer qualty. Commonly used values for n are 3 and 0 leadng to so called 3-fold and 0-fold cross-valdaton. Now, whle t s nce to have some dea of how well our decson functon wll generalze, we really want to use ths measure to gude the model selecton process. If there are only, say, two parameters to choose for the classfcaton algorthm, t s common to smply evaluate the generalzaton performance (usng cross valdaton) for all combnatons of the two parameters, over some reasonable range. As the number of parameters ncreases, however, ths soon becomes nfeasble due to the excessve number of parameter combnatons. Fortunately one can often get away wth just two parameters for the SVM algorthm, makng ths relatvely straght-forward model selecton methodology wdely applcable and qute effectve on real world problems. Now that we have a basc understandng of what supervsed learnng algorthms can do, as well as roughly how they should be used and evaluated, t s tme to take a peek under the hood of one n partcular, the SVM. Whle the man underlyng dea of the SVM s qute ntutve, t wll be necessary to delve nto some mathematcal detals n order to better apprecate why the method has been so successful. Man Thrust of the Chapter The SVM s a supervsed learnng algorthm that nfers from a set of labeled examples a functon that takes new examples as nput, and produces predcted labels as output. As such the output of the algorthm s a mathematcal functon that s defned on the space from whch our examples are taken, and takes on one of two values at all ponts n the space, correspondng to the two class labels that are consdered n bnary classfcaton. One of the theoretcally appealng thngs about the SVM s that the key underlyng dea s n fact extremely smple. Indeed, the standard dervaton of the SVM algorthm begns wth possbly the smplest class of decson functons: lnear ones. To llustrate what s meant by ths, Fgure 2 conssts of three lnear decson functons that happen to be correctly classfyng some smple 2D tranng sets. Fgure 2: A smple 2D classfcaton task, to separate the black dots from the crcles. Three feasble but dfferent lnear decson functons are depcted, whereby the classfer predcts that any new samples n the gray regon are black dots, and those n the whte regon are crcles. Whch s the best decson functon and why? Lnear decson functons consst of a decson boundary that s a hyperplane (a lne n 2D, plane n 3D, etc) separatng the two dfferent regons of the space. Such a decson functon can be expressed by a mathematcal functon of an nput vector x, the value of whch s the predcted label for x (ether + or -). The lnear classfer can therefore be wrtten as g( x) = sgn( f ( x)) where f ( x) =< w, x > + b.

5 In ths way we have parameterzed the functon by the weght vector w and the scalar b. The notaton <w,x> denotes the nner or scalar product of w and x, defned by < w,x d >= w x where d s the dmensonalty, and w s the -th component of w, where w s of the form (w, w 2, w d ). Havng formalzed our decson functon, we can now formalze the problem whch the lnear SVM addresses: Gven a tranng set of vectors x, x 2, x n wth correspondng class membershp labels y, y 2, y n that take on the values + or -, choose parameters w and b of the lnear decson functon that generalzes well to unseen examples. Perceptron Algorthm: Probably the frst algorthm to tackle ths problem was the Perceptron algorthm (Rosenblatt 958). The Perceptron algorthm smply used an teratve procedure to ncrementally adjust w and b untl the decson boundary was able to separate the two classes of the tranng data. As such, the Perceptron algorthm would gve no preference between the three feasble solutons n Fgure 2 any one of the three could result. Ths seems rather unsatsfactory as most people would agree that the rghtmost decson functon s the superor one. Moreover, ths ntutve preference can be justfed n varous ways, for example by consderng the effect of measurement nose on the data small perturbatons of the data could easly change the predcted labels of the tranng set n the frst two examples, whereas the thrd s far more robust n ths respect. In order to make use of ths ntuton, t s necessary to state more precsely why we prefer the thrd classfer: We prefer decson boundares that not only correctly separate two classes n the tranng set, but le as far from the tranng examples as possble. Ths smple ntuton s all that s requred to lead to the lnear SVM classfer, whch chooses the hyperplane that separates the two classes wth the maxmum margn. The margn s just the dstance from the hyperplane to the nearest tranng example. Before we contnue, t s mportant to note that whle the above example shows a 2D dataset, whch can be convenently represented by ponts n a plane, n fact we wll typcally be dealng wth hgher dmensonal data. For example, the example data n Table could easly be represented as ponts n four dmensons as follows. = x = [ ] ; y = + x 2 = [ ] ; y 2 = + x 3 = [ ] ; y 3 = - x 4 = [ ] ; y 4 = - Actually, there are some desgn decsons to be made by the practtoner when translatng attrbutes nto the above type of numercal format, whch we shall touch on n the next secton. For example here we have mapped the male/female column nto two new numercal ndcators. For now, just note that we have also lsted the labels y to y 4 whch take on the value + or -, n order to ndcate the class membershp of the examples (that s, y = means that x has broadband home nternet connecton). In order to easly fnd the maxmum margn hyperplane for a gven data set usng a computer, we would lke to wrte the task as an optmzaton problem. Optmzaton problems consst of an objectve functon, whch we typcally want to fnd the maxmum or mnmum value of, along wth a set of constrants, whch are condtons that we must satsfy whle fndng the best value of the objectve functon. A smple example s to mnmze x 2 subject to the constrant that x 2. The soluton to ths example optmzaton problem happens to be x =. To see how to compactly formulate the maxmum margn hyperplane problem as an optmzaton problem, take a look at Fgure 3.

6 <w,x>+b= <w,x>+b=+ <w,x>+b=0 Fgure 3: Lnearly separable classfcaton problem The Fgure shows some 2D data drawn as crcles and black dots, havng labels + and respectvely. As before, we have parameterzed our decson functon by the vector w and the scalar b, whch means that, n order for our hyperplane to correctly separate the two classes, we need to satsfy the followng constrants: < w, x < w, x > + b > 0,for all y > + b < 0,for all y = = To ad understandng, the frst constrant above may be expressed as: <w, x > + b must be greater than zero, whenever y s equal to one. It s easy to check that the two sets of constrants above can be combned nto the followng sngle set of constrants: ( < w, x > + b) y > 0, =... n However meetng ths constrant s not enough to separate the two classes optmally we need to do so wth the maxmum margn. An easy way to see how to do ths s the followng. Frst note that we have plotted the decson surface as a sold lne n Fgure 3, whch s the set satsfyng: < w, x > + b = 0. () The set of constrants that we have so far s equvalent to sayng that these data must le on the correct sde (accordng to class label) of ths decson surface. Next notce that we have also plotted as dotted lnes two other hyperplanes, whch are the hyperplanes where the functon <w,x> + b s equal to - (on the lower left) and + (on the upper rght). Now, n order to fnd the maxmum margn hyperplane, we can see ntutvely that we should keep the dotted lnes parallel and equdstant to the decson surface, and maxmze ther dstance from one another, whle satsfyng the constrant that the data le on the correct sde of the dotted lnes assocated wth that class. In mathematcal form, the fnal clause of ths sentence (the constrants) can be wrtten as y ( < w, x > + b) >,... n. = All we need to do then s to maxmze the dstance between the dotted lnes subject to the constrant set above. To ad n understandng, one commonly used analogy s to thnk of these data ponts as nals partally drven nto a board. Now we successvely place thcker and thcker peces of tmber between the nals representng the two classes untl the tmber just fts the centrelne of the tmber now represents the

7 optmal decson boundary. It turns out that ths dstance s equal to 2 / < w, w >, and snce maxmzng 2 / < w, w > s the same as mnmzng < w, w >, we end up wth the followng optmzaton problem, the soluton of whch yelds the parameters of the maxmum margn hyperplane. The term ½ n the objectve functon below can be gnored as t smply makes thngs neater from a certan mathematcal pont of vew: mn w, b 2 w w such that y ( w x + b) for all =, 2,..., m The above problem s qute smple, but t encompasses the key phlosophy behnd the SVM maxmum margn data separaton. If the above problem had been scrbbled onto a cocktal napkn and handed to the poneers of the Perceptron back n the 960 s, then the Machne Learnng dscplne would probably have progressed a great deal further than t has to date! We cannot relax just yet however, as there s a major problem wth the above method: What f these data are not lnearly separable? That s f t s not possble to fnd a hyperplane that separates all of the examples n each class from all of the examples n the other class? In ths case there would be no combnaton of w and b that could ever satsfy the set of constrants above, let alone do so wth maxmum margn. Ths stuaton s depcted n Fgure 4, where t becomes apparent that we need to soften the constrant that these data le on the correct sde of the + and - hyperplanes, that s we need to allow some, but not too many data ponts to volate these constrants by a preferably small amount. Ths alternatve approach turns out to be very useful not only for datasets that are not lnearly separable, but also, and perhaps more mportantly, n allowng mprovements n generalzaton. <w,x>+b= <w,x>+b=+ <w,x>+b=0 Fgure 4: Lnearly nseparable classfcaton problem Usually when we start talkng about vague concepts such as not too many and a small amount, we need to ntroduce a parameter nto our problem, whch we can vary n order to balance between varous goals and objectves. The followng optmzaton problem, known as the -norm soft margn SVM, s probably the one most commonly used to balance the goals of maxmum margn separaton, and correctness of the tranng set classfcaton. It acheves varous trade-offs between these goals for varous values of the parameter C, whch s usually chosen by cross-valdaton on a tranng set as dscussed earler.

8 mn w, b, ξ 2 w w+ C m = such that y ( w x + b) + ξ (2) ξ for all =, 2,..., m. The easest way to understand ths problem s by comparson wth the prevous formulaton that we gave, whch s known as the hard margn SVM, n reference to the fact that the margn constrants are hard, and are not allowed to be volated at all. Frst note that we have an extra term n our objectve functon that s equal to the sum of the ξ s. Snce we are mnmzng the objectve functon, t s safe to say that we are lookng for a soluton that keeps the ξ values small. Moreover, snce the ξ term s added to the orgnal objectve functon after multplcaton by C, we can say that as C ncreases we care less about the sze of the margn, and more about keepng the ξ s small. The true meanng of the ξ s can only be seen from the constrant set, however. Here, nstead of constranng the functon y (<w,x > + b) to be greater than, we constran t to be greater than - ξ. That s, we allow the pont x to volate the margn by an amount ξ. Thus, the value of C trades between how large of a margn we would prefer, as opposed to how many of the tranng set examples volate ths margn (and by how much). So far, we have seen that the maxmally separatng hyperplane s a good startng pont for lnear classfers. We have also seen how to wrte down the problem of fndng ths hyperplane as an optmzaton problem consstng of an objectve functon and constrants. After ths we saw a way of dealng wth data that s not lnearly separable, by allowng some tranng ponts to volate the margn somewhat. The next lmtaton we wll address s n the form of solutons avalable. So far we have only consdered very smple lnear classfers, and as such we can only expect to succeed n very smple cases. Fortunately t s possble to extend the prevous analyss n an ntutve manner, to more complex classes of decson functons. The basc dea s llustrated n Fgure 5. Fgure 5: An example of a mappng Ф to a feature space n whch the data become lnearly separable. The example n Fgure 5 shows on the left a data set that s not lnearly separable. In fact, the data s not even close to lnearly separable, and one could never do very well wth a lnear classfer for the tranng set gven. In spte of ths, t s easy for a person to look at the data and suggest a smple ellptcal decson surface that ought to generalze well. Imagne however that there s a mappng Ф whch transforms these data to some new, possbly hgher dmensonal space, n whch the data s lnearly separable. If we knew Ф then we could map all of the data to the feature space, and perform normal SVM classfcaton n ths space. If we can acheve a reasonable margn n the feature space, then we can expect a reasonably good generalzaton performance, n spte of a possble ncrease n dmensonalty.

9 The last sentence of the prevous paragraph s far deeper than t may frst appear. For some tme, Machne Learnng researchers have feared the curse of dmensonalty, a name gven to the wdely-held belef that f the dmenson of the feature space s large n comparson to the number of tranng examples, then t s dffcult to fnd a classfer that generalzes well. It took the theory of Vapnk and Chervonenks (Vapnk 998) to put a serous dent n ths belef. In a nutshell, they formalzed and proved the last sentence of the prevous paragraph, and thereby paved the way for methods that map data to very hgh dmensonal feature spaces where they then perform maxmum margn lnear separaton. Actually, a trcky practcal ssue also had to be overcome before the approach could floursh: f we map to a feature space that s too hgh n dmenson, then t wll become mpossble to perform the requred calculatons (that s, to fnd w and b) that s, t would take too long on a computer. It s not obvous how to overcome ths dffculty, and t took untl 995 for researchers to notce the followng elegant and qute remarkable possblty. The usual way of proceedng s to take the orgnal soft margn SVM, and convert t to an equvalent Lagrangan dual problem. The dervaton s not especally enlghtenng however, so we wll skp to the result, whch s that the soluton to the followng dual or equvalent problem gves us the soluton to the orgnal SVM problem. The dual problem, whch s to be solved by varyng the α ι s, s as follows (Vapnk 998) m mn yyαα ( x x ) α 2 j j j, j= = such that y α = 0 = 0 α C, =,2,..., m. m m α (3) The α ι s are known as the dual varables, and they defne the correspondng prmal varables w and b by the followng relatonshps: m w = α = y α ( y ( < w, x > + b) ) = 0 Note that by the lnearty of the nner product (that s, the fact that <a+b,c> = <a,c> + <b,c>), we can wrte the decson functon n the followng form: f ( x) =< w, x > + b = α y < x, x > + b Recall that t s the sgn of f(x) that gves us the predcted label of x. A qute remarkable thng s that n order to determne the optmal values of the α ι s and b, and also to calculate f(x), we do not actually need to know any of the tranng or testng vectors, we only need to know the scalar value of ther nner product wth one another. Ths can be seen by notng that the vectors only ever appear by way of ther nner product wth one another. The elegant thng s that rather than explctly mappng all of the data to the new space and performng lnear SVM classfcaton, we can operate n the orgnal space, provded we can fnd a socalled kernel functon k(.,.) whch s equal to the nner product of the mapped data. That s, we need a kernel functon k(.,.) satsfyng: m = x k ( x, y) =< Φ( x), Φ( y) > In practce, the practtoner need not concern hm or herself wth the exact nature of the mappng Ф. In fact, t s usually more ntutve to concentrate on propertes of the kernel functons anyway, and the prevalng wsdom states that the functon k(x,y) should be a good measure of the smlarty of the vectors x and y. Moreover, not just any functon k can be used t must also satsfy certan techncal condtons, known as

10 Mercer s condtons. Ths procedure of mplctly mappng the data va the functon k s typcally often called the kernel trck and has found wde applcaton after beng popularzed by the success of the SVM (Schölkopf, & Smola 998). The two most wdely used kernel functons are the followng. Polynomal Kernel k(x, y) = ( < x, y > + ) d The polynomal kernel s vald for all postve ntegers d. The kernel corresponds to a mappng Ф that computes all degree d monomal terms of the ndvdual vector components of the orgnal space. The polynomal kernel has been used to great effect on dgt recognton problems. 2 x y Gaussan Kernel k(x, y) = exp( - ) 2 σ The Gaussan kernel, whch s smlar to the Gaussan probablty dstrbuton from whch t gets ts name, s one of a group of kernel functons known as radal bass functons (RBFs). RBFs are kernel functons that depend only on the geometrc dstance between x and y. The kernel s vald for all non-zero values of the kernel wdth σ, and corresponds to a mappng Ф nto an nfnte dmensonal, and therefore somewhat less nterpretable, feature space. Nonetheless, the Gaussan s probably the most useful and commonly used kernel functon. Now that we know the form of the SVM dual problem, as well as how to generalze t usng kernel functons, the only thng left s to see s how to actually solve the optmzaton problem, n order to fnd the α ι s. The optmzaton problem s one example of a class of problems known as Quadratc Programs (QPs). The term program, as t s used here, s somewhat antquated and n fact means a mathematcal optmzaton problem, not a computer program. Fortunately there are many computer programs that can solve QP s such as ths, these computer programs beng known as Quadratc Program (QP) solvers. An mportant factor to note here s that there s consderable structure n the QP that arses n SVM tranng, and whle t would be possble to use almost any QP solver on the problem, there are a number of sophstcated software packages talored to take advantage of ths structure, n order to decrease the requrements of computer tme and memory. One property of the SVM QP that can be taken advantage of s ts sparsty the fact that n many cases, at the optmal soluton most of the α ι s wll equal zero. It s nterestng to see what ths means n terms of the decson functon f(x): those vectors wth α ι = 0 do not actually enter nto the fnal form of the soluton. In fact, t can be shown that one can remove all of the correspondng tranng vectors before tranng even commences, and get the same fnal result. The vectors wth non-zero values of α ι are known as the Support Vectors, a term that has ts root n the theory of convex sets. As t turns out, the Support Vectors are the hard cases the tranng examples that are most dffcult to classfy correctly (and that le closest to the decson boundary). In our prevous practcal analogy, the support vectors are lterally the nals that support the block of wood! Now that we have an understandng of the machnery underlyng t, we wll soon proceed to solve a practcal problem usng the freely avalable SVM software package lbsvm wrtten by Hsu and Ln. Relatonshp to Other Methods We noted n the ntroducton that the SVM s an especally easy to use method that typcally produces good results even when treated as a processng black box. Ths s ndeed the case, and to better understand ths t s necessary to consder what s nvolved n usng some other methods. We wll focus n detal on the extremely prevalent class of algorthms known as artfcal neural networks, but frst we provde a bref overvew of some other related methods. Lnear Dscrmnant Analyss (Hand 98, Wess & Kulkowsk 99) s wdely used n busness and marketng applcatons, can work n multple dmensons, and s well-grounded n the mathematcal lterature. It nonetheless has two major drawbacks. The frst s that lnear dscrmnant functons, as the

11 name mples, can only successfully classfy lnearly separable data thus lmtng ther applcaton to relatvely smple problems. If we extend the method to hgher order functons such as quadratc dscrmnators, generalzaton suffers. Indeed such degradaton n performance wth ncreased numbers of parameters corroborated the belef n the curse of dmensonalty fnally dsproved by Vapnk (Vapnk, 998). The second problem s smply that generalzaton performance on real problems s usually sgnfcantly worse than ether decson trees or artfcal neural networks (e.g., see the comparsons n Wess & Kulkowsk 99). Decson Trees are commonly used n classfcaton problems wth categorcal data (Qunlan 993), although t s possble to derve categorcal data from ordnal data by ntroducng bnary valued features such as age s less than 20. Decson trees construct a tree of questons to be asked of a gven example n order to determne the class membershp by way of class labels assocated wth leaf nodes of the decson tree. Ths approach s smple and has the advantage that t produces decson rules that can be nterpreted by a human as well as a machne; however the SVM s more approprate for complex problems wth a many ordnal features. Nearest Neghbour methods are very smple and therefore sutable for extremely large data sets. These methods smply search the tranng data set for the k examples that are closest (by the crtera of Eucldean dstance for example) to the gven nput. The most common class label that assocated wth these k s then assgned to the gven query example. When the tranng and testng computaton tmes are not so mportant however, the dscrmnatve nature of the SVM wll usually yeld sgnfcantly mproved results. Artfcal Neural Network (ANN) algorthms have become extremely wdespread n the area of data mnng and pattern recognton (Bshop, 995). These methods were orgnally nspred by the neural connectons that comprse the human bran the basc dea beng that n the human bran many smple unts (neurons) are connected together n a manner that produces complex, powerful behavour. To smulate ths phenomenon, neurons are modeled by unts whose output y s related to the nput x by some actvaton functon g by the relatonshp y = g(x). These unts are then connected together n varous archtectures, whereby the output of a gven unt s multpled by some constant weght and then fed forward as nput to the next unt, possbly n summaton wth a smlarly scaled output from some other unt(s). Ultmately all of the nputs are fed to one sngle fnal unt, the output of whch s typcally compared to some threshold n order to produce a class membershp predcton. Ths s a very general framework that provdes many avenues for customsaton: Choce of actvaton functon Choce of network archtecture (number of unts and the manner n whch they are connected) Choce of the weghts by whch the output of a gven unt s multpled to produce the nput of another unt. Algorthm for determnng the weghts gven the tranng data. In comparson to the SVM, both the strength and weakness of the ANN les n t s flexblty typcally a consderable amount of expermentaton s requred n order to acheve good results, and moreover snce the optmzaton problems that are typcally used to fnd the weghts of the chosen network are non-convex, many numercal trcks are requred n order to fnd a good soluton to the problem. Nonetheless, gven suffcent skll and effort n engneerng a soluton wth an ANN, one can often talor the algorthm very specfcally to a gven problem n a process that s lkely to eventually yeld superor results to the SVM. Havng sad ths, there are cases, for example n handwrtten dgt recognton, n whch SVM performance s on par wth hghly engneered ANN solutons (DeCoste 2002). By way of comparson, the SVM approach s lkely to yeld a very good soluton wth far less effort than s requred for a good ANN soluton. Practcal Applcaton of the SVM As we have seen, the theoretcal underpnnngs of the SVM are very compellng, especally snce the algorthm nvolves very lttle tral and error, and s easy to apply. Nonetheless, the usefulness of the

12 algorthm can only be borne out by practcal experence, and so n ths sub-secton we survey a number of studes that use the SVM algorthm n practcal problems. Before we menton such specfc cases, we frst dentfy the general characterstcs of those problems to whch the SVM s partcularly well suted. One key consderaton s that n ts basc form the SVM has lmted capacty to deal wth large tranng data sets. Typcally the SVM can only handle problems of up to approxmately 00,000 tranng examples before approxmatons must be made n order to yeld reasonable tranng tmes. Havng sad ths, the tranng tmes depend only margnally on the dmensonalty of the features t s often sad that SVMs can often defy the so-called curse of dmensonalty the dffculty that often occurs when the dmensonalty s hgh n comparson wth the number of tranng samples. It should also be noted that, wth the excepton of the strng kernel case, the SVM s most naturally suted to ordnal features rather than categorcal ones, although as we shall see n the next Secton, t s possble to handle both cases. Before turnng to some specfc busness and marketng cases, t s mportant to note that some of the most successful applcatons of the SVM have been n mage processng n partcular handwrtten dgt recognton (DeCoste 2002) and face recognton (Osuna 997). In these areas, a common theme of the applcaton of SVMs s not so much ncreased accuracy, but rather a greatly smplfed desgn and mplementaton process. As such, when consderng popular areas such as face recognton, t s mportant to understand that very smple SVM mplementatons are often compettve wth the complex and hghly tuned systems that were developed over a longer perod pror to the advent of the SVM. Another nterestng applcaton area for SVMs s on strng data, for example n text mnng or the analyss of genome sequences (Joachms 2002). The key reason for the great success of SVMs n ths area s the exstence of strng kernels these are kernel functons defned on strngs that elegantly avod many of the combnatorc problems assocated wth other methods, whlst havng the advantage over generatve probablty models such as the Hdden Markov Model that the SVM learns to dscrmnate between the two classes va the maxmsaton of the margn. The practcal use of text categorsaton systems s extremely wdespread, wth most large enterprses relyng on such analyss of ther customer nteractons n order to provde automated response systems that are nonetheless talored to the ndvdual. Furthermore, the SVM has been successfully used n a study of text and data mnng for drect marketng applcatons (Cheung 2003) n whch relatvely lmted customer nformaton was automatcally supplanted wth the preferences of a larger populaton, n order to determne effectve marketng strateges. To conclude ths survey note that whle the majorty of the marketng teams do not publsh ther methodologes, snce many of the mportant data mnng software packages (for example Oracle Data Mnng and SAS Enterprse Mner) have ncorporated the SVM, t s lkely that there s a sgnfcant and ncreasng use of the SVM n ndustral settngs. A Worked Example In A Practcal Gude to Support Vector Classfcaton (Chang et al 2003) a smple procedure for applyng the SVM classfer s provded for nexperenced practtoners of the SVM classfer. The procedure s ntended to be easy to follow, quck, and capable of producng reasonable generalzaton performance. The steps they advocate can be paraphrased as follows:. Convert the data to the nput format of the SVM software you ntend to use 2. Scale the ndvdual components of the data nto a common range 3. Use the Gaussan kernel functon 4. Use cross-valdaton to fnd the best parameters C (margn softness) and σ (Gaussan wdth) 5. Wth the values of C and σ determned by cross-valdaton, retran on the entre tranng set The above tasks are easly accomplshed usng, for example, the free lbsvm software package, as we wll demonstrate n detal n ths secton. We have chosen ths tool because t s free, easy to use and of a hgh qualty, although the majorty of our dscusson apples equally well to other SVM software packages wheren the same steps wll necessarly be requred. The pont of ths chapter, then, s to llustrate n a concrete fashon the process of applyng an SVM. The lbsvm software package wth whch we do ths conssts of three man command-lne tools, as well as a helper scrpt n the python language. The basc functons of these tools are summarzed here:

13 svm-scale grd.py svm-tran svm-predct Ths smple program smply rescales the data as n step 2 above. The nput s a data set, and the output s a new data set that has been rescaled. Ths functon can be used to assst n the cross valdaton parameter selecton process. It smply calculates a cross valdaton estmate of generalzaton performance for a range of values of C and the Gaussan kernel wdth σ. The results are then llustrated as a two dmensonal contour plot of generalzaton performance versus C and σ. Ths s the most sophstcated part of lbsvm, whch takes as nput a fle contanng the tranng examples, and outputs a model fle a lst of Support Vectors and correspondng α s, as well as the bas term and kernel parameters. The program also takes a number of nput arguments that are used to specfy the type of kernel functon and margn softness parameter. As well as some more techncal optons, the program also has the opton (used by grd.py) of computng an n-fold cross valdaton estmate of the generalzaton performance. Havng run svm-tran, svm-predct can be used to predct the class labels of a new set of unseen data. The nput to the program s a model fle and a dataset, and the output s a fle contanng the predcted labels, sgn(f(x)), for the gven dataset. Detaled nstructons for nstallng the software can be found on the lbsvm webste, We wll now demonstrate these three steps usng the example dataset at the begnnng of the chapter, n order to predct whch customers are lkely to be home broadband nternet users. To make the procedure clear, we wll gve detals of all the requred nput fles (contanng the labelled and unlabelled data), the output fle (contanng the learnt decson functon), and the command lne statements requred to produce and process these fles. Preprocessng (svm-scale) All of our dscussons so far have consdered the nput tranng examples as numercal vectors. In fact ths s not necessary as t s possble to defne kernels on dscrete quanttes, but we wll not worry about that here. Instead, notce that n our example tranng data n Table, each tranng example has several ndvdual features, both numercal and categorcal. There are three numercal features (age, ncome and years of educaton), and one categorcal feature (gender). In constructng tranng vectors for the SVM from these tranng examples, the numercal features are drectly assgned to ndvdual components of the tranng vectors Categorcal features, however, must be dealt wth slghtly dfferently. Typcally, f the categorcal feature belongs to one of m dfferent categores (here the categores are male and female so that our m s 2), then we map ths sngle categorcal feature nto m ndvdual bnary valued numercal features. A tranng vector whose categorcal feature corresponds to feature n (the orderng s rrelevant), wll have all zero values for these nto bnary valued numercal features, except for the n-th one, whch we set to. Ths s a smple way of ndcatng that the features are not related to one another by relatve magntudes. Once agan, the data n table one would thusly be represented by these four vectors, wth correspondng class labels y : x = [ ] ; y = + x 2 = [ ] ; y 2 = + x 3 = [ ] ; y 3 = - x 4 = [ ] ; y 4 = - In order to use the lbsvm software, we must represent the above data n a fle that s formatted accordng to the lbsvm standard. The format s very smple, and best descrbed wth an example. The above data would be represented by a sngle fle that looks lke ths:

14 + :30 2: :6 4: + : :2 4: - :6 2:2000 3: 4: - :35 2: :2 4: Each lne of the tranng fle represents one tranng example, and begns wth the class label (+ or -), followed by a space and then an arbtrary number of ndex:value pars. There should be no spaces between the colons and the ndexes or values, only between the ndvdual ndex:value pars. Note that f a feature takes on the value zero, t need not be ncluded as an ndex:value par, allowng data wth many zeros to be represented by a smaller fle. Now that we have our tranng data fle, we are ready to run svm_scale. As we dscovered n the frst secton, ultmately all our data wll be represented by the kernel functon evaluaton between ndvdual vectors. The purpose of ths program s to make some very smple adjustments to the data n order for t to be better represented by these kernel evaluatons. In accordance wth step 3 above we wll be usng the Gaussan kernel, whch can be expressed by 2 x (x ) k(x,y) = exp( - ) = exp(- ). σ 2 D y d y d 2 2 d= σ Here we have wrtten out the D ndvdual components of the vectors x and y, whch correspond to the (D = 5) ndvdual numercal features of our tranng examples. It s clear from the summaton on the rght, that f a gven feature has a much larger range of varaton than another feature t wll domnate the sum, and the feature wth the smaller range of varaton wll essentally be gnored. For our example, ths means that the ncome feature, whch has the largest range of values, wll receve an undue amount of attenton from the SVM algorthm. Clearly ths s a problem, and whle the Machne Learnng communty has yet to gve the fnal word on how to deal wth t n an optmal manner, many practtoners smply rescale the data so that each feature falls n the same range, for example between zero and one. Ths can be easly acheved usng svm_scale, whch takes as nput a data fle n lbsvm format, and outputs both a rescaled data fle and a set of scalng parameters. The rescaled data should then be used to tran the model, and the same scalng (as stored n the scalng parameters fle) should be appled to any unlabelled data before applyng the learnt decson functon. The format of the command s as follows: svm-scale s scalng_parameters_fle tranng_data_fle > rescaled_tranng_data_fle In order to apply the same scalng transformaton to the unlabelled set, svm_scale must be executed agan wth the followng arguments: svm-scale r scalng_parameters_fle unlabelled_data_fle > rescaled_unlabelled_data_fle Here the fle unlabelled_data_fle contans the unlabelled data, and has an dentcal format to the tranng fle, asde from the fact that the labels + and - are optonal, and wll be gnored f they exst. Parameter selecton (grd.py) The parameter selecton process s wthout doubt the most dffcult step n applyng an SVM. Fortunately the smplstc method we prescrbe here s not only relatvely straghtforward, but also usually qute effectve. Our goal s to choose the C and σ values for our SVM. Followng the prevous dscusson about parameter or model selecton, our basc method of tacklng ths problem s to make a cross valdaton estmate of the generalzaton performance for a range of values of C and σ, and examne the results vsually. Gven the outcome of ths step, we may ether choose values for C and σ, or conduct a further search based on the results we have already seen. The followng command wll construct a plot of the cross valdaton performance for our scaled dataset:

15 grd.py log2c -5,5, log2g -20,0, v 0 rescaled_tranng_data_fle The search range of the C and σ values are specfed by the log2c and log2g commands respectvely. In both cases the numbers that follow take the form begn,end,stepsze to ndcate that we wsh to search logarthmcally usng the values 2 begn,2 begn+ stepsze Specfyng -v n ndcates that we wsh to do n-fold cross valdaton (n the above command n = 0), and the last argument to the command ndcates whch data fle to use. The output of the program s a contour plot, saved n an mage fle of the name rescaled_tranng_data_fle.png. The output mage for the above command s depcted n Fgure end. Fgure 6: A contour plot of cross-valdaton accuracy for a gven tranng set as produced by grd.py The contour plot ndcates wth varous lne colours the cross-valdaton accuracy of the classfer, as a functon of C and σ ths s measured as a percentage of correct classfcatons, so that we prefer large values. Note that σ s n fact referred to as gamma by the lbsvm software the varable name s of course arbtrary, but we choose to refer to t as σ for compatblty wth the majorty of SVM lterature. Gven such a contour plot of performance, as stated prevously there are generally two conclusons to be reached:. The optmal (or at least satsfactory) values of C and σ are contaned wthn the plottng regon. 2. It s necessary to contnue the search for C and σ, over a dfferent range than that of the plot, n order to acheve better performance.

16 In the frst case, we can read the optmal values of C and σ from the output of the program on the command wndow. Each lne of output ndcates the best parameters that have been encountered up to that pont, and so we can take the last lne as our operatng parameters. In the second case, we must choose whch drecton to contnue the search. From fgure 6 t seems feasble to keep searchng over a range of smaller σ and larger C. Ths whole procedure s usually qute effectve, however there can be no denyng that the search for the correct parameters s stll somethng of a black art. Gven ths, we nvte nterested readers to experment for themselves, n order to get a basc feel for how thngs behave. For our purposes, we shall assume that a good choce s C = 2-2 = 0.25 and σ = 2-2 = 0.25, and proceed to the next step. Tranng (svm-tran) As we have seen, the cross valdaton process does not use all of the data for tranng at each teraton some of the tranng data must be excluded for evaluaton purposes. For ths reason t s stll necessary to do a fnal tranng run on the entre tranng set, usng the parameters that we have determned n the prevous parameter selecton process. The command to tran s: svm-tran g 0.25 c 0.25 rescaled_tranng_data_fle model_fle Ths command sets C and σ usng the c and g swtches, respectvely. The other two arguments are the name of the tranng data, and fnally the fle name for the learnt decson functon or model. Predcton (svm-predct) The fnal step s very smple. Now that we have a decson functon, stored n the fle model_fle as well as a properly scaled set of unlabelled data, we can compute the predcted label of each of the examples n the set of unlabelled data by executng the command: svm-predct rescaled_unlabelled_data_fle model_fle predctons_fle After executng ths command, we wll have a new fle of the name predctons_fle, Each lne of ths fle wll contan ether + or - dependng on the predcted label of the correspondng entry n the fle rescaled_unlabelled_data. Summary The general problem of nducton s an mportant one, and can add a great deal of value to large corporate databases. Analysng ths data s not always smple however, and t s fortunate that methods that are both easy to apply and effectve have fnally arsen, such as the Support Vector Machne. The basc concept underlyng the Support Vector Machne s qute smple and ntutve, and nvolves separatng our two classes of data from one another usng a lnear functon that s the maxmum possble dstance from the data. Ths basc dea becomes a powerful learnng algorthm, when one overcomes the ssue of lnear separablty (by allowng margn errors), and mplct mappng to more descrptve feature spaces (through the use of kernel functons). Moreover, there exst free and easy to use software packages, such as lbsvm, that allow one to obtan good results wth a mnmum of effort. The contnued uptake of these tools s nevtable, but s often mpeded by the poor results obtaned by novces. We hope that ths chapter s a useful ad n avodng ths problem, as t quckly affords a basc understandng of both the theory and practce of the SVM.

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law) Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

Classification / Regression Support Vector Machines

Classification / Regression Support Vector Machines Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

Edge Detection in Noisy Images Using the Support Vector Machines

Edge Detection in Noisy Images Using the Support Vector Machines Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona

More information

Complex Numbers. Now we also saw that if a and b were both positive then ab = a b. For a second let s forget that restriction and do the following.

Complex Numbers. Now we also saw that if a and b were both positive then ab = a b. For a second let s forget that restriction and do the following. Complex Numbers The last topc n ths secton s not really related to most of what we ve done n ths chapter, although t s somewhat related to the radcals secton as we wll see. We also won t need the materal

More information

Support Vector Machines. CS534 - Machine Learning

Support Vector Machines. CS534 - Machine Learning Support Vector Machnes CS534 - Machne Learnng Perceptron Revsted: Lnear Separators Bnar classfcaton can be veed as the task of separatng classes n feature space: b > 0 b 0 b < 0 f() sgn( b) Lnear Separators

More information

Mathematics 256 a course in differential equations for engineering students

Mathematics 256 a course in differential equations for engineering students Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

Announcements. Supervised Learning

Announcements. Supervised Learning Announcements See Chapter 5 of Duda, Hart, and Stork. Tutoral by Burge lnked to on web page. Supervsed Learnng Classfcaton wth labeled eamples. Images vectors n hgh-d space. Supervsed Learnng Labeled eamples

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour 6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the

More information

CMPS 10 Introduction to Computer Science Lecture Notes

CMPS 10 Introduction to Computer Science Lecture Notes CPS 0 Introducton to Computer Scence Lecture Notes Chapter : Algorthm Desgn How should we present algorthms? Natural languages lke Englsh, Spansh, or French whch are rch n nterpretaton and meanng are not

More information

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1)

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1) Secton 1.2 Subsets and the Boolean operatons on sets If every element of the set A s an element of the set B, we say that A s a subset of B, or that A s contaned n B, or that B contans A, and we wrte A

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

Problem Set 3 Solutions

Problem Set 3 Solutions Introducton to Algorthms October 4, 2002 Massachusetts Insttute of Technology 6046J/18410J Professors Erk Demane and Shaf Goldwasser Handout 14 Problem Set 3 Solutons (Exercses were not to be turned n,

More information

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 48 CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 3.1 INTRODUCTION The raw mcroarray data s bascally an mage wth dfferent colors ndcatng hybrdzaton (Xue

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes SPH3UW Unt 7.3 Sphercal Concave Mrrors Page 1 of 1 Notes Physcs Tool box Concave Mrror If the reflectng surface takes place on the nner surface of the sphercal shape so that the centre of the mrror bulges

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS46: Mnng Massve Datasets Jure Leskovec, Stanford Unversty http://cs46.stanford.edu /19/013 Jure Leskovec, Stanford CS46: Mnng Massve Datasets, http://cs46.stanford.edu Perceptron: y = sgn( x Ho to fnd

More information

Intro. Iterators. 1. Access

Intro. Iterators. 1. Access Intro Ths mornng I d lke to talk a lttle bt about s and s. We wll start out wth smlartes and dfferences, then we wll see how to draw them n envronment dagrams, and we wll fnsh wth some examples. Happy

More information

Wishing you all a Total Quality New Year!

Wishing you all a Total Quality New Year! Total Qualty Management and Sx Sgma Post Graduate Program 214-15 Sesson 4 Vnay Kumar Kalakband Assstant Professor Operatons & Systems Area 1 Wshng you all a Total Qualty New Year! Hope you acheve Sx sgma

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics Introducton G10 NAG Fortran Lbrary Chapter Introducton G10 Smoothng n Statstcs Contents 1 Scope of the Chapter... 2 2 Background to the Problems... 2 2.1 Smoothng Methods... 2 2.2 Smoothng Splnes and Regresson

More information

Classifying Acoustic Transient Signals Using Artificial Intelligence

Classifying Acoustic Transient Signals Using Artificial Intelligence Classfyng Acoustc Transent Sgnals Usng Artfcal Intellgence Steve Sutton, Unversty of North Carolna At Wlmngton (suttons@charter.net) Greg Huff, Unversty of North Carolna At Wlmngton (jgh7476@uncwl.edu)

More information

Hermite Splines in Lie Groups as Products of Geodesics

Hermite Splines in Lie Groups as Products of Geodesics Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 A mathematcal programmng approach to the analyss, desgn and

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

GSLM Operations Research II Fall 13/14

GSLM Operations Research II Fall 13/14 GSLM 58 Operatons Research II Fall /4 6. Separable Programmng Consder a general NLP mn f(x) s.t. g j (x) b j j =. m. Defnton 6.. The NLP s a separable program f ts objectve functon and all constrants are

More information

Machine Learning: Algorithms and Applications

Machine Learning: Algorithms and Applications 14/05/1 Machne Learnng: Algorthms and Applcatons Florano Zn Free Unversty of Bozen-Bolzano Faculty of Computer Scence Academc Year 011-01 Lecture 10: 14 May 01 Unsupervsed Learnng cont Sldes courtesy of

More information

TN348: Openlab Module - Colocalization

TN348: Openlab Module - Colocalization TN348: Openlab Module - Colocalzaton Topc The Colocalzaton module provdes the faclty to vsualze and quantfy colocalzaton between pars of mages. The Colocalzaton wndow contans a prevew of the two mages

More information

Parallel matrix-vector multiplication

Parallel matrix-vector multiplication Appendx A Parallel matrx-vector multplcaton The reduced transton matrx of the three-dmensonal cage model for gel electrophoress, descrbed n secton 3.2, becomes excessvely large for polymer lengths more

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

An Entropy-Based Approach to Integrated Information Needs Assessment

An Entropy-Based Approach to Integrated Information Needs Assessment Dstrbuton Statement A: Approved for publc release; dstrbuton s unlmted. An Entropy-Based Approach to ntegrated nformaton Needs Assessment June 8, 2004 Wllam J. Farrell Lockheed Martn Advanced Technology

More information

Three supervised learning methods on pen digits character recognition dataset

Three supervised learning methods on pen digits character recognition dataset Three supervsed learnng methods on pen dgts character recognton dataset Chrs Flezach Department of Computer Scence and Engneerng Unversty of Calforna, San Dego San Dego, CA 92093 cflezac@cs.ucsd.edu Satoru

More information

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET 1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School

More information

Brave New World Pseudocode Reference

Brave New World Pseudocode Reference Brave New World Pseudocode Reference Pseudocode s a way to descrbe how to accomplsh tasks usng basc steps lke those a computer mght perform. In ths week s lab, you'll see how a form of pseudocode can be

More information

Review of approximation techniques

Review of approximation techniques CHAPTER 2 Revew of appromaton technques 2. Introducton Optmzaton problems n engneerng desgn are characterzed by the followng assocated features: the objectve functon and constrants are mplct functons evaluated

More information

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for

More information

A Modified Median Filter for the Removal of Impulse Noise Based on the Support Vector Machines

A Modified Median Filter for the Removal of Impulse Noise Based on the Support Vector Machines A Modfed Medan Flter for the Removal of Impulse Nose Based on the Support Vector Machnes H. GOMEZ-MORENO, S. MALDONADO-BASCON, F. LOPEZ-FERRERAS, M. UTRILLA- MANSO AND P. GIL-JIMENEZ Departamento de Teoría

More information

Meta-heuristics for Multidimensional Knapsack Problems

Meta-heuristics for Multidimensional Knapsack Problems 2012 4th Internatonal Conference on Computer Research and Development IPCSIT vol.39 (2012) (2012) IACSIT Press, Sngapore Meta-heurstcs for Multdmensonal Knapsack Problems Zhbao Man + Computer Scence Department,

More information

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization Problem efntons and Evaluaton Crtera for Computatonal Expensve Optmzaton B. Lu 1, Q. Chen and Q. Zhang 3, J. J. Lang 4, P. N. Suganthan, B. Y. Qu 6 1 epartment of Computng, Glyndwr Unversty, UK Faclty

More information

AP PHYSICS B 2008 SCORING GUIDELINES

AP PHYSICS B 2008 SCORING GUIDELINES AP PHYSICS B 2008 SCORING GUIDELINES General Notes About 2008 AP Physcs Scorng Gudelnes 1. The solutons contan the most common method of solvng the free-response questons and the allocaton of ponts for

More information

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS

NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS ARPN Journal of Engneerng and Appled Scences 006-017 Asan Research Publshng Network (ARPN). All rghts reserved. NUMERICAL SOLVING OPTIMAL CONTROL PROBLEMS BY THE METHOD OF VARIATIONS Igor Grgoryev, Svetlana

More information

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr)

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr) Helsnk Unversty Of Technology, Systems Analyss Laboratory Mat-2.08 Independent research projects n appled mathematcs (3 cr) "! #$&% Antt Laukkanen 506 R ajlaukka@cc.hut.f 2 Introducton...3 2 Multattrbute

More information

SVM-based Learning for Multiple Model Estimation

SVM-based Learning for Multiple Model Estimation SVM-based Learnng for Multple Model Estmaton Vladmr Cherkassky and Yunqan Ma Department of Electrcal and Computer Engneerng Unversty of Mnnesota Mnneapols, MN 55455 {cherkass,myq}@ece.umn.edu Abstract:

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices Internatonal Mathematcal Forum, Vol 7, 2012, no 52, 2549-2554 An Applcaton of the Dulmage-Mendelsohn Decomposton to Sparse Null Space Bases of Full Row Rank Matrces Mostafa Khorramzadeh Department of Mathematcal

More information

Chapter 6 Programmng the fnte element method Inow turn to the man subject of ths book: The mplementaton of the fnte element algorthm n computer programs. In order to make my dscusson as straghtforward

More information

Performance Evaluation of Information Retrieval Systems

Performance Evaluation of Information Retrieval Systems Why System Evaluaton? Performance Evaluaton of Informaton Retreval Systems Many sldes n ths secton are adapted from Prof. Joydeep Ghosh (UT ECE) who n turn adapted them from Prof. Dk Lee (Unv. of Scence

More information

A New Approach For the Ranking of Fuzzy Sets With Different Heights

A New Approach For the Ranking of Fuzzy Sets With Different Heights New pproach For the ankng of Fuzzy Sets Wth Dfferent Heghts Pushpnder Sngh School of Mathematcs Computer pplcatons Thapar Unversty, Patala-7 00 Inda pushpndersnl@gmalcom STCT ankng of fuzzy sets plays

More information

Discriminative classifiers for object classification. Last time

Discriminative classifiers for object classification. Last time Dscrmnatve classfers for object classfcaton Thursday, Nov 12 Krsten Grauman UT Austn Last tme Supervsed classfcaton Loss and rsk, kbayes rule Skn color detecton example Sldng ndo detecton Classfers, boostng

More information

Using Neural Networks and Support Vector Machines in Data Mining

Using Neural Networks and Support Vector Machines in Data Mining Usng eural etworks and Support Vector Machnes n Data Mnng RICHARD A. WASIOWSKI Computer Scence Department Calforna State Unversty Domnguez Hlls Carson, CA 90747 USA Abstract: - Multvarate data analyss

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervsed Learnng and Clusterng Why consder unlabeled samples?. Collectng and labelng large set of samples s costly Gettng recorded speech s free, labelng s tme consumng 2. Classfer could be desgned

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

SUMMARY... I TABLE OF CONTENTS...II INTRODUCTION...

SUMMARY... I TABLE OF CONTENTS...II INTRODUCTION... Summary A follow-the-leader robot system s mplemented usng Dscrete-Event Supervsory Control methods. The system conssts of three robots, a leader and two followers. The dea s to get the two followers to

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

General Vector Machine. Hong Zhao Department of Physics, Xiamen University

General Vector Machine. Hong Zhao Department of Physics, Xiamen University General Vector Machne Hong Zhao (zhaoh@xmu.edu.cn) Department of Physcs, Xamen Unversty The support vector machne (SVM) s an mportant class of learnng machnes for functon approach, pattern recognton, and

More information

A MOVING MESH APPROACH FOR SIMULATION BUDGET ALLOCATION ON CONTINUOUS DOMAINS

A MOVING MESH APPROACH FOR SIMULATION BUDGET ALLOCATION ON CONTINUOUS DOMAINS Proceedngs of the Wnter Smulaton Conference M E Kuhl, N M Steger, F B Armstrong, and J A Jones, eds A MOVING MESH APPROACH FOR SIMULATION BUDGET ALLOCATION ON CONTINUOUS DOMAINS Mark W Brantley Chun-Hung

More information

Backpropagation: In Search of Performance Parameters

Backpropagation: In Search of Performance Parameters Bacpropagaton: In Search of Performance Parameters ANIL KUMAR ENUMULAPALLY, LINGGUO BU, and KHOSROW KAIKHAH, Ph.D. Computer Scence Department Texas State Unversty-San Marcos San Marcos, TX-78666 USA ae049@txstate.edu,

More information

The Codesign Challenge

The Codesign Challenge ECE 4530 Codesgn Challenge Fall 2007 Hardware/Software Codesgn The Codesgn Challenge Objectves In the codesgn challenge, your task s to accelerate a gven software reference mplementaton as fast as possble.

More information

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009.

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009. Farrukh Jabeen Algorthms 51 Assgnment #2 Due Date: June 15, 29. Assgnment # 2 Chapter 3 Dscrete Fourer Transforms Implement the FFT for the DFT. Descrbed n sectons 3.1 and 3.2. Delverables: 1. Concse descrpton

More information

5 The Primal-Dual Method

5 The Primal-Dual Method 5 The Prmal-Dual Method Orgnally desgned as a method for solvng lnear programs, where t reduces weghted optmzaton problems to smpler combnatoral ones, the prmal-dual method (PDM) has receved much attenton

More information

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration Improvement of Spatal Resoluton Usng BlockMatchng Based Moton Estmaton and Frame Integraton Danya Suga and Takayuk Hamamoto Graduate School of Engneerng, Tokyo Unversty of Scence, 6-3-1, Nuku, Katsuska-ku,

More information

LECTURE NOTES Duality Theory, Sensitivity Analysis, and Parametric Programming

LECTURE NOTES Duality Theory, Sensitivity Analysis, and Parametric Programming CEE 60 Davd Rosenberg p. LECTURE NOTES Dualty Theory, Senstvty Analyss, and Parametrc Programmng Learnng Objectves. Revew the prmal LP model formulaton 2. Formulate the Dual Problem of an LP problem (TUES)

More information

X- Chart Using ANOM Approach

X- Chart Using ANOM Approach ISSN 1684-8403 Journal of Statstcs Volume 17, 010, pp. 3-3 Abstract X- Chart Usng ANOM Approach Gullapall Chakravarth 1 and Chaluvad Venkateswara Rao Control lmts for ndvdual measurements (X) chart are

More information

Outline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like:

Outline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like: Self-Organzng Maps (SOM) Turgay İBRİKÇİ, PhD. Outlne Introducton Structures of SOM SOM Archtecture Neghborhoods SOM Algorthm Examples Summary 1 2 Unsupervsed Hebban Learnng US Hebban Learnng, Cntd 3 A

More information

INF 4300 Support Vector Machine Classifiers (SVM) Anne Solberg

INF 4300 Support Vector Machine Classifiers (SVM) Anne Solberg INF 43 Support Vector Machne Classfers (SVM) Anne Solberg (anne@f.uo.no) 9..7 Lnear classfers th mamum margn for toclass problems The kernel trck from lnear to a hghdmensonal generalzaton Generaton from

More information

3D vector computer graphics

3D vector computer graphics 3D vector computer graphcs Paolo Varagnolo: freelance engneer Padova Aprl 2016 Prvate Practce ----------------------------------- 1. Introducton Vector 3D model representaton n computer graphcs requres

More information

Biostatistics 615/815

Biostatistics 615/815 The E-M Algorthm Bostatstcs 615/815 Lecture 17 Last Lecture: The Smplex Method General method for optmzaton Makes few assumptons about functon Crawls towards mnmum Some recommendatons Multple startng ponts

More information

A Statistical Model Selection Strategy Applied to Neural Networks

A Statistical Model Selection Strategy Applied to Neural Networks A Statstcal Model Selecton Strategy Appled to Neural Networks Joaquín Pzarro Elsa Guerrero Pedro L. Galndo joaqun.pzarro@uca.es elsa.guerrero@uca.es pedro.galndo@uca.es Dpto Lenguajes y Sstemas Informátcos

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15 CS434a/541a: Pattern Recognton Prof. Olga Veksler Lecture 15 Today New Topc: Unsupervsed Learnng Supervsed vs. unsupervsed learnng Unsupervsed learnng Net Tme: parametrc unsupervsed learnng Today: nonparametrc

More information

Programming in Fortran 90 : 2017/2018

Programming in Fortran 90 : 2017/2018 Programmng n Fortran 90 : 2017/2018 Programmng n Fortran 90 : 2017/2018 Exercse 1 : Evaluaton of functon dependng on nput Wrte a program who evaluate the functon f (x,y) for any two user specfed values

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

Solving two-person zero-sum game by Matlab

Solving two-person zero-sum game by Matlab Appled Mechancs and Materals Onlne: 2011-02-02 ISSN: 1662-7482, Vols. 50-51, pp 262-265 do:10.4028/www.scentfc.net/amm.50-51.262 2011 Trans Tech Publcatons, Swtzerland Solvng two-person zero-sum game by

More information

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

Steps for Computing the Dissimilarity, Entropy, Herfindahl-Hirschman and. Accessibility (Gravity with Competition) Indices

Steps for Computing the Dissimilarity, Entropy, Herfindahl-Hirschman and. Accessibility (Gravity with Competition) Indices Steps for Computng the Dssmlarty, Entropy, Herfndahl-Hrschman and Accessblty (Gravty wth Competton) Indces I. Dssmlarty Index Measurement: The followng formula can be used to measure the evenness between

More information

Skew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach

Skew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach Angle Estmaton and Correcton of Hand Wrtten, Textual and Large areas of Non-Textual Document Images: A Novel Approach D.R.Ramesh Babu Pyush M Kumat Mahesh D Dhannawat PES Insttute of Technology Research

More information

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

Active Contours/Snakes

Active Contours/Snakes Actve Contours/Snakes Erkut Erdem Acknowledgement: The sldes are adapted from the sldes prepared by K. Grauman of Unversty of Texas at Austn Fttng: Edges vs. boundares Edges useful sgnal to ndcate occludng

More information

RECOGNIZING GENDER THROUGH FACIAL IMAGE USING SUPPORT VECTOR MACHINE

RECOGNIZING GENDER THROUGH FACIAL IMAGE USING SUPPORT VECTOR MACHINE Journal of Theoretcal and Appled Informaton Technology 30 th June 06. Vol.88. No.3 005-06 JATIT & LLS. All rghts reserved. ISSN: 99-8645 www.jatt.org E-ISSN: 87-395 RECOGNIZING GENDER THROUGH FACIAL IMAGE

More information

Module Management Tool in Software Development Organizations

Module Management Tool in Software Development Organizations Journal of Computer Scence (5): 8-, 7 ISSN 59-66 7 Scence Publcatons Management Tool n Software Development Organzatons Ahmad A. Al-Rababah and Mohammad A. Al-Rababah Faculty of IT, Al-Ahlyyah Amman Unversty,

More information

Some Advanced SPC Tools 1. Cumulative Sum Control (Cusum) Chart For the data shown in Table 9-1, the x chart can be generated.

Some Advanced SPC Tools 1. Cumulative Sum Control (Cusum) Chart For the data shown in Table 9-1, the x chart can be generated. Some Advanced SP Tools 1. umulatve Sum ontrol (usum) hart For the data shown n Table 9-1, the x chart can be generated. However, the shft taken place at sample #21 s not apparent. 92 For ths set samples,

More information

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,

More information

SENSITIVITY ANALYSIS IN LINEAR PROGRAMMING USING A CALCULATOR

SENSITIVITY ANALYSIS IN LINEAR PROGRAMMING USING A CALCULATOR SENSITIVITY ANALYSIS IN LINEAR PROGRAMMING USING A CALCULATOR Judth Aronow Rchard Jarvnen Independent Consultant Dept of Math/Stat 559 Frost Wnona State Unversty Beaumont, TX 7776 Wnona, MN 55987 aronowju@hal.lamar.edu

More information

CSCI 104 Sorting Algorithms. Mark Redekopp David Kempe

CSCI 104 Sorting Algorithms. Mark Redekopp David Kempe CSCI 104 Sortng Algorthms Mark Redekopp Davd Kempe Algorthm Effcency SORTING 2 Sortng If we have an unordered lst, sequental search becomes our only choce If we wll perform a lot of searches t may be benefcal

More information

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur FEATURE EXTRACTION Dr. K.Vjayarekha Assocate Dean School of Electrcal and Electroncs Engneerng SASTRA Unversty, Thanjavur613 41 Jont Intatve of IITs and IISc Funded by MHRD Page 1 of 8 Table of Contents

More information

(1) The control processes are too complex to analyze by conventional quantitative techniques.

(1) The control processes are too complex to analyze by conventional quantitative techniques. Chapter 0 Fuzzy Control and Fuzzy Expert Systems The fuzzy logc controller (FLC) s ntroduced n ths chapter. After ntroducng the archtecture of the FLC, we study ts components step by step and suggest a

More information