Knowledge-Based Systems

Size: px
Start display at page:

Download "Knowledge-Based Systems"

Transcription

1 Knowledge-Based Systems 54 (13) Contents lsts avalable at ScenceDrect Knowledge-Based Systems journal homepage: Mult-vew classfcaton wth cross-vew must-lnk and cannot-lnk sde nformaton Qang Qan a, Songcan Chen a,, udong Zhou b a Department of Computer Scence and Engneerng, Nanjng Unversty of Aeronautcs and Astronautcs, Nanjng 2116, P.R. Chna b Informaton Engneerng College, Yangzhou Unversty, Yangzhou 2259, P.R. Chna artcle nfo abstract Artcle hstory: Receved 22 October 12 Receved n revsed form 1 September 13 Accepted 5 September 13 Avalable onlne 9 October 13 Keywords: Classfcaton Mult-vew learnng Wthout correspondence Unpared mult-vew data Cross-vew sde nformaton Sde nformaton, lke must-lnk (ML) and cannot-lnk (CL), has been wdely used n sngle-vew classfcaton tasks. However, so far such nformaton has never been appled n mult-vew classfcaton tasks. In many real world stuatons, data wth multple representatons or vews are frequently encountered, and most proposed algorthms for such learnng stuatons requre that all the mult-vew data should be pared. Yet ths requrement s dffcult to satsfy n some settngs and the mult-vew data could be totally unpared. In ths paper, we propose an learnng framework to desgn the mult-vew classfers by only employng the weak sde nformaton of cross-vew must-lnks (CvML) and cross-vew cannotlnks (CvCL). The CvML and the CvCL generalze the tradtonal sngle-vew must-lnk (SvML) and sngle-vew cannot-lnk (SvCL), and to the best of our knowledge, are frst defntely ntroduced and appled nto the mult-vew classfcaton stuatons. Fnally, we demonstrate the effectveness of our method n our experments. Ó 13 Elsever B.V. All rghts reserved. 1. Introducton Tradtonal learnng only nvolves data wth sngle vew. However, n many real world crcumstances, data wth multple natural feature representatons or vews are frequently encountered. For example, web pages can be represented by both ts own content and hyperlnks pontng to t. To tackle ths data type, mult-vew learnng has been developed snce the poneer works [7,31]. So far many approaches [7,8,24,] have been proposed and acheved emprcal and theoretcal successes. All of those approaches rely on two common assumptons, compatblty and ndependence between vews [7]. However, to make the two assumptons work, one requrement that should be fulflled s such mult-vew data should be pared. Specfcally, for the representaton of a sample n one vew, there s a correspondng representaton pared n the other vew. Sometmes ths requrement s over-rgorous n some crcumstances. For nstance, n a wreless sensor network, collected data could be mssed or polluted durng data transmsson due to devce malfuncton or malcous attacks. Thus only partal data are pared whle the rest are unpared [15]. In[19,15,18,25], some methods have been proposed to deal wth ths scenaro. More extremely, n some crcumstances where all data are even unpared, for Correspondng author. Tel.: E-mal addresses: qan.qang.yx@gmal.com (Q. Qan), s.chen@nuaa.edu.cn (S. Chen), xdzhou@nuaa.edu.cn (. Zhou). example, web pages from Englsh Routers and French Routers are unpared. We may not easly know whch Englsh web page corresponds to whch French web page. Ths paper focuses on the most dffcult totally-unpared extreme crcumstance. Snce no parng nformaton between multple vews exsts, we ntroduce a new type of sde nformaton, called cross-vew must-lnk and cross-vew cannot-lnk, to help learnng. Must-lnk and cannot-lnk sde nformaton s usually used n the classfcaton [3,34] and clusterng learnng [33,28] on sngle-vew data (called SvML and SvCL n ths paper). Two samples belongng to the SvML set share the same label, whle n the SvCL set possess dfferent labels. Compared wth commonly-used supervsed labels, such SvMLs and SvCLs are weaker n characterzng supervson nformaton. Vrtually, we can nfer both the SvML and SvCL between samples f knowng ther label nformaton, but cannot reversely. The SvML and the SvCL only provde the label relatons between samples wthn vew, thus cannot help the totally-unpared mult-vew learnng. To acheve ths goal, some cross-vew relatons are needed. Consequently, n ths paper, we ntroduce the cross-vew must-lnk (CvML) and cross-vew cannot-lnk (CvCL). Two representatons n dfferent vews n CvML set ndcates that ther labels are the same, whle n CvCL set ndcates that ther labels are dfferent. Unlke SvML and SvCL, CvML and CvCL buld the mplct label relatons across dfferent vews. As a result, we can transfer mutually the learnng nformaton between dfferent vews through these CvMLs and CvCLs. Compared wth explct label nformaton, CvML and CvCL are lkewse also weaker /$ - see front matter Ó 13 Elsever B.V. All rghts reserved.

2 138 Q. Qan et al. / Knowledge-Based Systems 54 (13) n supervson lke SvML and SvCL for the same reason. Moreover, the pared nformaton belongs to the CvML because pared representatons must own the same label, but the CvML does not mean pared representatons because two representatons lnked by CvML could come from dfferent samples. Intutvely, by forcng the outputs of the target classfcaton functons n each vew to obey the CvML and the CvCL constrants, the outputs learned n one vew can be shfted to that n the other vew through both of them and ad the classfcaton learnng n that vew. To the best of our knowledge, such sde nformaton has never defntely been ntroduced and appled n mult-vew classfer desgn. The proposed framework s based on the classcal regularzaton frameworks [27,2] and the new regularzaton terms whch encode the CvMLs and the CvCLs sde nformaton,.e. forcng the outputs of the representatons n the CvMLs to be the same and the outputs of the representatons n the CvCLs to be dfferent. Snce the true (strongly-supervsed) labels are unknown, the classcal regularzaton framework has to be modfed by ntroducng probablstc ndcators whch ndcate how possble the sample belongs to a gven class. Such a modfed framework leads to a block-wse convex optmzaton problem whch can be teratvely solved effectvely by the classc block coordnate descent method wth guarantee of teratve convergence to a statonary pont [5]. Our experments demonstrate ts effectveness as well. We summarze our contrbuton as follows: We ntroduce and deepen the concepts of the CvML and the CvCL, whch are extensons of the SvML and the SvCL, to ad the jont classfcaton learnng n dfferent vews. We develop a classfcaton learnng framework whch utlzes the cross-vew sde nformaton to learn classfers n the tough unpared mult-vew settngs. The rest of the paper s organzed as follows. In Secton 2, we revew some related work. Then we ntroduce our framework n Secton 3. Next we llustrate our experment results n Secton 4. And fnally, we conclude ths paper and present the future work n Secton Related work Our work s related to both classfcaton learnng wth the SvML and SvCL sde nformaton and mult-vew classfcaton. Thus we revew the two parts respectvely. Snce the ML and CL sde nformaton has never been used n mult-vew classfcaton, we manly revew the related work on the sngle-vew crcumstance SvMLs and SvCLs for classfcaton The ML and CL sde nformaton n sngle vew has demonstrated ts value n classfcaton tasks. Yan et al. [3] formulated both MLs and CLs nto a convex parwse loss functon and ntegrated t nto the tradtonal margn-based learnng framework. Thus the proposed framework can handle both the label and MLs/CLs together. Nguyen and Caruana [22] ncorporated both MLs and CLs nto the margn-based learnng framework and proposed PCSVM algorthm. Zhang and Yan [34] frst transformed both the ML and the CL pars of samples nto a new space and learned an estmator there, then transformed the estmator back nto the orgnal sample space. They proved that the fnal estmator s sgnnsenstvely consstent wth the optmal decson boundary and gave out ts asymptotc varance. Rather than those drectly ncorporatng both the MLs and the CLs nto classfcaton models, metrc learnng goes along another lne. It frst learns a Mahalanobs metrc whch obeys MLs and CLs constrants, then uses the dstance-based classfers lke the k nearest neghbor to classfy the test data. Typcal works nclude [1,29,13,23]. The ML and CL sde nformaton s also used to learn proper kernel matrces for the later kernel machne algorthms. L et al. [21] forced the entres of kernel matrx correspondng to MLs and CLs to be 1 and respectvely and developed a kernel learnng algorthm PCP. PCP s computatonally ntensve because t s solved by Semdefnte Programmng (SDP). Hu et al. [17] proposed kernel propagaton method to avod solvng SDP on full kernel matrx. The man dea s frst to learn a small kernel matrx then propagate t nto full kernel matrx. The SvMLs and the SvCLs are also appled n other tasks lke clusterng, mage segmentaton et al. For user s reference we name a few works n typcal applcaton domans lke mage segmentaton [33], vdeo survellance [16], clusterng [28,3,32]. Despte so many works on the SvMLs and the SvCLs, the CvMLs and the CvCLs are almost never touched to the best of our knowledge Mult-vew learnng Mult-vew learnng s a very natural learnng settngs. It was frst touched n Yarowsky s [31] and Blum et al. s [7] works. Blum et al. proposed the renowned co-tranng algorthm. It alternatvely trans the predctor n one vew and uses the predcted labels to ad the tranng n another vew. Dasgupta et al. [9] proved a theoretcal PAC-style generalzaton bound of the co-tranng. Sndhwan et al. [24] ntroduced the co-regularzaton algorthm. The co-regularzaton algorthm drectly models the cross-vew agreement and ncorporates t nto a regularzaton framework. They ntroduced a famly of algorthms wth dfferent regularzaton frameworks (the classcal regularzaton framework and manfold regularzaton framework). The formulaton s a convex optmzaton problem rather than the style of alternatvely learnng on each vew lke the co-tranng. The formulaton s related to our framework to some extent. We wll compare wth t n Secton 3.4. Snce full pared mult-vew data are over-rgorous n some applcatons, some methods were proposed for partally pared crcumstance. Kmura et al. [18] consdered the stuaton where addtonal unpared data are provded and developed Sem Canoncal Correlaton Analyss (SemCCA) algorthm. They used both the pared and unpared data to regularze CCA through PCA-type penalty. Lampert and Kromer [19] proposed a modfed Maxmum Covarance Analyss algorthm for weakly-pared multmodal data. They guessed the parng between data vews and optmzed t along wth dmenson reducton parameter matrx. Blaschko et al. [6] modfed Kernel Canoncal Correlaton Analyss wth Laplacan regularzaton by usng unpared data and propose SemLRKCCA algorthm. However, they only embedded the local structure on constrants but not n ts objectve and had too many model parameters. The PPLCA algorthm propsed by Gu et al. [15] overcomes the shortcomngs of SemLRKCCA. PPLCA smultaneously embeds the local structure nto both objectves and constrants and has less model parameters. Sun et al. [25] developed dscrmnatve canoncal correlaton analyss n partally pared stuatons. They proposed DCCAM by estmatng the wthn-class and between-class correlaton on both the pared and unpared data. 3. Mult-vew learnng under cross-vew MLs and CLs In ths secton, we ntroduce our framework. Later, we compare our framework wth the co-regularzaton framework [24] at last. Our tranng process can formally be conducted n two steps: n the 1st step, we employ the avalable CvMLs and CvCLs to desgn a (sgn) classfer whch s used to decde whether any gven two

3 Q. Qan et al. / Knowledge-Based Systems 54 (13) samples are from the same class or not. In fact, ths step does not nvolve any label, thus the classfer s not used to decde a real class of a sample but just returns the sgn label of ether +1 or 1 to ndcate whether the sample shares the same real class or not wth some tranng sample. In order to determne ts real label fnally, we need formally the 2nd step n whch we adopt a few labeled samples provded to determne whch real label the prevous sgn label corresponds to. Fg. 1 shows the two steps Learnng classfers determnng sgn labels Our framework s based on the classcal regularzaton framework [27,2] for supervsed learnng. It solves the followng optmzaton problem: 1 N mn Vðx ; c ; f Þþkkfk 2 H f 2H k 2 k ð1þ ¼1 where H k s an Reproducng Kernel Hlbert Space (RKHS) nduced by a kernel k, c s the label of sample x, and V s a loss functon, such as the squared loss or the hnge loss. In our settngs, only the CvMLs and CvCLs sde nformaton are at hand and the sample labels are unavalable. Consequently, we are not certan of whch classes the samples belong to. The brought uncertanty s handled by ntroducng the probablstc ndcators just lke fuzzy c-means clusterng algorthm [11], and Eq. (1) s modfed as follows 1 mn f 2H k ;u 2 N ¼1 r2fþ; g u 2 r Vðx ; c r ; f Þþkkf k 2 H k where u for each x s the probablstc ndcator vector whch subjects to non-negatvty and untary summaton constrants. +, are the postve and negatve labels. We let c + = 1 and c = 1 respectvely. Usually n order to avod hard assgnment of labels, the exponents of u r are set to 2 rather than 1 [11]. Eq. (2) looks lke the fuzzy c-means (FCM) formulaton, however essentally there s a major dfference: n FCM the clusterng performs n sample space, thus the clusterng centers are a set of vector prototypes to be optmzed and have the same dmensonalty as the sample space, whle our algorthm performs n (label) output space wth dfferent dmensonalty from the sample space, and the centers are fxed to 1 and 1. The CvMLs and the CvCLs are used to regularze the learnng n both vews. The underlyng prncple s to force the classfers outputs n separate vews to obey both the CvMLs and the CvCLs constrants. For CvMLs, we smply use the square of dfference of the correspondng outputs. It s a convex formulaton as follows: ðf x ðx Þ f y ðy j ÞÞ 2 ð3þ ð;jþ2m where M denotes the CvML set. It mples that large output dfferences would ncur large penaltes. But for the CvCLs, t s not so easy to formulate [14,26]. Here we employ Goldberg et al. s [14] method and formulate the CvCLs nto a convex penalty: ðf x ðx Þþf y ðy j ÞÞ 2 ð4þ ð;jþ2c Fg. 1. Formal two-step tranng process. ð2þ where C denotes the CvCL set. Note that the mnus n Eq. (3) s substtuted for the plus n Eq. (4). The penalty s zero f f x (x ) and f y (y j ) have the same absolute value but opposte sgns, thus mnmzng the penalty mples dfferent output labels. The trval case f x (x )=f y (y j ) = s avoded because t wll rase classfcaton error. Ths dea could also be appled n mult-vew clusterng and dmenson reducton tasks f we can properly penalze the outputs whch volate the CvMLs and the CvCLs constrants. For example, f the outputs of cluster algorthm are multnomal random varables, we can consder penalze large (small) Kullback Lebler dvergences of the representatons n MðCÞ. Integratng them together suggests the followng optmzaton problem: mn J ¼ 1 Nx f x;f y;u x ;u y 2 ¼1 r2fþ; g Ny þ 1 2 j¼1 r2fþ; g þ k 3ðN x þ N y Þ 2jMj þ k 3ðN x þ N y Þ 2jCj u x 2 r Vðx ; c k ; f x Þþ k 1 2 kf xk 2 H kx u y 2 jr Vðy j ; c k ; f y Þþ k 2 2 kf yk 2 H ky ðf x ðx Þ f y ðy j ÞÞ 2 ð;jþ2m ðf x ðx Þþf y ðy j ÞÞ 2 ð;jþ2c s:t: u x r >¼ ; uy jr >¼ ; for r n fþ; g u x þ þ ux ¼ 1; uy þ jþ uy ¼ 1 j where x 1 ;...; x Nx and y 1 ;...; y Ny are tranng representatons n two vews respectvely and N x, N y are the number of them. jmj and jcj are the cardnalty of M and C respectvely. We dvde the CvMLs and the CvCLs penaltes by jmj and jcj to balance the dfferent numbers of CvMLs and CvCLs. In ths formulaton, the frst two lnes are the probablstc classfcaton regularzaton framework n each vews, and the last two lnes are the CvMLs and the CvCLs penaltes. Note that besdes CvMLs and CvCLs, no label nformaton s used here. Thus the learned classfers can only gve out +1 and 1 sgn labels. Later, a few labeled samples wll be used to determne whch real label the sgn label corresponds to Optmzaton Wthout loss of generalty, we use the square loss and propose the regularzed least square (RLS) under Cross-Vew MLs and CLs (RLSCVMC) algorthm. It s easy to see that the representer theorem holds (see appendx for the proof), thus the mnmzer f H x ; f H y have the followng forms: f H x f H y ¼ Nx ¼ Ny j a k x ðx; x Þ b j k y ðy; y j Þ By substtutng them for f x, f y n Eq. (5), we get the followng optmzaton problem: mn J ¼ 1 Nx a;b;u x ;u y 2 ¼1 r2fþ; g Ny þ 1 2 j¼1 r2fþ; g u x 2 k ða T k x c k Þ 2 þ k 1 þ k 3ðN x þ N y Þ 2jMj þ k 3ðN x þ N y Þ 2jCj 2 at K x a u y 2 k ðb T k yj c k Þ 2 þ k 2 2 bt K y b ða T k x b T k yj Þ 2 ð;jþ2m ða T k x þ b T k yj Þ 2 ð;jþ2c s:t: u x r >¼ ; uy jr >¼ ; for r n fþ; g u x þ þ ux ¼ 1; uy þ jþ uy ¼ 1 j ð5þ ð6þ ð7þ

4 1 Q. Qan et al. / Knowledge-Based Systems 54 (13) where matrces K x, K y are kernel matrx wth ther every entry be k x (x, x j ) and k y (y, y j ) respectvely. For notaton convenence, we let k x ¼½kðx ; x 1 Þ;...; kðx ; x Nx ÞŠ T and k yj ¼½kðy j ; y 1 Þ;...; kðy ; y Ny ÞŠ T. Note that the objectve s convex wth respect to each component though s nonconvex per se. Classc block coordnate descent method could be used to solve ths problem [5]. The basc dea s optmzng one block varable whle keepng other varable blocks fxed and repeatng ths step untl meetng some stop crtera. Because the objectve functon value decreases constantly after each step, ths procedure guarantees to converge to a statonary pont [5]. We optmze (u x, u y ), a and b teratvely whle keepng the rest varable blocks fxed. When a and b are fxed, optmzng u x and u y can be decoupled nto two ndependent but analogous problems smlar to Fuzzy c-means. Eq. (8) calculates current optmal soluton u x, and u y has an analogous formula. u x r ¼ 1=ða T k x c r Þ 2 Pr2þ; 1=ðaT k x c r Þ 2 ð8þ When u x, u y and b s fxed, the objectve s convex and quadratc wth respect to a. The current optmal soluton could be obtaned by settng ts dervatve to zero. Eqs. (1) and (9) are a s coeffcents of lnear and quadratc terms. By solvng the lnear equatons H a a = g a, we get the current optmal a. H a ¼ Nx g a ¼ Nx r2fþ; g þ k 3ðN x þ N y Þ jcj r2fþ; g k 3ðN x þ N y Þ jcj u x 2 r k x k T x þ k 1 K x þ k 3ðN x þ N y Þ jmj k x k T x ð;jþ2c u x 2 r c k k x k 3ðN x þ N y Þ jmj b T k yj k x ð;jþ2c k x k T x ð;jþ2m b T k yj k x ð;jþ2m ð9þ ð1þ When u x, u y and a s fxed, b can be optmzed by analogy wth the updatng formula of a. We omt the related formula of b here. Note that we have closed-form solutons for updatng each varable block. We summarze the whole algorthm nto Algorthm 1. Algorthm 1. RLS under the CvMLs and the CvCLs Input: data matrx,y, number of class, maxmum teraton number MaxIter ntalze a, b, whle ter < MaxIter do step 1: update u x, u y wth Eq. (8) and ts analogy for u y. step 2: update a = H a ng a by solvng a lnear equaton. step 3: update b by analogy wth updatng a end whle 3.3. Determnng real labels of the sgn labels In tranng data, besdes the CvMLs and CvCLs avalable, a few labeled samples are also provded. Snce the sgn labels of those samples have already been known after the classfers are learned, the sgn labels and the real labels from the same samples can be connected. For example, f a sample has the +1 sgn label as well as the frst real class label, we may say the +1 sgn label corresponds to the frst real class label. Note that we can only need as few as two labeled samples, one for each class, to connect the sgn labels to real classes. Table 1 Used supervson n Co-RLS and RLSCVMC. Here the cross-vew data correspondence n Co-RLS s categorzed as ML. Y w : unlabeled data s only used n co-regularzaton term not n least square loss. Labeled Unlabeled CvML CvCL Co-RLS Y Y w Y N RLSCVMC N Y Y Y 3.4. Comparson wth co-regularzaton RLSCVMC s related to Sndhwan et al. s co-regularzaton framework [24]. They proposed a famly of algorthms n the co-regularzaton framework: the Co-Regularzed Least Squares (Co-RLS), the Co-Regularzed Laplacan SVM and Least Squares (Co-LapSVM, Co-LapRLS). Co-LapSVM and Co-LapRLS place ther root on Manfold Regularzaton framework [4] whle our RLSCVMC s bult on classcal regularzaton framework. Thus we do not compare wth them and only compare wth Co-RLS algorthm lsted as follows: mn fx;f y l ðf x ðx Þ c Þ 2 þ l l þ c 2 kf y k 2 H ky þ cc ðl þ uþ lþu 2 f y ðy Þ c þ c 1 kf x k 2 H kx ðf x ðx Þ f y ðy ÞÞ 2 ð11þ where l, u mean the numbers of labeled and unlabeled samples respectvely. The formulaton conssts of two classcal regularzaton framework for each data vew and a co-regularzaton term. We can see that both Co-RLS and our framework are based on the classcal regularzaton framework. However our framework estmates the loss on unlabeled data by ntroducng probablstc ndcator vectors whle Co-RLS only estmates the loss on the labeled data. From the perspectve that the pared data can be treated as the CvML, the co-regularzaton term s exactly the same as the CvML penalty term n Eq. (3). Apparently Co-RLS does not employ the CvCL supervson explctly whle our framework explctly penalze the volaton of the CvCLs. We summarze the dfferent knds of nformaton used n Co-RLS and our framework n Table Experment In ths secton, we show the emprcal study on RLSCVMC algorthm. We frst ntroduce the datasets used n our experments, then show the performance of RLSCVMC under dfferent numbers of the CvMLs and the CvCLs supervson. In the next, we llustrate the classfcaton performance under dfferent parameter settngs. Fnally, we compare RLSCVMC wth Co-RLS algorthm. In all of our experments, lnear kernel s used as n Sndhwan et al. s co-regularzaton paper [24] Dataset descrpton In our experments, four mult-vew datasets lsted below are used. The Multple Feature (handwrtten) dgt data set (MFD). 1 Ths dataset comes from UCI machne learnng repostory [12]. It conssts of features of handwrtten numerals ( 9 ) extracted from a collecton of Dutch utlty maps. patterns per class (for a total of patterns) have been dgtzed n bnary mages. These dgts have sx feature sets. Here we only choose two feature sets as two vews used n our experments. They 1

5 Q. Qan et al. / Knowledge-Based Systems 54 (13) Table 2 Expermental dataset. NumP: #postve, NumN: #negatve, VName: Vew Name. Dataset Vew 1 Vew 2 Vew Dm Num NumP NumN Vew Dm Num NumP NumN MFD Px 2 Mor 6 Course Page Lnk ORL Px 124 LBP 124 MRC EN FR are px (2 pxel averages n 2 3 wndows) and mor (6 morphologcal features). We use as postve class and 1 as negatve class. Course dataset. Ths dataset conssts of 151 web pages collected from Computer Scence department webstes at four unverstes: Cornell, Unversty of Washngton, Unversty of Wsconsn and Unversty of Texas. These web pages are categorzed nto two classes: course and non-course. Two vews are web page content and text on the lnks to the web page. ORL. Ths dataset s a face dataset. It contans two feature sets. One s the cropped face mage (32 32) and the other s the LBP feature extracted from the mage. The two feature sets are treated as two vews n experments. We use the frst two persons as postve and negatve classes. Multlngual Reuters Collecton (MRC). 2 Ths dataset s defned and provded by [1]. It contans collectons of fve languages (EN,FR,GE,SP,IT) from sx large Reuters categores (CCAT, C15, ECAT, E21, GCAT and M11) extracted from RCV1 and RCV2. Ths dataset s totally unpared. In our experments, we use Englsh and French webpages as two vews and the CCAT/C15 categores as the postve and negatve class. Detals are lsted n Tabel 2. Note that, the frst three datasets have pared data whle the last does not. In all our experments, we reduce the dmensons of Course and ORL dataset to 1 and by PCA, and the dmenson of MRC to 1 by LSA Performance examnaton of RLSCVMC In ths experment, we show the performance of RLSCVMC on dfferent numbers of the CvMLs and the CvCLs. Snce CvMLs and CvCLs are weak supervson, we want to know how many CvMLs and CvCLs are need to gve out a satsfed performance. So we tran our framework on dfferent number of tranng set and see the predcton results. So we use the followng expermental settngs. The numbers of the CvMLs and the CvCLs grow from %(N x + N y )/2 to 1%(N x + N y )/2 at an nterval of 1%(N x + N y )/2. For each CvMLs and CvCLs number combnaton, ten trals are run and the mean accuraces are reported. For each tral, we randomly and evenly splt the dataset nto tranng set and testng set. The CvML s constructed by randomly choosng two representatons wth the same label n the tranng set of two vews, and the CvCL s constructed by randomly choosng two representatons wth dfferent labels n the tranng set of two vews. The a, b varables are randomly ntalzed, and n our experments we fnd out that our framework s nsenstve to the ntal values. Tradtonal parameter selecton method Cross-valdaton s not applcable here due to the absence of labeled data. So we tune the parameter heurstcally. For all experments we set k 3 = 1 whch approxmately balances the loss term and cross-vew sde nformaton regularzaton term. k 1 s set to 1 for dataset MFD and 1 for the rest datasets and k 2 s 2 set as the same wth k 1 for convenence. We draw the 3D-bar graph of the mean accuraces n Fg. 2. All subfgures n Fg. 2 demonstrate that the performance s ncreasng wth the MLs and the CLs numbers. However the ncrease of performance s not monotonous wth respect to the number of the CvMLs and the CvCLs especally for the ORL dataset. That may be caused by the unstable learned predctors on these datasets. Table 3 lsts the average standard devatons on the four dataset. Among them, the ORL dataset gets a hgh average standard devaton and obtans a vbratng ncrease of performance, whle the Course and the MRC datasets have very small average standard devaton and ther ncreasng are almost monotonous. Fg. 4 plots the dagonal bars n Fg. 2 for each dataset. On the Course and the MRC datasets, ther accuraces both rse quckly to about at % supervson, then keep almost stable. Whle on the ORL dataset, the accuracy ncreases approxmately lnearly. In addton, on all but the MFD datasets, the accuraces on ther two vews assume smlar trend whch may be partly due to the co-regularzaton among dfferent vews. An nterestng phenomenon s that the accuracy does not ncrease when the number of CvMLs (CvCLs) ncreases and the number of CvCLs (CvMLs) s zero, especally on the Course, the MRC and the ORL datasets. Actually t s a degenerated soluton of RLSCVMC. We demonstrate ths phenomenon by a toy problem n Fg. 3. It s a two-vew two-class dataset. Each class n each vew s generated from a Gaussan dstrbuton. We draw the classfcaton hyperplane (the blue lne), and label postve and negatve area on the 2D plane on both vews. Fg. 3(a)/(b) depcts the stuaton when CvCLs/CvMLs does not exst. When only CvMLs exst n Fg. 3(a), the classfers n both vews gve all data the same label. Apparently ths soluton ncurs lttle penalty on the CvMLs regularzaton because of the same label. Furthermore, due to adapton of the probablstc ndcator vector u x, u y n Eq. (2), t also ncurs small classfcaton loss. When only the CvCLs exst, the stuaton s smlar. The classfers gve data n dfferent vews opposte labels, thus ncur lttle penalty on the CvCLs regularzaton. And the classfcaton loss n Eq. (2) s also small. To avod ths knd of degenerated soluton, both the CvMLs and the CvCLs supervson are needed. As we see n Fg. 2(c), (f), (g), and (h), the accuraces dramatcally ncrease from %(N x + N y )/2 to 1%(N x + N y )/2 CvMLs (CvCLs) Parameter study In ths experment, we study the accuraces under dfferent parameter settngs. The experment settng follows the prevous experment. The parameters k 1, k 2 are set to be the same and are chosen from [1e1, 1e2, 1e3, 1e4, 1e5]. The parameter k 3 s chosen from [1e, 1e1, 1e2, 1e3, 1e4, 1e5]. The generaton of tranng data and testng data s the same wth the prevous experment. For each parameter settng, ten trals are run and the mean accuraces are reported. Due to the space lmtaton, we only llustrate the results when the number of the CvMLs and the CvCLs are set to be 3%(N + N y )/2. We also check the results of dfferent

6 142 Q. Qan et al. / Knowledge-Based Systems 54 (13) Testng of Px, ORL Testng of LBP, ORL Number of Must lnk Number of Must lnk Number of Cannot lnk Number of Cannot lnk (a) ORL-Px (b) ORL-LBP Testng of Px, MFD Testng of Mor, MFD Number of Must lnk Number of Must lnk Number of Cannot lnk Number of Cannot lnk (c) MFD-Px (d) MFD-Mor Testng of Page, Course Testng of Lnk, Course Number of Must lnk Number of Must lnk Number of Cannot lnk Number of Cannot lnk (e) Course-Page (f) Course-Lnk Testng of En, MRC Testng of Fr, MRC Number of Must lnk Number of Must lnk Number of Cannot lnk Number of Cannot lnk (g) MRC-EN (h) MRC-FR Fg. 2. Performance of LSCVMC. numbers of the CvMLs and the CvCLs and observe the smlar results. Fg. 5 shows the heat map of the accuraces of dfferent parameter settngs. On general, the accuraces do not vares very much under dfferent parameter settngs. It mples that our framework s not qute senstve to parameters. The ORL dataset has a relatvely unstable accuraces, whch could be caused by

7 Q. Qan et al. / Knowledge-Based Systems 54 (13) Table 3 Average standard devatons ORL MFD Course MRC Vew Vew the unstable classfers learned from too lttle tranng data rather than by dfferent parameter settngs. The accuraces on the MFD dataset keep above 9% for the most of the parameters and only drop on two group of parameters. On the Course dataset, the accuraces keep hgh when nether parameters are too large or too small Comparson wth Co-RLS In ths experments we compare RLSCVMC wth Co-RLS to test the effectveness of RLSCVMC. We dd not conduct more comparson experments wth other algorthms, manly because the ntroduced CvMLs and CvCLs concepts are relatvely new. So far as we know, at present there have not had related work based on such sde nformaton yet. However, loosely speakng, Co-RLS can be vewed as a related work. Snce Co- RLS only works only on full pared data, only the ORL, the MFD and the Course dataset are used n ths experment. In ths experment, the parameters n RLSCVMC are set to the same as n the above experment. For Co-RLS, we set l = 1 and k 1 = k 2 n Eq. (11) to make the parameter settng n Co-RLS smlar 1 Vew 1 1 Vew 2 2 Vew 1 3 Vew Negatve 1 Postve Negatve 5 4 Postve 25 Postve 3 5 Negatve Negatve Must lnks, Cannot lnks (a) CvMLs=1,CvCLs= 4 Postve Must lnks, 1 Cannot lnks (b) CvMLs=,CvCLs=1 Fg. 3. Degenerated soluton of RLSCVMC when the CvMLs or the CvCLs does not exst % 1% % 3% % 5% 6% 7% 8% 9% 1% Number of MLs and CLs ORL Px MFD Px Course Page MRC En ORL LBP MFD Mor Course Lnk MRC Fr Fg. 4. Trend of ncreasng from small to large set of the CvMLs and the CvCLs.

8 144 Q. Qan et al. / Knowledge-Based Systems 54 (13) (a) ORL-Px (b) ORL-LBP (c) MFD-Px (d) MFD-Mor (e) Course-Page (f) Course-Lnk Fg. 5. Accuraces under dfferent parameter settngs. The x and y axes ndcates the two parameters n our framework. The color ndcates the accuraces. Red color means a hgh accuracy, whle blue color means a low accuracy. (For nterpretaton of the references to color n ths fgure legend, the reader s referred to the web verson of ths artcle.) to that n RLSCVMC. We select the two parameters n Co-RLS by fvefold cross valdaton both n {.1,.1, 1, 1, 1}. The means and standard devatons of accuraces are lsted n Table 4. We test the performance on dfferent portons of labeled tranng data. We frst randomly select half as tranng set and the rest as testng set. Then we choose a part of tranng set, whch ncreases from 1%N to 1%N at an nterval of 1%N, where N s the number of tranng samples, as the labeled set. The rest s used as unlabeled set. In the next, we create every possble the CvMLs and the CvCLs from the labeled tranng set for RLSCVMC algorthm. Thus the supervsed nformaton used n Co-RLS and RLSCVMC s the same. And the dfference s that Co-RLS drectly uses the labels whle RLSCVMC frst converts the labels nto the CvMLs and the CvCLs and uses them nstead. Note that, besdes

9 Q. Qan et al. / Knowledge-Based Systems 54 (13) Table 4 comparson between RLSCVMC and Co-RLS. The boldface means t-test s passed. RLSCVMC Co-RLS ORL-Px ORL-LBP ORL-Px ORL-LBP 1%.58 ± ± ± ±.144 %.55 ±.31.6 ± ± ±.148 3%.68 ± ± ± ±.119 %.71 ± ± ± ±.91 5%.737 ± ± ± ±.57 6%.87 ± ± ± ±.153 7%.827 ± ± ± ±.65 8%.882 ±.77.8 ± ± ±.55 9%.882 ± ± ±.81.8 ±.92 1%.917 ± ±.48.9 ± ±.42 MFD-Px MFD-Mor MFD-Px MFD-Mor 1%.992 ± ± ± ±.8 %.991 ± ± ± ±.7 3%.99 ± ± ± ±.7 %.988 ± ± ± ±.7 5%.988 ± ± ± ±.8 6%.988 ± ± ± ±.5 7%.991 ± ± ± ±.6 8%.991 ± ± ±.1.99 ±.6 9%.986 ± ± ± ±.7 1%.991 ± ± ± ±.6 Course-Page Course-Lnk Course-Page Course-Lnk 1%.923 ± ± ± ±.19 %.943 ± ± ±.13.9 ±.16 3%.953 ± ± ± ±.14 %.954 ± ± ± ±.13 5%.96 ± ± ± ±.12 6%.962 ± ± ± ±.8 7%.964 ± ± ± ±.11 8%.963 ± ±.1.96 ± ±.12 9%.967 ± ± ± ±.6 1%.964 ± ± ± ±.5 the label set, Co-RLS also makes full use of all the unlabeled samples and the parng nformaton on the unlabeled sample set whle RLSCVMC does not. Thus n fact, Co-RLS uses much more nformaton than RLSCVMC. In such a crcumstance, the current comparsons wll naturally be much more prone to Co-RLS than our model. Through comparng the results of both Co-RLS and RLSCVMC n Table 4, t can be wtnessed that our RLSCVMC exhbts comparable performance on two datasets, and sgnfcant performance on MFD dataset. Though so, we can fnd that current RLSCVMC can stll leave a further room for ts performance promoton snce the present expermental settng s more favorable for Co-RLS than for our model. In addton, how to utlze CvMLs and CvCLs more effectvely s stll needed. On the ORL dataset, the performance of RLSCVMC and Co-RLS s comparable. Most of the results are statstcally not sgnfcant except two (% ORL-Px, and 7% ORL-LBP). On the Px vew under % labeled samples, the accuracy of Co- RLS s hgher than RLSCVMC by over 1 percent, however ts varance s also hgh (.128) thus s not so convncng. RLSCVMC outperforms Co-RLS on the MFD dataset especally on the Px vew. On ths vew, all the results are sgnfcantly better than Co-RLS s. On the Mor vew, three of ten results are better whle the rest are comparable. Note that on the Px vew, RLSCVMC acheves hgh accuraces even under a small labeled set whle Co- RLS does not. Wth less than or equal to 3% labeled set, RLSCVMC yelds an average 4.2% hgher accuracy. On the Course dataset, RLSCVMC stll beats Co-RLS. On the Page vew, half of the accuraces of RLSCVMC are sgnfcantly hgher than Co-RLS, and the rest are comparable. On the Lnk vew, each of both algorthms obtans a sgnfcantly hgher accuracy. So, RLSCVMC, whch only uses CvMLs and CvCLs, demonstrates ts learnng ablty, and t acheves better results on the MFD and the Course datasets, and comparatve results on the ORL dataset by usng less nformaton than Co-RLS. As dscussed n Secton 3.4, RLSCVMC estmates the classfcaton loss on unlabeled data and employs the CvCLs explctly, whch makes t acheve better relatvely performance than Co-RLS. 5. Concluson and future work In ths paper, we develop a framework whch utlzes the crossvew sde nformaton, specfcally the CvMLs and the CvCLs, to learn classfers n mult-vew crcumstance where vew data are totally not pared. We show the effectveness of our framework and demonstrate why the solutons are degenerated when only the CvMLs (CvCLs) are avalable. In our comparatve experments, we observe that our framework acheves better performance than Co-RLS algorthm under the same supervson. There are stll some problems deserved to study n the future. So far, our framework only works for two-vew dataset because of the lmtaton of modelng. How to extend t to the dataset wth more than two vews s a practcal and mportant queston. Furthermore, our CvMLs and CvCLs are general mult-vew sde nformaton. They are not just lmted to classfcaton tasks. How to apply them nto mult-vew clusterng and dmenson reducton tasks under totally-unpared dataset s also deserve examnng. The performance of RLSCVML only gets one sgnfcant mprovement comparng wth Co-RLS, although the experment s more favorable for Co-RLS. There s stll a further room for the performance promoton, and ths s our next work. Acknowledgments Ths work was supported n part by the NSFC of Chna Grant Nos , Natural Scence Foundaton of the Hgher Educaton Insttutons of Jangsu under Grant No. 12KJB518 and Sponsored by Jangsu QngLan Project. Appendx A of Eq. (5) admt the represen- Theorem 5.1. The optmzers fx H; f y H tatons of the form respectvely f H x f H y ¼ Nx ¼ Ny j a k x ðx; x Þ b j k x ðy; y j Þ ð12þ ð13þ Proof. We decompose f x 2H Kx (f y 2H Ky ) nto two parts. The frst part s n the subspace spanned by kernel functons k x ðx 1 ; Þ;...; k x ðx Nx ; Þ ðk y ðy 1 ; Þ;...; k y ðy Ny ; ÞÞ, and the second part s n ts orthogonal complement. f x ¼ f xk þ f x? ¼ Nx a kðx ; xþþf x? ðf y ¼ f yk þ f y? ¼ Ny b j kðy j ; yþþf y? Þ Then we may wrte f x (x k ) and f y (y k )as f x ðx k Þ¼ Nx a kðx ; x k Þþf x? ðx k Þ¼ Nx a kðx ; x k Þþ < f x? ; kðx k ; xþ > ¼ Nx a kðx ; x k Þ ð14þ ð15þ ð16þ

10 146 Q. Qan et al. / Knowledge-Based Systems 54 (13) f y ðy k Þ¼ Ny b j kðy j ; y k Þþf y? ðy k Þ¼ Ny b j kðy j ; y k Þþ < f y? ; kðy k ; yþ > ¼ Ny j b j kðy ; y k Þ And for all f x\ and f y\ we have ð17þ kf x k 2 H x ¼kf xk þ f x? k 2 H x ¼kf xk k 2 H x þkf x? k 2 H x P kf xk k 2 H x ð18þ kf y k 2 H y ¼kf yk þ f y? k 2 H y ¼kf yk k 2 H y þkf y? k 2 H y P kf yk k 2 H y ð19þ Thus for any fxed a, b j, the functon value of objectve s mnmzed for f x\ = and f y\ =. Snce these are also solutons, the theorem holds. h References [1] Massh-Reza Amn, Ncolas Usuner, Cyrl Goutte, Learnng from multple partally observed vews an applcaton to multlngual text categorzaton, n: Advances n Neural Informaton Processng Systems (NIPS 9), 1. [2] B. Schoelkopf, A.J. Smola, Learnng wth Kernels, MIT Press, Cambrdge, MA, 2. [3] S. Basu, A. Banerjee, R.J. Mooney, Actve sem-supervson for parwse constraned clusterng, n Proc. of the SIAM nternatonal conference on data mnng (SDM 4), 4. [4] M. Belkn, P. Nyog, V. Sndhwan, Manfold regularzaton: a geometrc framework for learnng from labeled and unlabeled examples, The Journal of Machne Learnng Research 7 (6) [5] D.P. Bertsekas, W.W. Hager, O.L. Mangasaran, Nonlnear Programmng, Athena Scentfc Belmont, MA, [6] M. Blaschko, C. Lampert, A. Gretton, Sem-supervsed Laplacan regularzaton of kernel canoncal correlaton analyss, Machne Learnng and Knowledge Dscovery n Databases (8) [7] A. Blum, T. Mtchell, Combnng labeled and unlabeled data wth co-tranng, n: Proc. of the 11th Annual Conference on Computatonal Learnng Theory (COLT 1998), [8] U. Brefeld, T. Scheffer, Co-em support vector learnng, n: Proc. of the 21st Internatonal Conference on Machne Learnng (ICML 4), 4. [9] S. Dasgupta, M.L. Lttman, D. McAllester, Pac generalzaton bounds for cotranng, n: Advances n Neural Informaton Processng Systems (NIPS 2), 2. [1] J.V. Davs, B. Kuls, P. Jan, S. Sra, I.S. Dhllon, Informaton-theoretc metrc learnng, n: Proceedngs of the 24th Internatonal Conference on Machne learnng (ICML7), 7. [11] R.O. Duda, P.E. Hart, D.G. Stork, Pattern Classfcaton, Wley, New York, 1. [12] A. Frank, A. Asuncon, UCI Machne Learnng Repostory, 1. [13] A. Globerson, S. Rowes, Metrc learnng by collapsng classes, n: Advances n Neural Informaton Processng Systems (NIPS6), 5. [14] A.B. Goldberg,. Zhu, S. Wrght, Dssmlarty n graph-based sem-supervsed classfcaton, n: Proc. of the 11th Artfcal Intellgence and Statstcs (AISTATS 7), 7. [15] Jngjng Gu, Songcan Chen, Tngka Sun, Localzaton wth ncompletely pared data n complex wreless sensor network, IEEE Transactons on Wreless Communcatons (11). [16] T. Hertz, N. Shental, A. Bar-Hllel, D. Wenshall, Enhancng mage and vdeo retreval: learnng va equvalence constrants, n: IEEE Computer Socety Conference on Computer Vson and Pattern Recognton (CVPR 3), 3. [17] E. Hu, S. Chen, D. Zhang,. Yn, Semsupervsed kernel matrx learnng by kernel propagaton, IEEE Transactons on Neural Networks 21 (1) [18] A. Kmura, H. Kameoka, M. Sugyama, T. Nakano, E. Maeda, H. Sakano, K. Ishguro, Semcca: Effcent sem-supervsed learnng of canoncal correlatons, n: Proc. of the th Internatonal Conference on Pattern Recognton (ICPR 1), 1. [19] C. Lampert, O. Kromer, Weakly-pared maxmum covarance analyss for multmodal dmensonalty reducton and transfer learnng, n: Proc. of the 11th European Conference on Computer Vson (ECCV 1), 1. [] G. L, S.C.H. Ho, K. Chang, Two-vew transductve support vector machnes, n: Proc. of the SIAM Internatonal Conference of Data Mnng (SDM 1), 1. [21] Z. L, J. Lu,. Tang, Parwse constrant propagaton by semdefnte programmng for sem-supervsed classfcaton, n: Proc. of the 25th Internatonal Conference on Machne Learnng (ICML 8), 8. [22] N. Nguyen, R. Caruana, Improvng classfcaton wth parwse constrants: a margn-based approach, Machne Learnng and Knowledge Dscovery n Databases (8) [23] S. Shalev-Shwartz, Y. Snger, A.Y. Ng, Onlne and batch learnng of pseudometrcs, n: Proceedngs of the 21st Internatonal Conference on Machne Learnng (ICML4), 4. [24] V. Sndhwan, P. Nyog, M. Belkn, A co-regularzaton approach to semsupervsed learnng wth multple vews, n: Workshop on Learnng wth Multple Vews at ICML, 5. [25] T. Sun, S. Chen, J. Yang,. Hu, P. Sh, Dscrmnatve canoncal correlaton analyss wth mssng samples, n: Computer Scence and Informaton Engneerng (CSIE 9), 9. [26] W. Tong, R. Jn, Sem-supervsed learnng by mxed label propagaton, n: Proc. of the 22th Natonal Conference on Artfcal Intellgence (AAAI 7), 7. [27] V.N. Vapnk, Statstcal Learnng Theory, Wley-Interscence, [28] K. Wagstaff, C. Carde, S. Rogers, S. Schrodl, Constraned k-means clusterng wth background knowledge, n: Proc. of the 18th Internatonal Conference on Machne Learnng (ICML 1), 1. [29] E.P. ng, A.Y. Ng, M.I. Jordan, S. Russell, Dstance metrc learnng, wth applcaton to clusterng wth sde-nformaton, n: Advances n Neural Informaton Processng Systems (NIPS2), 2. [3] R. Yan, J. Zhang, J. Yang, A.G. Hauptmann, A dscrmnatve learnng framework wth parwse constrants for vdeo object classfcaton, IEEE Transactons on Pattern Analyss and Machne Intellgence (6) [31] D. Yarowsky, Unsupervsed word sense dsambguaton rvalng supervsed methods, n: Proc. of the 33rd Annual Meetng on Assocaton for Computatonal Lngustcs (ACL 1995), [32] T. Yoshda, K. Okatan, A graph-based projecton approach for sem-supervsed clusterng, Knowledge Management and Acquston for Smart Systems and Servces (11) [33] S. Yu, J. Sh, Groupng wth drected relatonshps, n: Energy Mnmzaton Methods n Computer Vson and Pattern Recognton, 1. [34] J. Zhang, R. Yan, On the value of parwse constrants n classfcaton and consstency, n: Proc. of the 24th Internatonal Conference on Machne Learnng (ICML 7), 7.

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

Classification / Regression Support Vector Machines

Classification / Regression Support Vector Machines Classfcaton / Regresson Support Vector Machnes Jeff Howbert Introducton to Machne Learnng Wnter 04 Topcs SVM classfers for lnearly separable classes SVM classfers for non-lnearly separable classes SVM

More information

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for

More information

Discriminative Dictionary Learning with Pairwise Constraints

Discriminative Dictionary Learning with Pairwise Constraints Dscrmnatve Dctonary Learnng wth Parwse Constrants Humn Guo Zhuoln Jang LARRY S. DAVIS UNIVERSITY OF MARYLAND Nov. 6 th, Outlne Introducton/motvaton Dctonary Learnng Dscrmnatve Dctonary Learnng wth Parwse

More information

Face Recognition University at Buffalo CSE666 Lecture Slides Resources:

Face Recognition University at Buffalo CSE666 Lecture Slides Resources: Face Recognton Unversty at Buffalo CSE666 Lecture Sldes Resources: http://www.face-rec.org/algorthms/ Overvew of face recognton algorthms Correlaton - Pxel based correspondence between two face mages Structural

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS46: Mnng Massve Datasets Jure Leskovec, Stanford Unversty http://cs46.stanford.edu /19/013 Jure Leskovec, Stanford CS46: Mnng Massve Datasets, http://cs46.stanford.edu Perceptron: y = sgn( x Ho to fnd

More information

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

Announcements. Supervised Learning

Announcements. Supervised Learning Announcements See Chapter 5 of Duda, Hart, and Stork. Tutoral by Burge lnked to on web page. Supervsed Learnng Classfcaton wth labeled eamples. Images vectors n hgh-d space. Supervsed Learnng Labeled eamples

More information

Positive Semi-definite Programming Localization in Wireless Sensor Networks

Positive Semi-definite Programming Localization in Wireless Sensor Networks Postve Sem-defnte Programmng Localzaton n Wreless Sensor etworks Shengdong Xe 1,, Jn Wang, Aqun Hu 1, Yunl Gu, Jang Xu, 1 School of Informaton Scence and Engneerng, Southeast Unversty, 10096, anjng Computer

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

Support Vector Machines. CS534 - Machine Learning

Support Vector Machines. CS534 - Machine Learning Support Vector Machnes CS534 - Machne Learnng Perceptron Revsted: Lnear Separators Bnar classfcaton can be veed as the task of separatng classes n feature space: b > 0 b 0 b < 0 f() sgn( b) Lnear Separators

More information

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET 1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School

More information

Semi-Supervised Discriminant Analysis Based On Data Structure

Semi-Supervised Discriminant Analysis Based On Data Structure IOSR Journal of Computer Engneerng (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 3, Ver. VII (May Jun. 2015), PP 39-46 www.osrournals.org Sem-Supervsed Dscrmnant Analyss Based On Data

More information

Collaboratively Regularized Nearest Points for Set Based Recognition

Collaboratively Regularized Nearest Points for Set Based Recognition Academc Center for Computng and Meda Studes, Kyoto Unversty Collaboratvely Regularzed Nearest Ponts for Set Based Recognton Yang Wu, Mchhko Mnoh, Masayuk Mukunok Kyoto Unversty 9/1/013 BMVC 013 @ Brstol,

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints Australan Journal of Basc and Appled Scences, 2(4): 1204-1208, 2008 ISSN 1991-8178 Sum of Lnear and Fractonal Multobjectve Programmng Problem under Fuzzy Rules Constrants 1 2 Sanjay Jan and Kalash Lachhwan

More information

A New Approach For the Ranking of Fuzzy Sets With Different Heights

A New Approach For the Ranking of Fuzzy Sets With Different Heights New pproach For the ankng of Fuzzy Sets Wth Dfferent Heghts Pushpnder Sngh School of Mathematcs Computer pplcatons Thapar Unversty, Patala-7 00 Inda pushpndersnl@gmalcom STCT ankng of fuzzy sets plays

More information

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1)

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1) Secton 1.2 Subsets and the Boolean operatons on sets If every element of the set A s an element of the set B, we say that A s a subset of B, or that A s contaned n B, or that B contans A, and we wrte A

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law) Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes

More information

TN348: Openlab Module - Colocalization

TN348: Openlab Module - Colocalization TN348: Openlab Module - Colocalzaton Topc The Colocalzaton module provdes the faclty to vsualze and quantfy colocalzaton between pars of mages. The Colocalzaton wndow contans a prevew of the two mages

More information

Geometry-aware Metric Learning

Geometry-aware Metric Learning Zhengdong Lu LUZ@CS.UTEXAS.EDU Prateek Jan PJAIN@CS.UTEXAS.EDU Inderjt S. Dhllon INDERJIT@CS.UTEXAS.EDU Dept. of Computer Scence, Unversty of Texas at Austn, Unversty Staton C5, Austn, TX 78712 Abstract

More information

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour 6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

LECTURE : MANIFOLD LEARNING

LECTURE : MANIFOLD LEARNING LECTURE : MANIFOLD LEARNING Rta Osadchy Some sldes are due to L.Saul, V. C. Raykar, N. Verma Topcs PCA MDS IsoMap LLE EgenMaps Done! Dmensonalty Reducton Data representaton Inputs are real-valued vectors

More information

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION

CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 48 CHAPTER 3 SEQUENTIAL MINIMAL OPTIMIZATION TRAINED SUPPORT VECTOR CLASSIFIER FOR CANCER PREDICTION 3.1 INTRODUCTION The raw mcroarray data s bascally an mage wth dfferent colors ndcatng hybrdzaton (Xue

More information

Edge Detection in Noisy Images Using the Support Vector Machines

Edge Detection in Noisy Images Using the Support Vector Machines Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

Mathematics 256 a course in differential equations for engineering students

Mathematics 256 a course in differential equations for engineering students Mathematcs 56 a course n dfferental equatons for engneerng students Chapter 5. More effcent methods of numercal soluton Euler s method s qute neffcent. Because the error s essentally proportonal to the

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

Semi-supervised Classification Using Local and Global Regularization

Semi-supervised Classification Using Local and Global Regularization Proceedngs of the Twenty-Thrd AAAI Conference on Artfcal Intellgence (2008) Sem-supervsed Classfcaton Usng Local and Global Regularzaton Fe Wang 1, Tao L 2, Gang Wang 3, Changshu Zhang 1 1 Department of

More information

Learning a Class-Specific Dictionary for Facial Expression Recognition

Learning a Class-Specific Dictionary for Facial Expression Recognition BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 16, No 4 Sofa 016 Prnt ISSN: 1311-970; Onlne ISSN: 1314-4081 DOI: 10.1515/cat-016-0067 Learnng a Class-Specfc Dctonary for

More information

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng

More information

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces Range mages For many structured lght scanners, the range data forms a hghly regular pattern known as a range mage. he samplng pattern s determned by the specfc scanner. Range mage regstraton 1 Examples

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems A Unfed Framework for Semantcs and Feature Based Relevance Feedback n Image Retreval Systems Ye Lu *, Chunhu Hu 2, Xngquan Zhu 3*, HongJang Zhang 2, Qang Yang * School of Computng Scence Smon Fraser Unversty

More information

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION Paulo Quntlano 1 & Antono Santa-Rosa 1 Federal Polce Department, Brasla, Brazl. E-mals: quntlano.pqs@dpf.gov.br and

More information

Solving two-person zero-sum game by Matlab

Solving two-person zero-sum game by Matlab Appled Mechancs and Materals Onlne: 2011-02-02 ISSN: 1662-7482, Vols. 50-51, pp 262-265 do:10.4028/www.scentfc.net/amm.50-51.262 2011 Trans Tech Publcatons, Swtzerland Solvng two-person zero-sum game by

More information

Transductive Regression Piloted by Inter-Manifold Relations

Transductive Regression Piloted by Inter-Manifold Relations Huan Wang IE, The Chnese Unversty of Hong Kong, Hong Kong Shucheng Yan Thomas Huang ECE, Unversty of Illnos at Urbana Champagn, USA Janzhuang Lu Xaoou Tang IE, The Chnese Unversty of Hong Kong, Hong Kong

More information

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization Problem efntons and Evaluaton Crtera for Computatonal Expensve Optmzaton B. Lu 1, Q. Chen and Q. Zhang 3, J. J. Lang 4, P. N. Suganthan, B. Y. Qu 6 1 epartment of Computng, Glyndwr Unversty, UK Faclty

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

SVM-based Learning for Multiple Model Estimation

SVM-based Learning for Multiple Model Estimation SVM-based Learnng for Multple Model Estmaton Vladmr Cherkassky and Yunqan Ma Department of Electrcal and Computer Engneerng Unversty of Mnnesota Mnneapols, MN 55455 {cherkass,myq}@ece.umn.edu Abstract:

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervsed Learnng and Clusterng Why consder unlabeled samples?. Collectng and labelng large set of samples s costly Gettng recorded speech s free, labelng s tme consumng 2. Classfer could be desgned

More information

Three supervised learning methods on pen digits character recognition dataset

Three supervised learning methods on pen digits character recognition dataset Three supervsed learnng methods on pen dgts character recognton dataset Chrs Flezach Department of Computer Scence and Engneerng Unversty of Calforna, San Dego San Dego, CA 92093 cflezac@cs.ucsd.edu Satoru

More information

Module Management Tool in Software Development Organizations

Module Management Tool in Software Development Organizations Journal of Computer Scence (5): 8-, 7 ISSN 59-66 7 Scence Publcatons Management Tool n Software Development Organzatons Ahmad A. Al-Rababah and Mohammad A. Al-Rababah Faculty of IT, Al-Ahlyyah Amman Unversty,

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

A Robust LS-SVM Regression

A Robust LS-SVM Regression PROCEEDIGS OF WORLD ACADEMY OF SCIECE, EGIEERIG AD ECHOLOGY VOLUME 7 AUGUS 5 ISS 37- A Robust LS-SVM Regresson József Valyon, and Gábor Horváth Abstract In comparson to the orgnal SVM, whch nvolves a quadratc

More information

Load Balancing for Hex-Cell Interconnection Network

Load Balancing for Hex-Cell Interconnection Network Int. J. Communcatons, Network and System Scences,,, - Publshed Onlne Aprl n ScRes. http://www.scrp.org/journal/jcns http://dx.do.org/./jcns.. Load Balancng for Hex-Cell Interconnecton Network Saher Manaseer,

More information

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 A mathematcal programmng approach to the analyss, desgn and

More information

Face Recognition Based on SVM and 2DPCA

Face Recognition Based on SVM and 2DPCA Vol. 4, o. 3, September, 2011 Face Recognton Based on SVM and 2DPCA Tha Hoang Le, Len Bu Faculty of Informaton Technology, HCMC Unversty of Scence Faculty of Informaton Scences and Engneerng, Unversty

More information

Semi-Supervised Kernel Mean Shift Clustering

Semi-Supervised Kernel Mean Shift Clustering IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. XX, NO. XX, JANUARY XXXX 1 Sem-Supervsed Kernel Mean Shft Clusterng Saket Anand, Student Member, IEEE, Sushl Mttal, Member, IEEE, Oncel

More information

Discriminative classifiers for object classification. Last time

Discriminative classifiers for object classification. Last time Dscrmnatve classfers for object classfcaton Thursday, Nov 12 Krsten Grauman UT Austn Last tme Supervsed classfcaton Loss and rsk, kbayes rule Skn color detecton example Sldng ndo detecton Classfers, boostng

More information

Online Detection and Classification of Moving Objects Using Progressively Improving Detectors

Online Detection and Classification of Moving Objects Using Progressively Improving Detectors Onlne Detecton and Classfcaton of Movng Objects Usng Progressvely Improvng Detectors Omar Javed Saad Al Mubarak Shah Computer Vson Lab School of Computer Scence Unversty of Central Florda Orlando, FL 32816

More information

Quality Improvement Algorithm for Tetrahedral Mesh Based on Optimal Delaunay Triangulation

Quality Improvement Algorithm for Tetrahedral Mesh Based on Optimal Delaunay Triangulation Intellgent Informaton Management, 013, 5, 191-195 Publshed Onlne November 013 (http://www.scrp.org/journal/m) http://dx.do.org/10.36/m.013.5601 Qualty Improvement Algorthm for Tetrahedral Mesh Based on

More information

Local Quaternary Patterns and Feature Local Quaternary Patterns

Local Quaternary Patterns and Feature Local Quaternary Patterns Local Quaternary Patterns and Feature Local Quaternary Patterns Jayu Gu and Chengjun Lu The Department of Computer Scence, New Jersey Insttute of Technology, Newark, NJ 0102, USA Abstract - Ths paper presents

More information

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics

NAG Fortran Library Chapter Introduction. G10 Smoothing in Statistics Introducton G10 NAG Fortran Lbrary Chapter Introducton G10 Smoothng n Statstcs Contents 1 Scope of the Chapter... 2 2 Background to the Problems... 2 2.1 Smoothng Methods... 2 2.2 Smoothng Splnes and Regresson

More information

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation

An Iterative Solution Approach to Process Plant Layout using Mixed Integer Optimisation 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 An Iteratve Soluton Approach to Process Plant Layout usng Mxed

More information

A Semi-parametric Regression Model to Estimate Variability of NO 2

A Semi-parametric Regression Model to Estimate Variability of NO 2 Envronment and Polluton; Vol. 2, No. 1; 2013 ISSN 1927-0909 E-ISSN 1927-0917 Publshed by Canadan Center of Scence and Educaton A Sem-parametrc Regresson Model to Estmate Varablty of NO 2 Meczysław Szyszkowcz

More information

Adaptive Transfer Learning

Adaptive Transfer Learning Adaptve Transfer Learnng Bn Cao, Snno Jaln Pan, Yu Zhang, Dt-Yan Yeung, Qang Yang Hong Kong Unversty of Scence and Technology Clear Water Bay, Kowloon, Hong Kong {caobn,snnopan,zhangyu,dyyeung,qyang}@cse.ust.hk

More information

Efficient Text Classification by Weighted Proximal SVM *

Efficient Text Classification by Weighted Proximal SVM * Effcent ext Classfcaton by Weghted Proxmal SVM * Dong Zhuang 1, Benyu Zhang, Qang Yang 3, Jun Yan 4, Zheng Chen, Yng Chen 1 1 Computer Scence and Engneerng, Bejng Insttute of echnology, Bejng 100081, Chna

More information

Using Neural Networks and Support Vector Machines in Data Mining

Using Neural Networks and Support Vector Machines in Data Mining Usng eural etworks and Support Vector Machnes n Data Mnng RICHARD A. WASIOWSKI Computer Scence Department Calforna State Unversty Domnguez Hlls Carson, CA 90747 USA Abstract: - Multvarate data analyss

More information

Hierarchical clustering for gene expression data analysis

Hierarchical clustering for gene expression data analysis Herarchcal clusterng for gene expresson data analyss Gorgo Valentn e-mal: valentn@ds.unm.t Clusterng of Mcroarray Data. Clusterng of gene expresson profles (rows) => dscovery of co-regulated and functonally

More information

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY 1. SSDH: Semi-supervised Deep Hashing for Large Scale Image Retrieval

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY 1. SSDH: Semi-supervised Deep Hashing for Large Scale Image Retrieval IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY SSDH: Sem-supervsed Deep Hashng for Large Scale Image Retreval Jan Zhang, and Yuxn Peng arxv:607.08477v2 [cs.cv] 8 Jun 207 Abstract Hashng

More information

K-means and Hierarchical Clustering

K-means and Hierarchical Clustering Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your

More information

y and the total sum of

y and the total sum of Lnear regresson Testng for non-lnearty In analytcal chemstry, lnear regresson s commonly used n the constructon of calbraton functons requred for analytcal technques such as gas chromatography, atomc absorpton

More information

Human Face Recognition Using Generalized. Kernel Fisher Discriminant

Human Face Recognition Using Generalized. Kernel Fisher Discriminant Human Face Recognton Usng Generalzed Kernel Fsher Dscrmnant ng-yu Sun,2 De-Shuang Huang Ln Guo. Insttute of Intellgent Machnes, Chnese Academy of Scences, P.O.ox 30, Hefe, Anhu, Chna. 2. Department of

More information

Active Contours/Snakes

Active Contours/Snakes Actve Contours/Snakes Erkut Erdem Acknowledgement: The sldes are adapted from the sldes prepared by K. Grauman of Unversty of Texas at Austn Fttng: Edges vs. boundares Edges useful sgnal to ndcate occludng

More information

X- Chart Using ANOM Approach

X- Chart Using ANOM Approach ISSN 1684-8403 Journal of Statstcs Volume 17, 010, pp. 3-3 Abstract X- Chart Usng ANOM Approach Gullapall Chakravarth 1 and Chaluvad Venkateswara Rao Control lmts for ndvdual measurements (X) chart are

More information

Parallel Numerics. 1 Preconditioning & Iterative Solvers (From 2016)

Parallel Numerics. 1 Preconditioning & Iterative Solvers (From 2016) Technsche Unverstät München WSe 6/7 Insttut für Informatk Prof. Dr. Thomas Huckle Dpl.-Math. Benjamn Uekermann Parallel Numercs Exercse : Prevous Exam Questons Precondtonng & Iteratve Solvers (From 6)

More information

A Novel Adaptive Descriptor Algorithm for Ternary Pattern Textures

A Novel Adaptive Descriptor Algorithm for Ternary Pattern Textures A Novel Adaptve Descrptor Algorthm for Ternary Pattern Textures Fahuan Hu 1,2, Guopng Lu 1 *, Zengwen Dong 1 1.School of Mechancal & Electrcal Engneerng, Nanchang Unversty, Nanchang, 330031, Chna; 2. School

More information

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr)

Helsinki University Of Technology, Systems Analysis Laboratory Mat Independent research projects in applied mathematics (3 cr) Helsnk Unversty Of Technology, Systems Analyss Laboratory Mat-2.08 Independent research projects n appled mathematcs (3 cr) "! #$&% Antt Laukkanen 506 R ajlaukka@cc.hut.f 2 Introducton...3 2 Multattrbute

More information

Wishing you all a Total Quality New Year!

Wishing you all a Total Quality New Year! Total Qualty Management and Sx Sgma Post Graduate Program 214-15 Sesson 4 Vnay Kumar Kalakband Assstant Professor Operatons & Systems Area 1 Wshng you all a Total Qualty New Year! Hope you acheve Sx sgma

More information

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes SPH3UW Unt 7.3 Sphercal Concave Mrrors Page 1 of 1 Notes Physcs Tool box Concave Mrror If the reflectng surface takes place on the nner surface of the sphercal shape so that the centre of the mrror bulges

More information

Detection of an Object by using Principal Component Analysis

Detection of an Object by using Principal Component Analysis Detecton of an Object by usng Prncpal Component Analyss 1. G. Nagaven, 2. Dr. T. Sreenvasulu Reddy 1. M.Tech, Department of EEE, SVUCE, Trupath, Inda. 2. Assoc. Professor, Department of ECE, SVUCE, Trupath,

More information

Fitting: Deformable contours April 26 th, 2018

Fitting: Deformable contours April 26 th, 2018 4/6/08 Fttng: Deformable contours Aprl 6 th, 08 Yong Jae Lee UC Davs Recap so far: Groupng and Fttng Goal: move from array of pxel values (or flter outputs) to a collecton of regons, objects, and shapes.

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

General Vector Machine. Hong Zhao Department of Physics, Xiamen University

General Vector Machine. Hong Zhao Department of Physics, Xiamen University General Vector Machne Hong Zhao (zhaoh@xmu.edu.cn) Department of Physcs, Xamen Unversty The support vector machne (SVM) s an mportant class of learnng machnes for functon approach, pattern recognton, and

More information

Backpropagation: In Search of Performance Parameters

Backpropagation: In Search of Performance Parameters Bacpropagaton: In Search of Performance Parameters ANIL KUMAR ENUMULAPALLY, LINGGUO BU, and KHOSROW KAIKHAH, Ph.D. Computer Scence Department Texas State Unversty-San Marcos San Marcos, TX-78666 USA ae049@txstate.edu,

More information

CLUSTERING that discovers the relationship among data

CLUSTERING that discovers the relationship among data Ramp-based Twn Support Vector Clusterng Zhen Wang, Xu Chen, Chun-Na L, and Yuan-Ha Shao arxv:82.0370v [cs.lg] 0 Dec 208 Abstract Tradtonal plane-based clusterng methods measure the cost of wthn-cluster

More information

Taxonomy of Large Margin Principle Algorithms for Ordinal Regression Problems

Taxonomy of Large Margin Principle Algorithms for Ordinal Regression Problems Taxonomy of Large Margn Prncple Algorthms for Ordnal Regresson Problems Amnon Shashua Computer Scence Department Stanford Unversty Stanford, CA 94305 emal: shashua@cs.stanford.edu Anat Levn School of Computer

More information

Skew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach

Skew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach Angle Estmaton and Correcton of Hand Wrtten, Textual and Large areas of Non-Textual Document Images: A Novel Approach D.R.Ramesh Babu Pyush M Kumat Mahesh D Dhannawat PES Insttute of Technology Research

More information

CLASSIFICATION OF ULTRASONIC SIGNALS

CLASSIFICATION OF ULTRASONIC SIGNALS The 8 th Internatonal Conference of the Slovenan Socety for Non-Destructve Testng»Applcaton of Contemporary Non-Destructve Testng n Engneerng«September -3, 5, Portorož, Slovena, pp. 7-33 CLASSIFICATION

More information

Laplacian Eigenmap for Image Retrieval

Laplacian Eigenmap for Image Retrieval Laplacan Egenmap for Image Retreval Xaofe He Partha Nyog Department of Computer Scence The Unversty of Chcago, 1100 E 58 th Street, Chcago, IL 60637 ABSTRACT Dmensonalty reducton has been receved much

More information

Incremental Learning with Support Vector Machines and Fuzzy Set Theory

Incremental Learning with Support Vector Machines and Fuzzy Set Theory The 25th Workshop on Combnatoral Mathematcs and Computaton Theory Incremental Learnng wth Support Vector Machnes and Fuzzy Set Theory Yu-Mng Chuang 1 and Cha-Hwa Ln 2* 1 Department of Computer Scence and

More information

Proper Choice of Data Used for the Estimation of Datum Transformation Parameters

Proper Choice of Data Used for the Estimation of Datum Transformation Parameters Proper Choce of Data Used for the Estmaton of Datum Transformaton Parameters Hakan S. KUTOGLU, Turkey Key words: Coordnate systems; transformaton; estmaton, relablty. SUMMARY Advances n technologes and

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration Improvement of Spatal Resoluton Usng BlockMatchng Based Moton Estmaton and Frame Integraton Danya Suga and Takayuk Hamamoto Graduate School of Engneerng, Tokyo Unversty of Scence, 6-3-1, Nuku, Katsuska-ku,

More information