Linear Cross-Modal Hashing for Efficient Multimedia Search

Size: px
Start display at page:

Download "Linear Cross-Modal Hashing for Efficient Multimedia Search"

Transcription

1 Lnear Cross-Modal Hashng for Effcent Multmeda Search Xaofeng Zhu Z Huang Heng Tao Shen Xn Zhao College of CSIT, Guangx Normal Unversty, Guangx, 544,P.R.Chna School of ITEE, The Unversty of Queensland, QLD 472, Australa {zhux,huang,shenht}@tee.u.edu.au, x.zhao@u.edu.au ABSTRACT Most exstng cross-modal hashng methods suffer from the scalablty ssue n the tranng phase. In ths paper, we propose a novel cross-modal hashng approach wth a lnear tme complexty to the tranng data sze, to enable scalable ndexng for multmeda search across multple modals. Takng both the ntra-smlarty n each modal and the ntersmlarty across dfferent modals nto consderaton, the proposed approach ams at effectvely learnng hash functons from large-scale tranng datasets. More specfcally, for each modal, we frst partton the tranng data nto k clusters and then represent each tranng data pont wth ts dstances to k centrods of the clusters. Interestngly, such a k-dmensonal data representaton can reduce the tme complexty of the tranng phase from tradtonal O(n 2 )orhgher to O(n), where n s the tranng data sze, leadng to practcal learnng on large-scale datasets. We further prove that ths new representaton preserves the ntra-smlarty n each modal. To preserve the nter-smlarty among data ponts across dfferent modals, we transform the derved data representatons nto a common bnary subspace n whch bnary codes from all the modals are consstent and comparable. The transformaton smultaneously outputs the hash functons for all modals, whch are used to convert unseen data nto bnary codes. Gven a uery of one modal, t s frst mapped nto the bnary codes usng the modal s hash functons, followed by matchng the database bnary codes of any other modals. Expermental results on two benchmark datasets confrm the scalablty and the effectveness of the proposed approach n comparson wth the state of the art. Categores and Subject Descrptors H.3. [Content Analyss and Indexng]: Indexng Methods; H.3.3 [Informaton Search and Retreval]: Search Process Permsson to make dgtal or hard copes of all or part of ths work for personal or classroom use s granted wthout fee provded that copes are not made or dstrbuted for proft or commercal advantage and that copes bear ths notce and the full ctaton on the frst page. Copyrghts for components of ths work owned by others than ACM must be honored. Abstractng wth credt s permtted. To copy otherwse, or republsh, to post on servers or to redstrbute to lsts, reures pror specfc permsson and/or a fee. Reuest permssons from permssons@acm.org. MM 3, October 2 25, 23, Barcelona, Span. Copyrght 23 ACM /3/ $5.. Keywords Cross-modal, hashng, ndex, multmeda search. INTRODUCTION Hashng s ncreasngly popular to support approxmate nearest neghbor (ANN) search from multmeda data. The dea of hashng for ANN search s to learn the hash functons for convertng hgh-dmensonal data nto short bnary codes whle preservng the neghborhood relatonshps of orgnal data as much as possble [3, 5, 2, 3]. It has been shown that hash functon learnng (HFL) s the key process for effectve hashng [3, 2]. Extng hashng methods on sngle modal data (referred to un-modal hashng methods n ths paper) can be categorzed nto LSH-lke hashng (e.g., localty senstve hashng (LSH) [7, 8], KLSH [5], and SKLSH [2]) whch randomly selects lnear functons as hash functons, PCA-lke hashng (e.g., SH [33], PCAH [3], and ITQ []) whch uses the prncpal components of tranng data to learn hash functons, and manfold-lke hashng (e.g., MFH [26] and [34]) whch employs manfold learnng technues to learn hash functons. More recently, some hashng methods have been proposed to ndex data represented by multple modals (referred to mult-modal hashng n ths paper) [26, 36], whch can be used to facltate the retreval for data descrbed by multple modals n many real-lfe applcatons, such as nearduplcate mage retreval. Consderng an mage database where each mage s descrbed by multple modals, such as SIFT descrptor, color hstogram, bag-of-word, etc, multmodal hashng learns hash functons from all the modals to support effectve mage retreval, where the smlartes from all the modals are consdered n rankng the fnal results wth respect to a mult-modal uery. Cross-modal hashng also constructs hash functons from all the modals by analyzng ther correlatons. However, t serves for a dfferent purpose,.e., supportng cross-modal retreval where a uery of one modal can search for the relevant results of another modal [2, 6, 22, 37, 38]. For example, gven a uery descrbed by SIFT descrptor, relevant results descrbed by other modals such as color hstogram and bag-of-word can also be found and returned 2. Modal, feature and vew are often used wth subtle dfferences n multmeda research. In ths paper, we consstently use the term modal. 2 In ths sense, cross-modal retreval s defned more generally than tradtonal cross-meda retreval [35] where ueres and results can be of dfferent meda types, such as text document, mage, vdeo, and audo. Area Char: Roelof van Zwol 43

2 Offlne process Onlne process Tranng data Database () m m2 m3 m4 m (4) Database bnary codes (2) (2) (5) m m2 m3 m (3) m Hash functons Hash functons Query mage Query bnary codes Text results Fgure : Flowchart of the proposed lnear cross-modal hashng (). tranng phase becomes lnear to the tranng data sze,.e., O(n). To acheve hgh ualty hash functons, also preserves both the ntra-smlarty among data ponts n each modal and the nter-smlarty among data ponts across dfferent modals. The learnt hash functons ensure that all the data ponts descrbed by dfferent modals n the common bnary subspace are consstent (.e., relevant data of dfferent modals should have smlar bnary codes) and comparable (.e., bnary codes of dfferent modals can be drectly compared). Fg. llustrates the whole flowchart of the proposed. The tranng phase of s an offlne process and ncludes five key steps. In the first step, for each modal we partton ts data nto k clusters. In the second step, we represent each tranng data pont wth ts dstances to the k clusters centrods. In the thrd step, hash functons are learnt efficently wth a lnear tme complexty and effectvely wth the ntra- and nter-smlarty preservatons. In the fourth step, all the data ponts n the database are approxmated wth k-dmensonal representatons, whch are then mapped nto the bnary codes wth the learn hash functons n the fifth step. In the onlne search process, a uery of one modal s first approxmated wth ts k-dmensonal representaton n ths modal whch s then mapped nto the uery bnary codes wth the hash functons for ths modal, followed by matchng the database bnary codes to find relevant results of any other modal. Extensve expermental results on two benchmark datasets confirm the scalablty and the effectveness of the proposed approach n comparson wth the state of the art. The rest of the paper s organzed as follows. Related work s revewed n Secton 2. The proposed and ts analyss are presented n Secton 3. Secton 4 reports the results and the paper s concluded n Secton 5. Whle few attempts have been made towards effectve cross-modal hashng, most exstng cross-modal hashng methods [6, 22, 27, 37, 38] suffer from hgh tme complexty n the tranng phase (.e., O(n2 ) or hgher, where n s the tranng data sze) and thus fal to learn from large-scale tranng datasets n practcal amount of tme. Such a hgh complexty constrans the above methods from applcatons dealng wth large-scale datasets. For example, mult-modal latent bnary embeddng () [38] s a generatve model such that only a small-szed tranng dataset (e.g., 3 out of 8, data ponts) can be used n the tranng phase. Although cross-modal smlarty senstve hashng () [2] s able to learn from large-scale tranng datasets, t reures pror knowledge (.e., postve pars and negatve pars among tranng data ponts) to be predefined and known, whch s not practcal n most real-lfe applcatons. To enable cross-modal retreval, nter-meda hashng (IMH) [27] explores the correlatons among multple modals from dfferent data sources and acheves better hashng performance, but the tran process of IMH wth the tme complexty O(n3 ) s expensve for large-scale cross-modal hashng. In ths paper, we propose a novel hashng method, named lnear cross-modal hashng (), to address the scalablty ssues wthout usng any pror knowledge. acheves a lnear tme complexty to the tranng data sze n the tranng phase, enablng effectve learnng from largescale datasets. The key dea s to first partton the tranng data of each modal nto k clusters by applyng a lnear tme clusterng method, and then represent each tranng data pont usng ts dstances to the k clusters centrods. That s, we approxmate each data pont wth a k-dmensonal representaton. Interestngly, such a representaton leads to the tme complexty of O(kn) for the tranng phase. Gven a really large-scale tranng dataset, t s expected that k n. Snce k s a constant, the overall tme complexty of the 44

3 2. RELATED WORK In ths secton we revew exstng hashng methods n three major categores, ncludng un-modal hashng, mult-modal hashng and cross-modal hashng. In un-modal hashng, early work such as LSH-lke hashng methods [7, 8, 5, 2] construct hash functons based on random projectons and are typcally unsupervsed. Although they have some asymptotc theoretcal propertes, LSH-lke hashng methods often reure long bnary codes and multple hash tables to acheve reasonable retreval accuracy [2]. Ths leads to long uery tme and hgh storage cost. Recently machne learnng technues have been appled to mprove hashng performance. For example, PCAlke hashng [, 3, 33] learns hash functons va preservng the maxmal covarance of orgnal data and has been shown to outperform LSH-lke hashng n [4, 7, 29]. Manfoldlke hashng [8, 26] employs manfold learnng technues to learn hash functons. Besdes, some hashng methods conduct hash functon learnng by makng the best use of pror knowledge of data. For example, supervsed hashng methods [4, 7, 9, 24, 28] mprove the hashng performance usng the pre-provded pars of data wth the assumpton that there s smlar or dssmlar pars n datasets. There are also some sem-supervsed hashng methods [3, 34] n whch a supervsed term s used to mnmze the emprcal error on the labeled data whle an unsupervsed term s used to maxmze desrable propertes, such as varance and ndependence of ndvdual bts n the bnary codes. Mult-modal hashng s desgned to conduct hash functon learnng for encodng mult-modal data. To ths end, the method n [36] frst uses an teratve method to preserve the semantc smlartes among tranng examples, and then keeps the consstency between the hash codes and the correspondng hashng functons desgned for multple modals. The method multple feature hashng (MFH) [26] preserves the local structure nformaton of each modal and also globally consders the algnments for all the modals to learn a group of hash functons for real-tme large scale nearduplcate web vdeo retreval. Cross-modal hashng also encodes mult-modal data. However, t focuses more on dscoverng the correlatons among dfferent modals to enable cross-modal retreval. Cross-modal smlarty senstve hashng () [2] s the frst crossmodal hashng method for cross-modal retreval. However, only consders the nter-smlarty and gnores the ntra-smlarty. Cross-vew hashng () [6] extends spectral hashng [33] to the mult-modal case, amng at mnmzng the Hammng dstances for smlar ponts and maxmzng those for dssmlar ponts. However, t needs to construct the smlarty matrx for all the data ponts, whch leads to a uadratc tme complexty to the tranng data sze. Raswasa et al., [22] employs canoncal correlaton analyss (CCA) to conduct hash functon learnng, whch s a specal case of. Recently, mult-modal latent bnary embeddng () [38] uses a probablstc latent factor model to learn hash functons. Smlar to, t also has a uadratc tme complexty for constructng the smlarty matrx. Moreover, t uses a samplng method to solve the ssue of out-of-sample extenson. Co-regularzed hashng (CRH) [37] s a boosted co-regularzaton framework whch learns a group of hash functons for each bt of bnary codes n every modal. However, ts objectve functon s nonconvex. Inter-meda hashng (IMH) [27] ams to dscover a common Hammng space for learnng hash functons. IMH preserves the ntra-smlarty of each ndvdual modal va enforcng that the data wth smlar semantc should have smlar hash codes, and preserves the nter-smlarty among dfferent modals va preservng local structural nformaton embedded n each modal. 3. LINEAR CROSS-MODAL HASHING In ths secton we descrbe the detals of the proposed method. For the purpose of nterpretng our basc dea, we frst focus on hash functon learnng for bmodal data from Secton 3. to Secton 3.5, and then extend t to the general settng of mult-modal data n Secton 3.6. In ths paper, we use boldface uppercases, boldface lowercase and letter to denote matrces, vectors and scales, respectvely. Besdes, the transpose of X s denoted as X T, the nverse of X s denoted as X, and the trace operator of a matrx X s denoted as the symbol tr(x). 3. Problem formulaton Assume we have two modals, X () = {x (),, x() n }; =, 2, descrbng the same data ponts where n s the number of data ponts. For example, X () s the SIFT vsual feature extracted from the content of mages, and X (2) s the bagof-words feature extracted from the text surroundng the mages. In general the feature dmensonaltes of dfferent modals are dfferent. Wth the same assumpton n [4, ] that there s an nvarant common space among multple modals, the objectve of s to effectvely and effcently learn hash functons for dfferent modals to support cross-modal retreval. To ths end, needs to generate the hash functons: f () : x () b () = {, } c, =, 2, where c s the code length. Note that all the modals have the same code length. Moreover, needs to ensure that the neghborhood relatonshps wthn each ndvdual modal and across dfferent modals are preserved n the produced common Hammng space. To do ths, s devsed to preserve both the ntra-smlarty and the nter-smlarty of the orgnal feature spaces n the Hammng space. The man dea of learnng the hash functons goes as follows. Data of each ndvdual modal are frstly converted nto ther new representatons, denoted as Z (), for preservng the ntra-smlarty (see Secton 3.2). Data of all modals represented by Z are then mapped nto a common space where the nter-smlarty s preserved to generate hash functons (see Secton 3.3). Fnally, values generated from hash functons are bnarzed nto the Hammng space (see Secton 3.4). Wth the learnt hash functons, ueres and database data can be mapped nto the Hammng space to facltate fast search by effcent bnary code matchng. 3.2 Intra-smlarty preservaton Intra-smlarty preservaton s desgned to mantan the neghborhood relatonshps among tranng data ponts n each ndvdual modal after they are mapped nto the new space spanned by ther new representatons. To acheve ths, manfold-lke hashng [26, 27, 36, 39] constructs a smlarty matrx, where each entry represents the dstance between two data ponts. In such a matrx, each data pont can be regarded as a n-dmensonal representaton ndcatng ts dstance to n data ponts. Typcally, a neghborhood for a data pont s descrbed by ts few nearest neghbors. To 45

4 preserve the neghborhood of each data pont, only few dmensons correspondng to ts nearest neghbors n the n- dmensonal representaton are non-zero. In other words, the n-dmensonal representaton s hghly sparse. However, to buld such a sparse matrx needs uadratc tme complexty,.e., O(n 2 ), whch s mpractcal for large-scale datasets. Observed from the sparse n-dmensonal representaton, only few data ponts are used to descrbe the neghborhood for a data pont. Ths motvates us to derve a smaller k-dmensonal representaton (.e., k n) to approxmate each tranng data pont, amng at reducng the tme complexty for buldng the neghborhood structures. The dea s to select k most representatve data ponts from the tranng dataset and approxmate each tranng data pont usng ts dstances to these k representatve data ponts. To do ths, n ths paper we use a scalable k-means clusterng method [5] to generate k centrods whch are taken as k most representatve data ponts ponts n the tranng dataset. It has been shown that k centrods have a strong representaton power to adeuately cover large-scale datasets [5]. More specfcally, gven a tranng dataset n the frst modal X (), nstead of mappng each tranng data pont x () nto the n-dmensonal representaton leadng to uadratc tme complexty, we convert t nto the k-dmensonal representaton z (), usng the obtaned k centrods whch are denoted by m (), =, 2,, k. For a z (),tsj-thdmenson carres the dstance from x () to the j-th centrod m () j, denoted as z () j. To obtan the value of z () j, we frst calculate the Eucldean dstance between x () and m () j,.e., z () j = x () m () j 2, () where. stands for the Eucldean norm. As n [9], the value of z () j can be further defned as a functon of the Eucldean to better ft the Gaussan dstrbuton n real applcatons. Denote the redefned value of z () j as p () j,wehave: p () exp( z () j j = /σ) k l= exp( z() l /σ), (2) where σ s a tunng parameter for controllng the decay rate of z () j. For smplcty, we set σ = n ths paper, whle an adaptve settng of σ can lead to better results. Let p () =[p () ; ; p() j ; ; p() k ], p() forms the new representaton of x (). It can be seen that the ratonale of defnng p () s smlar to that of kernel densty estmaton wth a Gaussan kernel,.e., f x () s near to the j-th centrod, p () j wll be relatvely hgh; otherwse, p () j wll decay. To preserve the neghborhood of each tranng data pont n the new k-dmensonal space, here we also represent each tranng data pont usng several (say s and s k) nearest centrods so that the new representaton p () of x () s sparse. Therefore, n the mplementaton, for each tranng data pont we only keep the values to ts s nearest centrods n p () and set the rest as. After ths, we normalze the derved value to generate the fnal value of z () j. Accordng to the perspectve of geometrc reconstructon n the lteratures [23, 25, 32], we can easly show that the ntra-smlarty can be well preserved n the derved k-dmensonal representaton,.e., the nvarance to rotatons, rescalngs, and translatons. Accordng to Es.-2, we convert the tranng data X () nto ther k-dmensonal representatons Z (), =, 2. That s,wecanuseak n matrx to approxmate the orgnal n n smlarty matrx wth ntra-smlarty preservaton. The advantage s to reduce the complexty from O(n 2 )too(kn). Note that one can select dfferent numbers of centrods for each modal. For smplcty, n ths paper we select the same numbers of centrods n our experments. The next problem s to preserve the nter-smlarty between Z () and Z (2) va seekng a common latent space between them. 3.3 Inter-smlarty preservaton It s well known that multmeda data wth same semantcs can exst n dfferent types of modals. For example, a text document and an mage can descrbe exactly the same topc. Research has shown that f data descrbed n dfferent modal spaces are related to the same event or topc, they are expected to have some common latent space [6, 38]. Ths suggests that mult-modal data wth the same semantc should share some common space n whch relevant data are close to each other. Such a property s understood as nter-smlarty preservaton when modal-modal data are mapped nto the common space. In our problem settng, mult-modal data are eventually represented by bnary codes n the common Hammng space. To ths end, we frst learn a semantc brdge for each modal Z () n ts k-dmensonal space to map Z () nto the common Hammng space. To ensure nter-smlarty preservaton, n the Hammng space, data descrbng the same object from dfferent modals should have same or smlar bnary codes. For example, n Fg.2, we map both the mages vsual modal and textual modal va learnt semantc brdges (.e., the arrows n Fg.2) nto the Hammng space (.e., the crcle n Fg.2), n whch two modals of an mage are represented wth same or smlar bnary codes n the Hammng space. That s, consstency across dfferent modals s acheved. Fgure 2: An llustraton on nter-smlarty preservaton. More formally, gven Z () R n k and Z (2) R n k where n s sample sze and k s the number of centrods, we learn the transformaton matrx (.e., semantc brdge ) W () R k c and W (2) R k c for convertng Z () and Z (2) nto the new representaton B () {, } n c and B (2) {, } n c 46

5 n a common Hammng space, n whch each sample par (descrbng the same object,.e., B () and B (2) descrbng the -th object wth dfferent modals) has the mnmal Hammng dstance,.e., the maxmal consstency. Ths leads to the followng objectve functon: mn B (),B (2) B () B (2) 2 F s.t., B ()T e =, b () {, }, B ()T B () = I c, =, 2, where. F means a Frobenus norm, e s a n vector whose each entry s and I c s a c c dentty matrx, the constrant B ()T e = reures each bt has eual chance to be or -, the constrant B ()T B () = I c reures the bts to be obtaned ndependently, and the loss functon term B () B (2) 2 F acheves the mnmal dfference (or the maxmal consstency) on two representatons of an object. The optmzaton problem n E.3 euals to the ssue of balanced graph parttonng and s NP-hard. Followng the lteratures [6, 33], we frst denote Y () as the real-valued representaton of B () and solve the derved objectve functon on Y () n ths subsecton, and then bnarze Y () nto bnary codes based on the medan threshold method n Secton 3.4. To map Z () nto Y () R n c va the transformaton matrx W (),welety () = Z () W (). Accordng to E.3, we have the objectve functon as Z mn () W () Z (2) W (2) 2 W (),W (2) F s.t., W ()T W () = I, W (2)T W (2) = I, where orthogonal constrants are set to avod trval solutons. To optmze the objectve functon n E.4, we frst convert ts loss functon term nto (3) (4) 3.4 Bnarzaton After obtanng all Y (), we get the medan vector of Y () u () =medan(y () ) R c, (8) we then bnarze Y () as follows: b () jl = f y () jl u () l b () jl = f y () jl <u () l where Y () =[y () (9),, y() n ] T, =, 2; j =,, n and l =,, c. E.9 generates the fnal bnary codes B for the tranng data X, n whch the medan value of each dmenson s used as the threshold for bnarzaton. The learnt hash functons and bnarzaton step are used to map unseen data (e.g., database and uery) nto the Hammng space. In the onlne search phase, gven a uery x () from the -th modal, we frst approxmate t wth ts dstances to k centrods,.e., z () compute ts y () y () to generate ts bnary codes b () dstances between b () usng Es.(-2), and then usng E.7, followed by the bnarzaton on. Fnally the Hammng and database bnary codes are com- n any other modal. puted to fnd the neghbors of x () 3.5 Summary and analyss We summarze the proposed approach n Algorthm (tranng phase) and Algorthm 2 (search phase). Algorthm : Pseudo code of tranng phase Input: X, c, k Output: u () R c ; W () R k c,=,2 Perform scalable k-means on X () to obtan m () ; 2 Compute Z () by E.-2; 3 Generate W () by E.6; 4 Generate u () by E.8; Z () W () Z (2) W (2) 2 F = tr(w ()T Z ()T Z () W () + W (2)T Z (2)T Z (2) W (2) (5) W ()T Z ()T Z (2) W (2) W (2)T Z (2)T Z () W () ) = tr(w T ZW), where W =[W ()T ; W (2)T ] T R 2k c and ( ) Z ()T Z () Z ()T Z (2) Z = R 2k 2k. Z (2)T Z () Z (2)T Z (2) Then the objectve functon n E.4 becomes: max W tr(w T ZW) s.t., W T W = I. (6) E.6 s an egenvalue problem. We can obtan the optmal W va solvng the egenvalue problem on Z. W represents the hash functons to generate Y as follows: Y () = tr(z () W () ) (7) where W () = W( : k, :) and W (2) = W(k +:end, :). Algorthm 2: Pseudo code of search phase Input: x (), u (), W () Output: Nearestneghborsofx () n another modal Compute z () by E.-2; 2 Compute y () by E.7; 3 Generate b () t by E.9; 4 Match b () wth database bnary codes n another modal; In the tranng phase of, tme cost manly comes from the clusterng process, new representaton generaton, and egenvalue decomposton n generatng hash functons. Applyng a scalable clusterng method, such as [5], clusters can be generated n lnear tme complexty to the tranng data sze n. Generatngthek-dmensonal representatons Z takes the complexty of O(kn). The tme complexty to generate W s O(mn{nk 2,k 3 }). Snce k n for largescale tranng datasets, O(k 3 ) s the complexty to generate 47

6 hash functons. Therefore, the overall tme complexty s O(max{kn, k 3 }). Gven that k n, weexpectthatk 2 <n or both have smlar scale. Ths leads to the approxmaton of O(kn) tme complexty for the tranng phase. Havng k as a constant, the fnal tme complexty becomes lnear to the tranng data sze. In the search phase, the tme complexty s constant. 3.6 Extenson We present an extenson of Algorthm and Algorthm 2 to the case of more than two modals, whch makes us to use the nformaton avalable n all the possble modals to acheve better learnng results. To do ths, we frst generate new representatons of each modal accordng to Secton 3. for preservng ntra-smlarty, and then transform new representatons of all the modals nto a common latent space for preservng nter-smlarty across any par of modals. The objectve functon for preservng nter-smlarty s defned as: mn B (),=,,p p = <j p B () B (j) 2 F s.t., B ()T e =, b () {, }, B ()T B () = I c, =,, p, () where e s a n vector,p s the number of dfferent modals, I c s a c c dentty matrx, the constrant B ()T e = reures each bt has eual chance to be or -, and the constrant B ()T B () = I d reures the bts of each modal to be obtaned ndependently. To solve E., we frst relax t to: p p mn Z () W () Z (j) W (j) 2 W (),=,,p = <j F () s.t., W ()T W () = I,=,, p. We then obtan max W tr(wt ZW) s.t., W T W = I, (2) where W =[W ()T ; ; W (p)t ] T, W R pk c and Z ()T Z () Z ()T Z (2) Z ()T Z (p) Z = Z (2)T Z () Z (2)T Z (2) Z (2)T Z (p), Z (p)t Z () Z (p)t Z (2) Z (p)t Z (p) where Z R pk pk. After solvng the egenvalue problem n E.2, we obtan the hash functons of multple modals (smlar to E.7 to E.9 n Secton 3.4). Wth hash functons and medan thresholds, we can transform database data and ueres nto the Hammng space, to support cross-modal retreval va effcent bnary code comparsons. 4. EXPERIMENTAL ANALYSIS We conduct our experments on two benchmark datasets,.e., Wk [22] and NUS-WIDE [6], so far the largest publcly avalable mult-modal datasets that are fully pared and labeled [38]. The two datasets are bmodal wth both vsual and textual modals n dfferent representatons. In our experments, each dataset s parttoned nto a uery set and a database set whch s used for tranng. 4. Comparson algorthms The comparson algorthms nclude a baselne algorthm B and state-of-the-art algorthms, ncludng [6], [2] and [38]. B s our wthout preservng ntra-smlarty, wth the purpose to test the effect of ntra-smlarty preservaton n our method. We compare wth the comparson algorthms on two cross-modal retreval tasks. Specfcally, one task s to use a text uery n the textual modal to search relevant mages n the vsual modal (shorted for Text uery vs. Image data ) and the other s to use an mage uery n the vsual modal to search relevant texts from the textual modal (shorted for Image uery vs. Text data ). 4.2 Evaluaton Metrcs We use mean Average Precson (map) [38] as one of performance measures. Gven a uery and a lst of R retreved results, the value of ts Average Precson s defned as R AP = { P (r)δ(r)}, (3) l r= where l s the number of true neghbors n ground truth, P (r) denotes the precson of the top r retreved results, and δ(r) = fther-th retreved result s a true neghbor of the uery, otherwse δ(r) =. map s the mean of all the ueres average precson. Clearly, the larger the map, the better the performance s. In our experments, we set R as the number of tranng data ponts whose Hammng dstances to the uery are not larger than 2. We also report the results on two other types of measures, ncludng recall curves wth dfferent retreved results and tme cost for generatng hash functons and searchng database bnary codes. Both map and recall curves are used to reflect the retreval effectveness and tme cost s used to evaluate the effcency. 4.3 Parameters settng By default, we set the parameter k = 3 for dataset Wk and k = 6 for dataset NUS-WIDE. Amongk centrods, we set s = 3 for representng each tranng data pont wth s nearest centrods. In our experments, we vary the length of hash codes (.e., the number of hash bts) n the range of [8, 6, 24] for dataset Wk and [8, 6, 32] for dataset NUS-WIDE. Moreover, for calculatng the value of recall curves, we set the number of retreved results n the range of [25, 5, 75,, 25, 5, 7, 2] for Wk and [, 2, 5, 8,, 2, 5] on NUS-WIDE. For all the comparson algorthms, the codes are provded by the authors. We tune the parameters accordng to the correspondng lteratures. All the experments are conducted on a computer whch has Intel Xeon(R) 2.9GHz 2 processors, 92 GB RAM and the 64-bt Wndows 7 operatng system. 4.4 Results on Wk dataset The dataset Wk [22] s generated from a group of 2,866 Wkpeda documents. In Wk, each object s an magetext par and s labeled wth exactly one of semantc 48

7 classes. The mages are represented by 28-dmensonal SIFT feature vectors. The text artcles are represented by the probablty dstrbutons over topcs, whch are derved from a latent Drchlet allocaton (LDA) model []. Followng the settng n the lterature [22], 273 data ponts form the database set and the remanng 693 data ponts form the uery set. Due to that the dataset s fully annotated, semantc neghbors for a uery s regarded as the ground truth, based on the assocated labels. The map results of all the algorthms on dfferent code lengths are reported n Fg.3.(a-b). The recall curves for two uery tasks on dfferent code lengths are plotted n Fg.4. Accordng to the expermental results, we can see that consstently performs best. For example, the maxmal dfference between and the second best one (.e., ) s about 4% n Fg.3.(a) and about 8% n Fg.3.(b) whle the code length s 24. Moreover, both and are better than, whch s wth the same concluson as n [38]. Besdes, we also have three observatons based on our expermental results. Frst,, and outperform B and whch only consder the nter-smlarty across modals and gnore the ntra-smlarty wthn a modal. Therefore, we can make an concluson that t s useful for consderng both the ntra-smlarty and the nter-smlarty together to buld cross-modal hashng. Second, although both and B consder the nter-smlarty, mproves B slghtly snce employs pror knowledge, such as the predefned smlar pars and dssmlar pars [2]. Thrd, accordng to the expermental results on map and recall curves, we see that all algorthms acheve ther best performance when the number of hash bts s 6 for dataset Wk. After achevng ther peak, the performance of all algorthms degrades. A possble reason s that a longer bnary code representaton may lead to less retreved results gven the fxed Hammng dstance threshold, whch affects ts precson and recall. Such a phenomenon has also been dscussed n [8, 38]. Table : Runnng tme for all algorthms whle fxng code length as 6 for dataset Wk and dataset NUS-WIDE. Both tranng tme and search tme are recorded n second. Wk NUS-WIDE Task Methods Image uery vs. Text data Text uery vs. Image data tran search tran search B B Table shows the tme cost of the tranng phase and the search phase of all the algorthms. We can see that s most tme consumng snce t s a generatve model, followed by, and. Snce does not consder the ntra-smlarty, t s faster than. However, has unsatsfactory performance n search ualty as shown n Fg Results on NUS-WIDE dataset The dataset NUS-WIDE orgnally contans 269, 648 mages assocated wth 8 ground truth concept tags. Followng the lterature [8, 3], we prune orgnal NUS-WIDE to form a new dataset NUS-WIDE consstng of 95, 969 mage-tag pars by keepng the pars that belong to one of the 2 most freuent tags, such as anmal, buldngs, person, etc. In our NUS-WIDE, each par s annotated by at least one of 2 labels. The mages are represented by 5-dmensonal SIFT feature vectors and the texts are represented by -dmensonal feature vectors obtaned by performng PCA on the orgnal tag occurrence features. Followng the settng n the lteratures [8, 3], we unformly sample mages from each of the selected 2 tags to form auerysetof2, mages and the left 93, 869 magetag pars servng as the database set. The ground truth s defned based on whether two mages share at least one common tag n our experments. As shown n n Fg.3.(c-d), Fg.5, and Table, one can see that the rankng of all the algorthms on dataset NUS- WIDE s largely consstent wth that on dataset Wk. The maxmal dfference between and the second best one (.e., ) s about 6% n Fg.3.(c) and about 5% n Fg.3.(d) whle the code length s 6. Table 2: Runnng tme wth dfferent number of centrods whle fxng code length as 6 for dataset Wk and dataset NUS-WIDE. Both tranng tme and search tme are recorded n second. Task Centrods Wk NUS-WIDE tran search tran search Image uery k = vs. k = Text data k = Text uery vs. Image data k = k = k = Parameters senstvty In ths secton, we test the senstvty of dfferent parameters. Frst, we look at the effect of k. We set the dfferent values on k (.e., the number of clusters n the tranng phase) and report parts of results n Fg.6. From Fg.6, we can see that a larger k value leads to better results, snce the k-dmensonal representaton can be more accurate n capturng the orgnal data dstrbuton n the tranng dataset. Nonetheless, more tranng cost occurs for a larger k value, as shown n Table 2. Our results show that a relatvely small k value (e.g., k=3 and 6 for Wk and NUS-WIDE) can acheve reasonably good results. Due to the space lmt, we do not report the results on dfferent s values. Generally, a good choce of s s between 3 to CONCLUSION In ths paper we have proposed a novel and effectve crossmodal hashng approach, namely lnear cross-modal hashng (). The man dea s to represent each tranng data pont wth a smaller k-dmensonal approxmaton whch can preserve the ntra-smlarty and reduce the tme and space complexty n learnng hash functons. We then map the 49

8 map.25.2 B map B map B map B c = 8 c = 6 c = 24 c = 8 c = 6 c = 24 c = 8 c = 6 c = 32 c = 8 c = 6 c = 32 Dfferent code lengths (c) Dfferent code lengths (c) Dfferent code lengths (c) Dfferent code lengths (c) (a) Image uery vs. Text data (b) Text uery vs. Image data (c) Image uery vs. Text data (d) Text uery vs. Image data Fgure 3: map comparson wth dfferent code lengths for dateset Wk (a-b) and for dataset NUS WIDE (c-d)..8 B.8 B.8 B No. of Retreved Samples (a) code length = No. of Retreved Samples (b) code length = No. of Retreved Samples (c) code length = 24.8 B.8 B.8 B No. of Retreved Samples (d) code length = No. of Retreved Samples (e) code length = No. of Retreved Samples (f) code length = 24 Fgure 4: curves wth dfferent code lengths for dataset Wk. The upper row (a-c) s the task of mage uery vs. text data, the bottom row (d-f) s the task of text uery vs. mage data. new representatons of the tranng data from all modals to a common latent space n whch the nter-smlarty s preserved and hash functons of each modal are obtaned. Gven a uery, t s frst transformed nto ts k-dmensonal representaton whch s then mapped nto the Hammng space wth the learnt hash functons, to match wth database bnary codes. Snce bnary codes from dfferent modals are comparable n the Hammng space, cross-modal retreval can be effectvely and effectvely supported by. The expermental results on two benchmark datasets demonstrate that outperforms the state of the art sgnfcantly wth practcal tme cost. 6. ACKNOWLEDGEMENTS Ths work was supported by the Australa Research Councl (ARC) under research Grant DP94678 and the Nature Scence Foundaton (NSF) of Chna under grants

9 .8 B.8 B.8 B No. of Retreved Samples x 4 (a) code length = No. of Retreved Samples x 4 (b) code length = No. of Retreved Samples x 4 (c) code length = 32.8 B.8 B.8 B No. of Retreved Samples x 4 (d) code length = No. of Retreved Samples x 4 (e) code length = No. of Retreved Samples x 4 (f) code length = 32 Fgure 5: curves wth dfferent code lengths for dataset NUS WIDE. The upper row (a-c) s the task of mage uery vs. text data, the bottom row (d-f) s the task of text uery vs. mage data..8 k = 3 k = 6 k =.8 k = 3 k = 6 k =.8 k = 3 k = 6 k =.8 k = 3 k = 6 k = No. of Retreved Samples No. of Retreved Samples No. of Retreved Samples x 4 No. of Retreved Samples x 4 (a) Image uery vs. Text data (b) Text uery vs. Image data (c) Image uery vs. Text data (d) Text uery vs. Image data Fgure 6: curves wth dfferent number of centrods whle fxng the code length as 6 for dateset Wk (a-b) and NUS WIDE (c-d). 7. REFERENCES [] D. M. Ble, A. Y. Ng, and M. I. Jordan. Latent drchlet allocaton. J. Mach. Learn. Res., 3:993 22, 23. [2] M.M.Bronsten,A.M.Bronsten,F.Mchel,and N. Paragos. Data fuson through cross-modalty metrc learnng usng smlarty-senstve hashng. In CVPR, pages , 2. [3] R. Chaudhry and Y. Ivanov. Fast approxmate nearest neghbor methods for non-eucldean manfolds wth applcatons to human actvty analyss n vdeos. In ECCV, pages , 2. [4] M. Chen, K. Q. Wenberger, and J. C. Bltzer. Co-tranng for doman adaptaton. In NIPS, pages 9, 2. [5] X. Chen and D. Ca. Large scale spectral clusterng wth landmark-based representaton. In AAAI, pages 33 38, 2. [6] T.-S. Chua, J. Tang, R. Hong, H. L, Z. Luo, and Y. Zheng. Nus-wde: a real-world web mage database 5

10 from natonal unversty of sngapore. In CIVR, pages 48 56, 29. [7] M. Datar, N. Immorlca, P. Indyk, and V. S. Mrrokn. Localty-senstve hashng scheme based on p-stable dstrbutons. In SOCG, pages , 24. [8] A. Gons, P. Indyk, and R. Motwan. Smlarty search n hgh dmensons va hashng. In VLDB, pages , 999. [9] J.Goldberger,S.T.Rowes,G.E.Hnton,and R. Salakhutdnov. Neghbourhood components analyss. In NIPS, pages 9, 24. [] Y. Gong, S. Lazebnk, A. Gordo, and F. Perronnn. Iteratve uantzaton: A procrustean approach to learnng bnary codes for large-scale mage retreval. IEEE Trans. Pattern Anal. Mach. Intell., page accepted, 22. [] R. Gopalan, R. L, and R. Chellappa. Doman adaptaton for object recognton: An unsupervsed approach. In ICCV, pages 999 6, 2. [2] P. Jan, B. Kuls, and K. Grauman. Fast mage search for learned metrcs. In CVPR, pages 8, 28. [3] H. Jégou, M. Douze, and C. Schmd. Product uantzaton for nearest neghbor search. In CVPR, pages 7 28, 2. [4] B. Kuls and T. Darrell. Learnng to hash wth bnary reconstructve embeddngs. In NIPS, pages 42 5, 29. [5] B. Kuls and K. Grauman. Kernelzed localty-senstve hashng for scalable mage search. In ICCV, pages , 29. [6] S. Kumar and R. Udupa. Learnng hash functons for cross-vew smlarty search. In IJCAI, pages , 2. [7] W. Lu, J. Wang, R. J, Y.-G. Jang, and S.-F. Chang. Supervsed hashng wth kernels. In CVPR, pages , 22. [8] W. Lu, J. Wang, S. Kumar, and S.-F. Chang. Hashng wth graphs. In ICML, pages 8, 2. [9] M. Norouz and D. J. Fleet. Mnmal loss hashng for compact bnary codes. In ICML, pages , 2. [2] M. Norouz, A. Punjan, and D. J. Fleet. Fast search n hammng space wth mult-ndex hashng. In CVPR, pages 38 35, 22. [2] M. Ragnsky and S. Lazebnk. Localty-senstve bnary codes from shft-nvarant kernels. In NIPS, pages 59 57, 29. [22] N. Raswasa, J. C. Perera, E. Covello, and G. Doyle. A new approach to cross-modal multmeda retreval. In ACM MM, pages 25 26, 2. [23] S. Rowes and L. Saul. Nonlnear dmensonalty reducton by locally lnear embeddng. Scence, 29(55): , 2. [24] R. Salakhutdnov and G. E. Hnton. Semantc hashng. Int. J. Approx. Reasonng, 5(7): , 29. [25] L. K. Saul and S. T. Rowes. Thnk globally, ft locally: Unsupervsed learnng of low dmensonal manfold. J.Mach.Learn.Res., 4:9 55, 23. [26] J. Song, Y. Yang, Z. Huang, H. T. Shen, and R. Hong. Multple feature hashng for real-tme large scale near-duplcate vdeo retreval. In ACM MM, pages , 2. [27] J. Song, Y. Yang, Y. Yang, Z. Huang, and H. T. Shen. Inter-meda hashng for large-scale retreval from heterogenous data sources. In SIGMOD, pages , 23. [28] C. Strecha, A. A. Bronsten, M. M. Bronsten, and P. Fua. Ldahash: Improved matchng wth smaller descrptors. IEEE Trans. Pattern Anal. Mach. Intell., 34():66 78, 22. [29] A. Torralba, R. Fergus, and Y. Wess. Small codes and large mage databases for recognton. In CVPR, pages 8, 28. [3] J. Wang, O. Kumar, and S.-F. Chang. Sem-supervsed hashng for scalable mage retreval. In CVPR, pages , 2. [3] J. Wang, S. Kumar, and S.-F. Chang. Seuental projecton learnng for hashng wth compact codes. In ICML, pages 27 34, 2. [32] K. Q. Wenberger, B. D. Packer, and L. K. Saul. Nonlnear dmensonalty reducton by semdefnte programmng and kernel matrx factorzaton. In AISTATS, pages , 25. [33] Y. Wess, A. Torralba, and R. Fergus. Spectral hashng. In NIPS, pages , 28. [34] C.Wu,J.Zhu,D.Ca,C.Chen,andJ.Bu. Sem-supervsed nonlnear hashng usng bootstrap seuental projecton learnng. IEEE Trans. Knowl. Data Eng., 99:, 22. [35] Y. Yang, D. Xu, F. Ne, J. Luo, and Y. Zhuang. Rankng wth local regresson and global algnment for cross meda retreval. In ACM MM, pages 75 84, 29. [36] D. Zhang, F. Wang, and L. S. Composte hashng wth multple nformaton sources. In SIGIR, pages , 2. [37] Y. Zhen and D.-Y. Yeung. Co-regularzed hashng for multmodal data. In NIPS, pages , 22. [38] Y. Zhen and D.-Y. Yeung. A probablstc model for multmodal hash functon learnng. In SIGKDD, pages , 22. [39] X. Zhu, Z. Huang, H. Cheng, J. Cu, and H. T. Shen. Sparse hashng for fast multmeda search. ACM Trans. Inf. Syst., 3(2):59 57,

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY 1. SSDH: Semi-supervised Deep Hashing for Large Scale Image Retrieval

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY 1. SSDH: Semi-supervised Deep Hashing for Large Scale Image Retrieval IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY SSDH: Sem-supervsed Deep Hashng for Large Scale Image Retreval Jan Zhang, and Yuxn Peng arxv:607.08477v2 [cs.cv] 8 Jun 207 Abstract Hashng

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

LECTURE : MANIFOLD LEARNING

LECTURE : MANIFOLD LEARNING LECTURE : MANIFOLD LEARNING Rta Osadchy Some sldes are due to L.Saul, V. C. Raykar, N. Verma Topcs PCA MDS IsoMap LLE EgenMaps Done! Dmensonalty Reducton Data representaton Inputs are real-valued vectors

More information

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data A Fast Content-Based Multmeda Retreval Technque Usng Compressed Data Borko Furht and Pornvt Saksobhavvat NSF Multmeda Laboratory Florda Atlantc Unversty, Boca Raton, Florda 3343 ABSTRACT In ths paper,

More information

MULTI-VIEW ANCHOR GRAPH HASHING

MULTI-VIEW ANCHOR GRAPH HASHING MULTI-VIEW ANCHOR GRAPH HASHING Saehoon Km 1 and Seungjn Cho 1,2 1 Department of Computer Scence and Engneerng, POSTECH, Korea 2 Dvson of IT Convergence Engneerng, POSTECH, Korea {kshkawa, seungjn}@postech.ac.kr

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

Data-dependent Hashing Based on p-stable Distribution

Data-dependent Hashing Based on p-stable Distribution Data-depent Hashng Based on p-stable Dstrbuton Author Ba, Xao, Yang, Hachuan, Zhou, Jun, Ren, Peng, Cheng, Jan Publshed 24 Journal Ttle IEEE Transactons on Image Processng DOI https://do.org/.9/tip.24.2352458

More information

Collaboratively Regularized Nearest Points for Set Based Recognition

Collaboratively Regularized Nearest Points for Set Based Recognition Academc Center for Computng and Meda Studes, Kyoto Unversty Collaboratvely Regularzed Nearest Ponts for Set Based Recognton Yang Wu, Mchhko Mnoh, Masayuk Mukunok Kyoto Unversty 9/1/013 BMVC 013 @ Brstol,

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

Face Recognition University at Buffalo CSE666 Lecture Slides Resources:

Face Recognition University at Buffalo CSE666 Lecture Slides Resources: Face Recognton Unversty at Buffalo CSE666 Lecture Sldes Resources: http://www.face-rec.org/algorthms/ Overvew of face recognton algorthms Correlaton - Pxel based correspondence between two face mages Structural

More information

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization

Problem Definitions and Evaluation Criteria for Computational Expensive Optimization Problem efntons and Evaluaton Crtera for Computatonal Expensve Optmzaton B. Lu 1, Q. Chen and Q. Zhang 3, J. J. Lang 4, P. N. Suganthan, B. Y. Qu 6 1 epartment of Computng, Glyndwr Unversty, UK Faclty

More information

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

Cluster Analysis of Electrical Behavior

Cluster Analysis of Electrical Behavior Journal of Computer and Communcatons, 205, 3, 88-93 Publshed Onlne May 205 n ScRes. http://www.scrp.org/ournal/cc http://dx.do.org/0.4236/cc.205.350 Cluster Analyss of Electrcal Behavor Ln Lu Ln Lu, School

More information

AS THE major component of big data, multi-modal data

AS THE major component of big data, multi-modal data 1220 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 19, NO. 6, JUNE 2017 Cross-Modal Retreval Usng Multordered Dscrmnatve Structured Subspace Learnng Lang Zhang, Bngpeng Ma, Guorong L, Qngmng Huang, and Q Tan Abstract

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers

Content Based Image Retrieval Using 2-D Discrete Wavelet with Texture Feature with Different Classifiers IOSR Journal of Electroncs and Communcaton Engneerng (IOSR-JECE) e-issn: 78-834,p- ISSN: 78-8735.Volume 9, Issue, Ver. IV (Mar - Apr. 04), PP 0-07 Content Based Image Retreval Usng -D Dscrete Wavelet wth

More information

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION Paulo Quntlano 1 & Antono Santa-Rosa 1 Federal Polce Department, Brasla, Brazl. E-mals: quntlano.pqs@dpf.gov.br and

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

Discriminative Dictionary Learning with Pairwise Constraints

Discriminative Dictionary Learning with Pairwise Constraints Dscrmnatve Dctonary Learnng wth Parwse Constrants Humn Guo Zhuoln Jang LARRY S. DAVIS UNIVERSITY OF MARYLAND Nov. 6 th, Outlne Introducton/motvaton Dctonary Learnng Dscrmnatve Dctonary Learnng wth Parwse

More information

Image Alignment CSC 767

Image Alignment CSC 767 Image Algnment CSC 767 Image algnment Image from http://graphcs.cs.cmu.edu/courses/15-463/2010_fall/ Image algnment: Applcatons Panorama sttchng Image algnment: Applcatons Recognton of object nstances

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 15 CS434a/541a: Pattern Recognton Prof. Olga Veksler Lecture 15 Today New Topc: Unsupervsed Learnng Supervsed vs. unsupervsed learnng Unsupervsed learnng Net Tme: parametrc unsupervsed learnng Today: nonparametrc

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervsed Learnng and Clusterng Supervsed vs. Unsupervsed Learnng Up to now we consdered supervsed learnng scenaro, where we are gven 1. samples 1,, n 2. class labels for all samples 1,, n Ths s also

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Decson surface s a hyperplane (lne n 2D) n feature space (smlar to the Perceptron) Arguably, the most mportant recent dscovery n machne learnng In a nutshell: map the data to a predetermned

More information

Laplacian Eigenmap for Image Retrieval

Laplacian Eigenmap for Image Retrieval Laplacan Egenmap for Image Retreval Xaofe He Partha Nyog Department of Computer Scence The Unversty of Chcago, 1100 E 58 th Street, Chcago, IL 60637 ABSTRACT Dmensonalty reducton has been receved much

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervsed Learnng and Clusterng Why consder unlabeled samples?. Collectng and labelng large set of samples s costly Gettng recorded speech s free, labelng s tme consumng 2. Classfer could be desgned

More information

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

What Is the Most Efficient Way to Select Nearest Neighbor Candidates for Fast Approximate Nearest Neighbor Search?

What Is the Most Efficient Way to Select Nearest Neighbor Candidates for Fast Approximate Nearest Neighbor Search? IEEE Internatonal Conference on Computer Vson What Is the Most Effcent Way to Select Nearest Neghbor Canddates for Fast Approxmate Nearest Neghbor Search? Masakazu Iwamura, Tomokazu Sato and Koch Kse Graduate

More information

Machine Learning: Algorithms and Applications

Machine Learning: Algorithms and Applications 14/05/1 Machne Learnng: Algorthms and Applcatons Florano Zn Free Unversty of Bozen-Bolzano Faculty of Computer Scence Academc Year 011-01 Lecture 10: 14 May 01 Unsupervsed Learnng cont Sldes courtesy of

More information

Semantic Image Retrieval Using Region Based Inverted File

Semantic Image Retrieval Using Region Based Inverted File Semantc Image Retreval Usng Regon Based Inverted Fle Dengsheng Zhang, Md Monrul Islam, Guoun Lu and Jn Hou 2 Gppsland School of Informaton Technology, Monash Unversty Churchll, VIC 3842, Australa E-mal:

More information

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION

BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION BAYESIAN MULTI-SOURCE DOMAIN ADAPTATION SHI-LIANG SUN, HONG-LEI SHI Department of Computer Scence and Technology, East Chna Normal Unversty 500 Dongchuan Road, Shangha 200241, P. R. Chna E-MAIL: slsun@cs.ecnu.edu.cn,

More information

Private Information Retrieval (PIR)

Private Information Retrieval (PIR) 2 Levente Buttyán Problem formulaton Alce wants to obtan nformaton from a database, but she does not want the database to learn whch nformaton she wanted e.g., Alce s an nvestor queryng a stock-market

More information

Learning an Image Manifold for Retrieval

Learning an Image Manifold for Retrieval Learnng an Image Manfold for Retreval Xaofe He*, We-Yng Ma, and Hong-Jang Zhang Mcrosoft Research Asa Bejng, Chna, 100080 {wyma,hjzhang}@mcrosoft.com *Department of Computer Scence, The Unversty of Chcago

More information

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Multi-stable Perception. Necker Cube

Multi-stable Perception. Necker Cube Mult-stable Percepton Necker Cube Spnnng dancer lluson, Nobuuk Kaahara Fttng and Algnment Computer Vson Szelsk 6.1 James Has Acknowledgment: Man sldes from Derek Hoem, Lana Lazebnk, and Grauman&Lebe 2008

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

Unsupervised Learning

Unsupervised Learning Pattern Recognton Lecture 8 Outlne Introducton Unsupervsed Learnng Parametrc VS Non-Parametrc Approach Mxture of Denstes Maxmum-Lkelhood Estmates Clusterng Prof. Danel Yeung School of Computer Scence and

More information

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices

An Application of the Dulmage-Mendelsohn Decomposition to Sparse Null Space Bases of Full Row Rank Matrices Internatonal Mathematcal Forum, Vol 7, 2012, no 52, 2549-2554 An Applcaton of the Dulmage-Mendelsohn Decomposton to Sparse Null Space Bases of Full Row Rank Matrces Mostafa Khorramzadeh Department of Mathematcal

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

Multimodal Learning of Geometry-Preserving Binary Codes for Semantic Image Retrieval

Multimodal Learning of Geometry-Preserving Binary Codes for Semantic Image Retrieval 600 IEICE TRANS. INF. & SYST., VOL.E100 D, NO.4 APRIL 2017 INVITED PAPER Specal Secton on Award-wnnng Papers Multmodal Learnng of Geometry-Preservng Bnary Codes for Semantc Image Retreval Go IRIE a), Hroyuk

More information

Determining the Optimal Bandwidth Based on Multi-criterion Fusion

Determining the Optimal Bandwidth Based on Multi-criterion Fusion Proceedngs of 01 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 5 (01) (01) IACSIT Press, Sngapore Determnng the Optmal Bandwdth Based on Mult-crteron Fuson Ha-L Lang 1+, Xan-Mn

More information

Hierarchical Image Retrieval by Multi-Feature Fusion

Hierarchical Image Retrieval by Multi-Feature Fusion Preprnts (www.preprnts.org) NOT PEER-REVIEWED Posted: 26 Aprl 207 do:0.20944/preprnts20704.074.v Artcle Herarchcal Image Retreval by Mult- Fuson Xaojun Lu, Jaojuan Wang,Yngq Hou, Me Yang, Q Wang* and Xangde

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

Lecture 4: Principal components

Lecture 4: Principal components /3/6 Lecture 4: Prncpal components 3..6 Multvarate lnear regresson MLR s optmal for the estmaton data...but poor for handlng collnear data Covarance matrx s not nvertble (large condton number) Robustness

More information

Improving Web Image Search using Meta Re-rankers

Improving Web Image Search using Meta Re-rankers VOLUME-1, ISSUE-V (Aug-Sep 2013) IS NOW AVAILABLE AT: www.dcst.com Improvng Web Image Search usng Meta Re-rankers B.Kavtha 1, N. Suata 2 1 Department of Computer Scence and Engneerng, Chtanya Bharath Insttute

More information

Human Face Recognition Using Generalized. Kernel Fisher Discriminant

Human Face Recognition Using Generalized. Kernel Fisher Discriminant Human Face Recognton Usng Generalzed Kernel Fsher Dscrmnant ng-yu Sun,2 De-Shuang Huang Ln Guo. Insttute of Intellgent Machnes, Chnese Academy of Scences, P.O.ox 30, Hefe, Anhu, Chna. 2. Department of

More information

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces

Range images. Range image registration. Examples of sampling patterns. Range images and range surfaces Range mages For many structured lght scanners, the range data forms a hghly regular pattern known as a range mage. he samplng pattern s determned by the specfc scanner. Range mage regstraton 1 Examples

More information

Recognizing Faces. Outline

Recognizing Faces. Outline Recognzng Faces Drk Colbry Outlne Introducton and Motvaton Defnng a feature vector Prncpal Component Analyss Lnear Dscrmnate Analyss !"" #$""% http://www.nfotech.oulu.f/annual/2004 + &'()*) '+)* 2 ! &

More information

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. *, NO. *, Dictionary Pair Learning on Grassmann Manifolds for Image Denoising

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. *, NO. *, Dictionary Pair Learning on Grassmann Manifolds for Image Denoising IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. *, NO. *, 2015 1 Dctonary Par Learnng on Grassmann Manfolds for Image Denosng Xanhua Zeng, We Ban, We Lu, Jale Shen, Dacheng Tao, Fellow, IEEE Abstract Image

More information

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique

The Greedy Method. Outline and Reading. Change Money Problem. Greedy Algorithms. Applications of the Greedy Strategy. The Greedy Method Technique //00 :0 AM Outlne and Readng The Greedy Method The Greedy Method Technque (secton.) Fractonal Knapsack Problem (secton..) Task Schedulng (secton..) Mnmum Spannng Trees (secton.) Change Money Problem Greedy

More information

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng

More information

Lobachevsky State University of Nizhni Novgorod. Polyhedron. Quick Start Guide

Lobachevsky State University of Nizhni Novgorod. Polyhedron. Quick Start Guide Lobachevsky State Unversty of Nzhn Novgorod Polyhedron Quck Start Gude Nzhn Novgorod 2016 Contents Specfcaton of Polyhedron software... 3 Theoretcal background... 4 1. Interface of Polyhedron... 6 1.1.

More information

Edge Detection in Noisy Images Using the Support Vector Machines

Edge Detection in Noisy Images Using the Support Vector Machines Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona

More information

TPL-Aware Displacement-driven Detailed Placement Refinement with Coloring Constraints

TPL-Aware Displacement-driven Detailed Placement Refinement with Coloring Constraints TPL-ware Dsplacement-drven Detaled Placement Refnement wth Colorng Constrants Tao Ln Iowa State Unversty tln@astate.edu Chrs Chu Iowa State Unversty cnchu@astate.edu BSTRCT To mnmze the effect of process

More information

A Posteriori Multi-Probe Locality Sensitive Hashing

A Posteriori Multi-Probe Locality Sensitive Hashing A Posteror Mult-Probe Localty Senstve Hashng Alexs Joly INRIA Rocquencourt Le Chesnay, 78153, France alexs.joly@nra.fr Olver Busson INA, France Bry-sur-Marne, 94360 obusson@na.fr ABSTRACT Effcent hgh-dmensonal

More information

Local Quaternary Patterns and Feature Local Quaternary Patterns

Local Quaternary Patterns and Feature Local Quaternary Patterns Local Quaternary Patterns and Feature Local Quaternary Patterns Jayu Gu and Chengjun Lu The Department of Computer Scence, New Jersey Insttute of Technology, Newark, NJ 0102, USA Abstract - Ths paper presents

More information

Semi-Supervised Discriminant Analysis Based On Data Structure

Semi-Supervised Discriminant Analysis Based On Data Structure IOSR Journal of Computer Engneerng (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 3, Ver. VII (May Jun. 2015), PP 39-46 www.osrournals.org Sem-Supervsed Dscrmnant Analyss Based On Data

More information

Reducing Frame Rate for Object Tracking

Reducing Frame Rate for Object Tracking Reducng Frame Rate for Object Trackng Pavel Korshunov 1 and We Tsang Oo 2 1 Natonal Unversty of Sngapore, Sngapore 11977, pavelkor@comp.nus.edu.sg 2 Natonal Unversty of Sngapore, Sngapore 11977, oowt@comp.nus.edu.sg

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

1. Introduction. Abstract

1. Introduction. Abstract Image Retreval Usng a Herarchy of Clusters Danela Stan & Ishwar K. Seth Intellgent Informaton Engneerng Laboratory, Department of Computer Scence & Engneerng, Oaland Unversty, Rochester, Mchgan 48309-4478

More information

Problem Set 3 Solutions

Problem Set 3 Solutions Introducton to Algorthms October 4, 2002 Massachusetts Insttute of Technology 6046J/18410J Professors Erk Demane and Shaf Goldwasser Handout 14 Problem Set 3 Solutons (Exercses were not to be turned n,

More information

Kent State University CS 4/ Design and Analysis of Algorithms. Dept. of Math & Computer Science LECT-16. Dynamic Programming

Kent State University CS 4/ Design and Analysis of Algorithms. Dept. of Math & Computer Science LECT-16. Dynamic Programming CS 4/560 Desgn and Analyss of Algorthms Kent State Unversty Dept. of Math & Computer Scence LECT-6 Dynamc Programmng 2 Dynamc Programmng Dynamc Programmng, lke the dvde-and-conquer method, solves problems

More information

Large-scale Web Video Event Classification by use of Fisher Vectors

Large-scale Web Video Event Classification by use of Fisher Vectors Large-scale Web Vdeo Event Classfcaton by use of Fsher Vectors Chen Sun and Ram Nevata Unversty of Southern Calforna, Insttute for Robotcs and Intellgent Systems Los Angeles, CA 90089, USA {chensun nevata}@usc.org

More information

Towards Semantic Knowledge Propagation from Text to Web Images

Towards Semantic Knowledge Propagation from Text to Web Images Guoun Q (Unversty of Illnos at Urbana-Champagn) Charu C. Aggarwal (IBM T. J. Watson Research Center) Thomas Huang (Unversty of Illnos at Urbana-Champagn) Towards Semantc Knowledge Propagaton from Text

More information

A Robust Method for Estimating the Fundamental Matrix

A Robust Method for Estimating the Fundamental Matrix Proc. VIIth Dgtal Image Computng: Technques and Applcatons, Sun C., Talbot H., Ourseln S. and Adraansen T. (Eds.), 0- Dec. 003, Sydney A Robust Method for Estmatng the Fundamental Matrx C.L. Feng and Y.S.

More information

A Novel Adaptive Descriptor Algorithm for Ternary Pattern Textures

A Novel Adaptive Descriptor Algorithm for Ternary Pattern Textures A Novel Adaptve Descrptor Algorthm for Ternary Pattern Textures Fahuan Hu 1,2, Guopng Lu 1 *, Zengwen Dong 1 1.School of Mechancal & Electrcal Engneerng, Nanchang Unversty, Nanchang, 330031, Chna; 2. School

More information

Learning-Based Top-N Selection Query Evaluation over Relational Databases

Learning-Based Top-N Selection Query Evaluation over Relational Databases Learnng-Based Top-N Selecton Query Evaluaton over Relatonal Databases Lang Zhu *, Wey Meng ** * School of Mathematcs and Computer Scence, Hebe Unversty, Baodng, Hebe 071002, Chna, zhu@mal.hbu.edu.cn **

More information

Outline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like:

Outline. Self-Organizing Maps (SOM) US Hebbian Learning, Cntd. The learning rule is Hebbian like: Self-Organzng Maps (SOM) Turgay İBRİKÇİ, PhD. Outlne Introducton Structures of SOM SOM Archtecture Neghborhoods SOM Algorthm Examples Summary 1 2 Unsupervsed Hebban Learnng US Hebban Learnng, Cntd 3 A

More information

Quality Improvement Algorithm for Tetrahedral Mesh Based on Optimal Delaunay Triangulation

Quality Improvement Algorithm for Tetrahedral Mesh Based on Optimal Delaunay Triangulation Intellgent Informaton Management, 013, 5, 191-195 Publshed Onlne November 013 (http://www.scrp.org/journal/m) http://dx.do.org/10.36/m.013.5601 Qualty Improvement Algorthm for Tetrahedral Mesh Based on

More information

Research Article A High-Order CFS Algorithm for Clustering Big Data

Research Article A High-Order CFS Algorithm for Clustering Big Data Moble Informaton Systems Volume 26, Artcle ID 435627, 8 pages http://dx.do.org/.55/26/435627 Research Artcle A Hgh-Order Algorthm for Clusterng Bg Data Fanyu Bu,,2 Zhku Chen, Peng L, Tong Tang, 3 andyngzhang

More information

A Bilinear Model for Sparse Coding

A Bilinear Model for Sparse Coding A Blnear Model for Sparse Codng Davd B. Grmes and Rajesh P. N. Rao Department of Computer Scence and Engneerng Unversty of Washngton Seattle, WA 98195-2350, U.S.A. grmes,rao @cs.washngton.edu Abstract

More information

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems A Unfed Framework for Semantcs and Feature Based Relevance Feedback n Image Retreval Systems Ye Lu *, Chunhu Hu 2, Xngquan Zhu 3*, HongJang Zhang 2, Qang Yang * School of Computng Scence Smon Fraser Unversty

More information

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints

Sum of Linear and Fractional Multiobjective Programming Problem under Fuzzy Rules Constraints Australan Journal of Basc and Appled Scences, 2(4): 1204-1208, 2008 ISSN 1991-8178 Sum of Lnear and Fractonal Multobjectve Programmng Problem under Fuzzy Rules Constrants 1 2 Sanjay Jan and Kalash Lachhwan

More information

KIDS Lab at ImageCLEF 2012 Personal Photo Retrieval

KIDS Lab at ImageCLEF 2012 Personal Photo Retrieval KD Lab at mageclef 2012 Personal Photo Retreval Cha-We Ku, Been-Chan Chen, Guan-Bn Chen, L-J Gaou, Rong-ng Huang, and ao-en Wang Knowledge, nformaton, and Database ystem Laboratory Department of Computer

More information

Fitting & Matching. Lecture 4 Prof. Bregler. Slides from: S. Lazebnik, S. Seitz, M. Pollefeys, A. Effros.

Fitting & Matching. Lecture 4 Prof. Bregler. Slides from: S. Lazebnik, S. Seitz, M. Pollefeys, A. Effros. Fttng & Matchng Lecture 4 Prof. Bregler Sldes from: S. Lazebnk, S. Setz, M. Pollefeys, A. Effros. How do we buld panorama? We need to match (algn) mages Matchng wth Features Detect feature ponts n both

More information

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law) Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes

More information

An Image Fusion Approach Based on Segmentation Region

An Image Fusion Approach Based on Segmentation Region Rong Wang, L-Qun Gao, Shu Yang, Yu-Hua Cha, and Yan-Chun Lu An Image Fuson Approach Based On Segmentaton Regon An Image Fuson Approach Based on Segmentaton Regon Rong Wang, L-Qun Gao, Shu Yang 3, Yu-Hua

More information

Positive Semi-definite Programming Localization in Wireless Sensor Networks

Positive Semi-definite Programming Localization in Wireless Sensor Networks Postve Sem-defnte Programmng Localzaton n Wreless Sensor etworks Shengdong Xe 1,, Jn Wang, Aqun Hu 1, Yunl Gu, Jang Xu, 1 School of Informaton Scence and Engneerng, Southeast Unversty, 10096, anjng Computer

More information

The Discriminate Analysis and Dimension Reduction Methods of High Dimension

The Discriminate Analysis and Dimension Reduction Methods of High Dimension Open Journal of Socal Scences, 015, 3, 7-13 Publshed Onlne March 015 n ScRes. http://www.scrp.org/journal/jss http://dx.do.org/10.436/jss.015.3300 The Dscrmnate Analyss and Dmenson Reducton Methods of

More information

TN348: Openlab Module - Colocalization

TN348: Openlab Module - Colocalization TN348: Openlab Module - Colocalzaton Topc The Colocalzaton module provdes the faclty to vsualze and quantfy colocalzaton between pars of mages. The Colocalzaton wndow contans a prevew of the two mages

More information

Learning a Class-Specific Dictionary for Facial Expression Recognition

Learning a Class-Specific Dictionary for Facial Expression Recognition BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 16, No 4 Sofa 016 Prnt ISSN: 1311-970; Onlne ISSN: 1314-4081 DOI: 10.1515/cat-016-0067 Learnng a Class-Specfc Dctonary for

More information

User Authentication Based On Behavioral Mouse Dynamics Biometrics

User Authentication Based On Behavioral Mouse Dynamics Biometrics User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA

More information

Multi-View Surveillance Video Summarization via Joint Embedding and Sparse Optimization

Multi-View Surveillance Video Summarization via Joint Embedding and Sparse Optimization IEEE TRANSACTIONS ON MULTIMEDIA, VOL. XX, NO. XX, DECEMBER 20XX 1 Mult-Vew Survellance Vdeo Summarzaton va Jont Embeddng and Sparse Optmzaton Rameswar Panda and Amt K. Roy-Chowdhury, Senor Member, IEEE

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

Journal of Process Control

Journal of Process Control Journal of Process Control (0) 738 750 Contents lsts avalable at ScVerse ScenceDrect Journal of Process Control j ourna l ho me pag e: wwwelsevercom/locate/jprocont Decentralzed fault detecton and dagnoss

More information

Two-Dimensional Supervised Discriminant Projection Method For Feature Extraction

Two-Dimensional Supervised Discriminant Projection Method For Feature Extraction Appl. Math. Inf. c. 6 No. pp. 8-85 (0) Appled Mathematcs & Informaton cences An Internatonal Journal @ 0 NP Natural cences Publshng Cor. wo-dmensonal upervsed Dscrmnant Proecton Method For Feature Extracton

More information

Steps for Computing the Dissimilarity, Entropy, Herfindahl-Hirschman and. Accessibility (Gravity with Competition) Indices

Steps for Computing the Dissimilarity, Entropy, Herfindahl-Hirschman and. Accessibility (Gravity with Competition) Indices Steps for Computng the Dssmlarty, Entropy, Herfndahl-Hrschman and Accessblty (Gravty wth Competton) Indces I. Dssmlarty Index Measurement: The followng formula can be used to measure the evenness between

More information

Hermite Splines in Lie Groups as Products of Geodesics

Hermite Splines in Lie Groups as Products of Geodesics Hermte Splnes n Le Groups as Products of Geodescs Ethan Eade Updated May 28, 2017 1 Introducton 1.1 Goal Ths document defnes a curve n the Le group G parametrzed by tme and by structural parameters n the

More information

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration Improvement of Spatal Resoluton Usng BlockMatchng Based Moton Estmaton and Frame Integraton Danya Suga and Takayuk Hamamoto Graduate School of Engneerng, Tokyo Unversty of Scence, 6-3-1, Nuku, Katsuska-ku,

More information

Hierarchical clustering for gene expression data analysis

Hierarchical clustering for gene expression data analysis Herarchcal clusterng for gene expresson data analyss Gorgo Valentn e-mal: valentn@ds.unm.t Clusterng of Mcroarray Data. Clusterng of gene expresson profles (rows) => dscovery of co-regulated and functonally

More information

Unsupervised Co-segmentation of 3D Shapes via Functional Maps

Unsupervised Co-segmentation of 3D Shapes via Functional Maps Unsupervsed Co-segmentaton of 3D Shapes va Functonal aps Jun Yang School of Electronc and Informaton Engneerng, Lanzhou Jaotong Unversty, Lanzhou 730070, P. R. Chna yangj@mal.lzjtu.cn Zhenhua Tan School

More information

A Deflected Grid-based Algorithm for Clustering Analysis

A Deflected Grid-based Algorithm for Clustering Analysis A Deflected Grd-based Algorthm for Clusterng Analyss NANCY P. LIN, CHUNG-I CHANG, HAO-EN CHUEH, HUNG-JEN CHEN, WEI-HUA HAO Department of Computer Scence and Informaton Engneerng Tamkang Unversty 5 Yng-chuan

More information

Feature Extractions for Iris Recognition

Feature Extractions for Iris Recognition Feature Extractons for Irs Recognton Jnwook Go, Jan Jang, Yllbyung Lee, and Chulhee Lee Department of Electrcal and Electronc Engneerng, Yonse Unversty 134 Shnchon-Dong, Seodaemoon-Gu, Seoul, KOREA Emal:

More information

Fuzzy C-Means Initialized by Fixed Threshold Clustering for Improving Image Retrieval

Fuzzy C-Means Initialized by Fixed Threshold Clustering for Improving Image Retrieval Fuzzy -Means Intalzed by Fxed Threshold lusterng for Improvng Image Retreval NAWARA HANSIRI, SIRIPORN SUPRATID,HOM KIMPAN 3 Faculty of Informaton Technology Rangst Unversty Muang-Ake, Paholyotn Road, Patumtan,

More information

Deep Classification in Large-scale Text Hierarchies

Deep Classification in Large-scale Text Hierarchies Deep Classfcaton n Large-scale Text Herarches Gu-Rong Xue Dkan Xng Qang Yang 2 Yong Yu Dept. of Computer Scence and Engneerng Shangha Jao-Tong Unversty {grxue, dkxng, yyu}@apex.sjtu.edu.cn 2 Hong Kong

More information

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

Face Recognition using 3D Directional Corner Points

Face Recognition using 3D Directional Corner Points 2014 22nd Internatonal Conference on Pattern Recognton Face Recognton usng 3D Drectonal Corner Ponts Xun Yu, Yongsheng Gao School of Engneerng Grffth Unversty Nathan, QLD, Australa xun.yu@grffthun.edu.au,

More information

Transductive Regression Piloted by Inter-Manifold Relations

Transductive Regression Piloted by Inter-Manifold Relations Huan Wang IE, The Chnese Unversty of Hong Kong, Hong Kong Shucheng Yan Thomas Huang ECE, Unversty of Illnos at Urbana Champagn, USA Janzhuang Lu Xaoou Tang IE, The Chnese Unversty of Hong Kong, Hong Kong

More information