Extracting Adaptive Contextual Cues from Unlabeled Regions

Size: px
Start display at page:

Download "Extracting Adaptive Contextual Cues from Unlabeled Regions"

Transcription

1 Extractng Adaptve Contextual Cues from Unlabeled Regons Congcong L Cornell Unversty Dev Parkh Toyota Technologcal Insttute Chcago Tsuhan Chen Cornell Unversty cl758@cornell.edu dparkh@ttc.edu tsuhan@ece.cornell.edu Abstract Exstng approaches to contextual reasonng for enhanced object detecton typcally utlze other labeled categores n the mages to provde contextual nformaton. As a consequence, they nadvertently commt to the granularty of nformaton mplct n the labels. Moreover, large portons of the mages may not belong to any of the manuallychosen categores, and these unlabeled regons are typcally neglected. In ths paper, we overcome both these drawbacks and propose a contextual cue that explots unlabeled regons n mages. Our approach adaptvely determnes the granularty (scene, nter-object, ntra-object, etc.) at whch contextual nformaton s captured. In order to extract the proposed contextual cue, we consder a scene to be a structured confguraton of objects and regons; just as an object s a composton of parts. We thus learn our proposed contextual meta-objects usng any off-the-shelf object detector, whch makes our proposed cue wdely accessble to the communty. Our results show that ncorporatng our proposed cue provdes a relatve mprovement of 12% over a state-of-the-art object detector on the challengng PASCAL dataset. 1. Introducton Object recognton s one of the central problems n computer vson. Many recent works leverage contextual nformaton surroundng the object-of-nterest for enhanced recognton [5, 11, 12, 14, 19, 22]. These typcally leverage labeled data from other object categores to learn contextual relatonshps. Ths leads to two undesrable consequences. Unlabeled regons: Frst, regons n the mages that are unlabeled are often neglected. In most scenaros such as the popular MSRC [1] or PASCAL [7] datasets, not all regons n the mages are accounted for by the manually chosen categores that are labeled. For nstance, 28.18% of the pxels n MSRC and 54.74% of pxels n the PASCAL 2007 dataset are not a part of the labeled categores. See Fgure 1. Do these unlabeled regons contan useful nformaton that s worth capturng? We conduct human studes on recognzng objects n low-resoluton mages1 from the PASCAL 1 Parkh et al. [19] showed that humans need contextual nformaton for recognton only when the appearance nformaton s mpovershed, such context for aeroplane context for potted-plant context for dnng-table Fgure 1. Many approaches to contextual reasonng for object detecton model the relatonshp of the object-of-nterest (yellow boxes) to other object categores labeled n the dataset. They do not leverage unlabeled regons n the mages that do not belong to these manually chosen labeled categores, resultng n a hghly myopc vew of the scene (1st and 3rd row). Our approach (blue boxes) ntellgently leverages nformaton present n these unlabeled regons to better detect the object of nterest. From left to rght: unlabeled regons lke sky and grass provde context for aeroplanes, and unlabeled objects lke wndows and pantngs are contextually relevant for potted-plants and dnng-tables respectvely. The granulartes of our blue boxes relatve to yellow boxes are learned adaptvely. dataset. We fnd that subjects perform better n scenaros where they see the entre mage ncludng the unlabeled regons (Fgure 2). Ths ndcates that the unlabeled regons, whch are often dscarded, ndeed contan useful contextual nformaton. Adaptve granularty: Second, n focusng only on labeled regons n the mage, models nadvertently lmt the contextual nformaton captured to the granularty mplct n the labels. For nstance, most works explorng contextual models for the PASCAL dataset consder only objectlevel nformaton, such as co-occurrence, relatve locaton and relatve sze [5, 9]. We argue that the granularty at whch contextual nformaton s most helpful s category specfc. Whle an aeroplane may only need to consder a as n low-resoluton mages.

2 58 Recognton Accuracy (%) Excludng Unlabeled Entre Image Fgure 2. Human subjects were shown mages excludng the unlabeled regons (left), as well as entre mages (mddle). The object to be recognzed s shown wth and wthout a yellow-outlne (to avod dstracton). Our results (rght) ndcate that subjects can recognze objects sgnfcantly more relably f nformaton from the unlabeled regons s avalable. These experments were conducted on 394 PASCAL 2007 mages contanng 897 objects from 20 categores. neghborng sky regon as context, dnng-table can beneft from the entre scene layout. Moreover, whle a sheep may be surrounded by relevant contextual nformaton all around t, the most relable contextual nformaton for a bcycle s the person on top. Therefore, t s crucal that context s extracted from regons that adapt to dfferent objects. In ths work, we overcome both these drawbacks and propose a contextual cue that explots unlabeled regons n mages and automatcally adapts to each object category. Formulaton: What knd of nformaton s contextually most useful? Informaton that s relevant.e. has consstent spatal locaton wth respect to the object-of-nterest, and nformaton that s relable.e. has consstent appearance across mages for relable detecton. Interestngly, these are also aspects that make object models effectve: capturng spatally coherent and vsually consstent object-parts. Instantatng the popular herarchcal vew of scenes [18, 20, 25] (parts are to objects as objects are to scenes), we can cast the problem of learnng relevant contextual regons n a scene as that of learnng detectors for objects, whch we call contextual meta-objects (CMOs). These CMOs form our proposed contextual cues. Any exstng object detector can be used to learn our CMOs. In fact, the detector that one trans to detect objects-of-nterest (OOIs) can be seamlessly (and convenently so!) used to learn CMOs, essentally boostng the performance of the OOI detector wthout desgnng complex algorthms or cumbersome learnng procedures. Summary of approach: How can we extract meanngful contextual nformaton from unlabeled regons? Our approach explots the followng key observaton: object boundng boxes do not smply provde us wth nformaton about what the object looks lke whch can be used to tran a detector for the object. The presence of an object at a certan locaton n a natural mage also provdes an anchor pont that suggests a meanngful algnment among the scenes contanng these objects. Gven mages labeled wth boundng-boxes of the OOI, we algn and cluster the mages such that each cluster contans mages wth smlar contextual nformaton surroundng the OOI. We dentfy a CMO regon around the OOI that best captures ths contextual nformaton. Note that the granularty at whch the CMO regon s defned s not fxed, and adapts to the content of the mages. Any off-the-shelf object detector can then be traned to detect our CMOs. Durng testng, we apply the CMO detector on the test mage, and the score of the detecton captures our contextual cue. Ths cue can be combned wth the OOI detecton score, or other contextual cues for enhanced OOI detecton. Contrbutons: In ths work, we effectvely explot the unlabeled regons n mages that are often neglected by most exstng works, to extract contextually relevant cues for enhanced object detecton. Our contrbutons are threefold. Frst, we dscover contextual regons that automatcally adapt to the object category of nterest n order to capture contextual nteractons at varyng granulartes (the entre scene, nter-object, even ntra-object) for dfferent categores. Second, we cast the problem of extractng contextual regons n scenes nto the problem of learnng object models. Ths allows us to employ any off-the-shelf object detector to learn our contextual cue; a convenent choce beng the object-of-nterest detector whose performance we hope to enhance va contextual reasonng. Ths smplcty makes our proposed cue easly accessble to the communty. Lastly, our approach acheves hgher detecton performance when usng a sngle labeled category, than the state-of-theart approach [9] that utlzes labels for all 20 object categores on the PASCAL 2007 dataset. Ths demonstrates our effectve use of unlabeled regons. We use our approach to boost performance of several object detectors, and show that our contextual cue compares favorably to, and complements other sources of context. Our adaptve selecton of the granularty of contextual nformaton outperforms the fxed-granularty counterpart. 2. Related Work Many recent works leverage contextual nformaton for enhanced recognton and localzaton of objects n natural mages. Varous sources of context have been explored, rangng from the global scene layout, nteractons between objects and regons, as well as local features. Dvvala et al. [6] survey and study the effectveness of dfferent contextual cues and combne them to acheve superor performance. We vew our novel cue that extracts useful contextual nformaton from unlabeled regons as complementary to exstng cues, and can be easly ntegrated n most contextual models. Fxed-granularty models: Many exstng works commt to a fxed granularty of contextual nformaton. To ncorporate scene-level nformaton, Torralba et al. [26] use the statstcs of low-level features across the entre scene to prme object detecton. Hoem et al. [13] use 3D scene nformaton to provde prors on potental object locatons. Park et al. [21] use the ground plane estmaton as contextual cues for pedestran detecton. Probablstc models have

3 been proposed to capture the local nteractons between neghborng regons [12, 14, 17], objects [5, 9, 19, 22, 30], or both [2, 10]. Sadegh et al. [24] propose and detect vsual phrases whch correspond to chunks of meanng bgger than objects and smaller than scenes. Whle the vsual phrases n [24] are manually labeled n the dataset, our work can be thought of as learnng the vsual phrases n an unsupervsed way. Our learned compostes, however, may not have a clearly defned semantc meanng. Whle most works focus on one level of nteracton, Gallegullos et al. [10] explore contextual nteractons at multple levels. However, ths mult-level aspect s ganed by explctly usng dfferent models for each nteracton level. In contrast, our approach allows for adaptvely pckng dfferent granulartes of contextual nformaton wthn the same framework. Leveragng unlabeled nformaton: A natural way to ncorporate nformaton from unlabeled regons n mages s to buld a global descrptor for the entre mage to provde contextual cues. Though global mage statstcs show great potental n prmng object detecton [26], very modest mprovement can be acheved when appled to datasets lke PASCAL (also confrmed n our experments), where mages have poor algnments due to the hgh varance n object scales, poses, etc. [3]. Instead, local neghborng regons tend to be more effectve. Wolf et al. [29] sample contextual nformaton from pre-defned relatve locatons. Dalal et al. [4] show that smply ncreasng the sze of the person boundng box by a small amount boosts the accuracy of pedestran detecton. Ths s equvalent to leveragng potentally unlabeled regons n close proxmty of the object-of-nterest. Our approach nstead automatcally and adaptvely determnes the extent of contextual nformaton to be captured around dfferent object categores. Felzenszwalb et al. [9] often detect parts that le slghtly outsde the sldng wndow, and use these detectons to refne ther boundng box predcton. Lee et al. [15] utlze a few labeled categores to dscover other object categores n the background that have consstent appearance and are contextually coherent wth the labeled categores. Whle smlar n phlosophy to our work, our soluton s qute dfferent, as s the problem settng of enhancng object detecton. 3. Approach In order to extract contextual cues that explot unlabeled regons n mages, we work n the most extreme settng where only one object category s labeled (wth boundngboxes) n the tranng dataset, leavng a large porton of the mages unlabeled. We can seamlessly ncorporate more labeled categores n our approach, as we descrbe later. Our goal then s to extract useful contextual regons, gven mages labeled wth ground-truth boundng-boxes for just one object category.e. the object-of-nterest (OOI). The ground-truth OOI may occupy dfferent proportons of (a) (b) Fgure 3. Context around objects (potted-plant and dog shown n yellow boxes) can vary n extent (left) as well as content (rght). mages across the dataset, as seen n Fgure 3(a). Moreover, mages may exhbt dfferent contextual settngs, as seen n Fgure 3(b). To better model these, we frst group the tranng mages that exhbt smlar contextual extent and content (Sectons 3.1 and 3.2). As stated earler, we cast our problem of extractng a contextual regon n a scene as that of learnng an object model, whch we call contextual metaobject (CMO), for whch any off-the-shelf detector can be used. Each cluster of mages s used to tran a dfferent component of our CMO detector (Secton 3.3). Durng testng, we run both the traned OOI and CMO detectors, and the score of the latter provdes contextual nformaton for the former. Whle any contextual reasonng model can be used to ntegrate the two, n our mplementaton we adopt the smple contextual re-scorng scheme from [9] (Secton 3.4) Contextual-extent-based clusterng We consder the contextual-extent of an mage to be the porton of the mage that les outsde of and surrounds the OOI. Intutvely, we wsh to group all mages that contan the OOI wth smlar poses, at smlar scales, and n smlar locatons wth respect to the rest of the scene. We use all the tranng mages, as well as ther leftrght flpped versons. To enhance consstency n the data, we frst dvde the mages nto 2 groups: one contanng mages wth the OOI ground-truth boundng boxes on the rght, and the other contanng mages wth the OOI on the left. We then employ the followng procedure for both groups. Consder an mage I consstng of nb groundtruth OOI boundng-boxes {Bj }, j {1,..., nb }. I s descrbed by a set of nb fve-dmensonal descrptors F = {f1,..., fj,..., fnb } fj = # " xtl x j ytl y j xbr x j ybr y j hj (1) sj sj sj sj wj where xtl, ytl and xbr, ybr are the co-ordnates of the top-left and bottom-rght corners of the mage I respec tvely, x j, y j are the co-ordnates of the center of the j th ground-truth OOI boundng-box n mage I, and sj, hj, and wj are the scale, heght and wdth of the j th OOI boundng-box respectvely. Ths descrptor captures the extent of the contextual nformaton n the mage (n terms of relatve locaton and scale) wth respect to the OOI present n the mage, as well as the aspect rato of the OOI. For each of the groups mentoned above, we cluster the descrptors N =1 F collected from all N tranng mages contanng the

4 OOI nto K clusters usng k-means clusterng. Ths results n a total of 2K context-extent based clusters of mages. In our mplementaton, we use K = 3. For all mages n the k th cluster, k {1,..., K}, we determne the extent of the CMO to be the largest boundngbox surroundng the OOI such that the CMO s entrely contaned wthn at least 80% of the mages n the cluster. Ths boundng-box ndcates the presence and extent of the CMO we wll learn va an off-the-shelf object detector. We note that snce the extent of the CMO s not ted to the OOI boundng-box tself, and nstead depends on the layout of the OOI n the scene, t can freely capture any relevant nformaton n the mage, ncludng unlabeled regons, at any granularty. In fact, CMOs correspondng to the dfferent clusters capture contextual nformaton at dfferent granulartes for the same object category. We note that a tranng mage I contanng n B ground-truth OOI nstances wll have n B correspondng nstances of the CMO for tranng Contextual-content-based clusterng Whle we ensure that the extent of context s smlar among the mages wthn the 2K clusters, the contextual settng or content of the mages could be qute vared (Fgure 3(b)). We further cluster each of the 2K groups based on ther content. We extract gst [27] features wthn the CMO box, and perform k-means clusterng on these features to dvde each of the 2K groups nto two clusters. In our mplementaton, when less than 40 mages are assgned to a cluster, we drop the cluster and re-assgn the correspondng mages to the remanng clusters. Therefore, we have M 4K clusters, whch vares wth categores. For the PASCAL VOC 2007 dataset, we fnd a total of 198 clusters across the 20 categores. These clusters now have CMOs defned that are consstent n appearance as well as spatal relatonshps wth respect to the OOI, and thus have the potental to provde useful contextual nformaton to enhance the OOI detecton Tranng CMO detectors We now descrbe how we use the above clusters to learn our CMO detector. As stated earler, we can use any offthe-shelf object detector, whch we treat as a black-box parameterzed by a model θ, learnt va a tranng procedure f tran that takes n tranng data of the form (P, N ). P = {(I 1 +, B 1),..., (I + k, B k)} s a set of postve mage and boundng-box pars, and N = {(I1 ),..., (I l )} s a set of negatve mages. The tranng procedure can be vewed as ˆθ = optmze f tran (P, N ; θ). (2) The detector can then be evaluated on a test mage I, va the nference procedure f nfer to obtan a detecton (B, s) ncludng a boundng box B and a score s (B, s) = f nfer I; ˆθ. (3) We see that the above formulaton holds for a varety of detectors, be t sldng-wndow based [4, 9, 28] or houghtransform based [16]. So how do we use one of these black-box detectors to tran our CMO detector usng the M clusters formed n Secton 3.2? Instead of commttng to the hard-clusterng whch was blnd to the choce of detector that would follow, we tran an M-component detector usng an EM style approach, smlar to the strategy employed n [9]. We ntalze the components usng the above clusterng, and tran M detectors to obtan ˆθ m usng f tran (P m, N m ; θ), where P m s the set of mages n the m th cluster and the CMO wndows contaned (Secton 3.1), and N m are negatve mages. We then nfer these M components on all postve tranng mages (across all M clusters), and re-assgn each ground-truth CMO box B (Secton 3.1) n the tranng set to the component that best explans t: m = argmax (s m t m ), (4) m {1,...,M} s.t.(b m, s m ) =f nfer I; ˆθ m, overlap(b m, B) > 0.5 (5) where t m s a bas term used to normalze the scores across the components, and s set to the 5 th percentle of the scores assgned by component m to all detectons n the tranng mages. Each component m s now re-traned, but wth the new set of postve examples, to obtan an updated ˆθ m, and the teratons contnue. We fnd that 3-5 teratons suffce. Durng testng, all M components are evaluated on the test mage. Non-maxmal suppresson s used on the resultant detectons to elmnate repeated detectons Contextual re-scorng We now descrbe how we use our CMO to provde context to the OOI detecton. We evaluate our CMO as well as the OOI detectors on the test mage. The presence as well as the locaton of the CMO can provde useful contextual nformaton. However n ths work, we only consder the presence.e. the CMO detecton score. Based on the CMO detectons, we wsh to re-score the detected OOI boundng boxes. Whle any contextual reasonng mechansm can be used to ths end (such as [6, 9, 11, 23]), we tran a classfer, smlar to Felzenszwalb et al. [9]. Let D be the set of detectons obtaned for the OOI detector. Each detecton (B, s), (B, s) D s formed of a boundng-box wth co-ordnates B = (x tl, y tl, x br, y br ) (overloaded from prevous sectons) and a score s. Let s C be the score of the hghest scorng CMO detecton (across all M components) found n the mage. To re-score an OOI detecton (B, s), we buld a 6-dmensonal descrptor consstng of the orgnal score of the OOI detecton, the top-left and bottom-rght boundng box coordnates normalzed by mage sze, and the contextual nformaton provded by the CMO detecton, captured va s C : ψ 1c = [σ(s) x tl y tl x br y br σ(s C)] (6) where σ(x) = 1/ (1 + exp( αx)), α = 1.5, as n [8].

5 aero bke brd boat bottle bus car cat char cow table dog horse mbk pers plant sheep sofa tran tv MEAN GAIN Base w/o context [9] Scene ( [26]) % EXO ( [4]) % CMO % Table 1. Average precson (AP) for all 20 categores n PASCAL VOC 2007, mean AP across 20 categores, and the average relatve mprovement on 20 categores compared to the Base method. All methods lsted use labels from only one object category. ( [ ]) means the method s smlar n sprt to the reference work. Ths ψ 1c descrptor s fed to a classfer h traned to separate correct OOI detectons from false postves. The OOI boundng-box B s assgned a new score s = h (ψ 1c ). In our mplementaton, the classfer h s an SVM wth a polynomal kernel (parameters set va cross-valdaton), smlar to [8, 9]. The tranng data are obtaned by runnng the traned OOI detector on labeled tranng data, and collectng correct detectons and false postves. We note that ψ 1c uses labeled data from only one object category, and stll captures contextual nformaton. Our approach s not restrcted to usng labels from only one object category. If more object categores are avalable, they can be seamlessly ncorporated. A CMO detector would be traned for each labeled object category. Durng test tme, let (D 1,..., D n ) be the set of OOI detectons obtaned for n dfferent object categores (for PASCAL n = 20 f all categores are consdered). Let (s C1,..., s Cn ) be the scores of the hghest scorng CMO detectons for each of the correspondng n CMO detectors. The contextual descrptor to re-score an OOI detecton (B, s) D s now n + 5 dmensonal ψ nc = [σ(s) x tl y tl x br y br σ(s C1 )... σ(s Cn )]. (7) Smlar to tradtonal approaches that explot context, Felzenszwalb et al. [9] only use other OOI object categores to provde context, va a descrptor ψ no = [σ(s) x tl y tl x br y br σ(s D1 )... σ(s Dn )] (8) where s D s the score of the hghest-scorng OOI detecton from the th OOI category. We note that ths descrptor s dentcal to the one used n [9], whch we compare to n our experments. Fnally, a contextual descrptor capturng contextual nformaton provded by all n CMO and OOI detectors s gven as ψ nco = [σ(s) x tl y tl x br y br γ(c) γ(d)] (9) γ(c) = [σ(s C1 )... σ(s Cn )] (10) γ(d) = [σ(s D1 )... σ(s Dn )]. (11) A specal case of Equaton 9 for n = 1 dffers from Equaton 6 by one-dmenson correspondng to σ(s D ), the score correspondng to the hghest-scorng OOI detecton. Snce the use of σ(s D ) does not requre addtonal tranng data, and s obtaned by usng labeled data from a sngle objectcategory, we replace Equaton 6 wth the followng n our experments. ψ 1co = [σ(s) x tl y tl x br y br σ(s C) σ(s D)] (12) CMO Rato OOI 37.2% Other Categores Unlabeled Regons 17.5% 45.3% Number of CMO components Proporton of a CMO component occuped by OOI Fgure 4. Left: The proporton of hghest-scorng CMO detectons on postve testng mages occuped by dfferent content. Rght: The average proporton of a CMO component detecton occuped by OOI. The truly mult-granular nature of our learnt contextual cues s evdent. 4. Experments and Results We evaluate our approach usng the PASCAL VOC challenge dataset and protocol [7], whch contans 9963 mages of realstc scenes, contanng ground-truth boundng boxes for 20 object categores. Whle we provde results wth other object detectors n subsequent experments, we frst perform several comparsons and analyses usng the publcly avalable mplementaton [8] of the state-of-the-art deformable parts-based object detector [9] Quanttatve results We frst provde several quanttatve evaluatons, followed by some qualtatve llustratons. Useful contextual nformaton from unlabeled regons: We frst evaluate whether our proposed cue captures useful contextual nformaton extracted from the unlabeled regons. We compare the performance of the baselne OOI detector to the contextually re-scored detector, usng our proposed cue as the source of contextual nformaton as descrbed n Equaton 12. The results can be seen n Table 1 (Base w/o context vs. CMO). We see that across all 20 categores, our method outperforms the state-of-the-art OOI detector, ndcatng that our contextual cue does n fact extract useful contextual nformaton from unlabeled regons. Further, n Fgure 4 (left) we show the average proporton of the CMO detectons occuped by dfferent contents (OOI, other labeled categores, and unlabeled). We see that almost half of the CMO detectons cover unlabeled regons. We also see that about 1/5 of the CMO detectons capture 2 We are not proposng a novel object detector. Instead, we wll be performng a varety of comparsons to demonstrate the effectveness of our proposed contextual cue. As recommended by the challenge organzers [7], we work wth the 2007 test-set, whch s the latest PASCAL test-set wth publcly avalable annotatons.

6 aero bke brd boat bottle bus car cat char cow table dog horse mbk pers plant sheep sofa tran tv MEAN GAIN 20OOI ([9]) % 20CMO+20OOI % Table 2. Average precson (AP) for all 20 categores n PASCAL VOC 2007, mean AP across 20 categores, and the average relatve mprovement on 20 categores compared to the Base method n Table 1). All methods lsted use labels from all 20 categores. # of labels for tranng fuson methods mean AP Scene+EXO CMO 34.2 Scene+EXO+CMO OOI OOI+20CMO OOI+Scene+EXO OOI+20CMO+Scene+EXO 35.3 Table 3. The mean AP across the 20 categores n PASCAL VOC 2007 for fusng varous sources of contextual nformaton. other labeled categores n the mages, even though our contextual cue does not use these labels to explctly learn contextual relatonshps between the OOI and other categores. Adaptve scalng helps: We now compare our contextual cue to two cues that also explot the unlabeled regons but at a fxed granularty. The frst cue ( Scene ) captures the entre scene, and thus operates at the global granularty, smlar to the work of Torralba et al. [26]. We tran a bnary RBF-kernel SVM classfer on the gst descrptors [27] extracted from the entre mages to dscrmnate between mages wth and wthout the object-of-nterest. We then use the same re-scorng scheme n Equaton 12, by replacng the σ(s C ) wth σ(s g ), where s g s the score of the gst-based classfer for the test mage. The second cue ( EXO ) s local n nature. We expand the OOI boundng-box by a fxed amount (smlar to [4]) of 20% n all four drectons. Instead of tranng our CMO on the co-ordnates determned n Secton 3.1, we use ths expanded OOI boundng-box to learn a contextual object whch we call EXO. We use the same re-scorng scheme n Equaton 12, by replacng σ(s D ) wth σ(s E ), where s E s the hghest score of the EXO detectons. Note all three approaches only requre labels from a sngle object category. Table 1 shows that our approach CMO outperforms both fxed-granularty methods. Ths demonstrates our ablty to effectvely determne the granularty of useful contextual nteractons. Further, Fgure 4 (rght) shows the dstrbuton of the average proporton of CMO component detectons occuped by OOI. Hgh values (rght of the hstogram) correspond to contextual cues that capture local context around the OOI, whle low values (left of the hstogram) correspond to models that capture scene level context wth respect to a relatvely small OOI. The large varance n the dstrbuton demonstrates the truly mult-granular nature of the contextual nformaton learnt. Complementary cue: We now test the ablty of our proposed cue to provde complementary contextual nformaton. We consder several dfferent sources of context popularly explored n lterature: the global scene-level context ( Scene ) and local context ( EXO ) descrbed above, as Mean AP over 20 categores n OOI + n CMO (Our algorthm) n OOI (Felzenszwalb et. al [9]) No. of category labels for learnng contextual nformaton (n) Fgure 5. The effect of number of labeled categores on the average AP (across 20 categores). well as object-level context provded by other labeled objects n the mages, be t a subset of the categores (nooi) or all 20 (20OOI). In Table 1 we saw that CMO performs better than the global and local context ndvdually. Table 3 shows that our proposed cue learnt from a sngle labeled category s comparable to (slghtly outperforms) the contextual nformaton provded by all 20 labeled object categores (20OOI). We note that 20OOI corresponds exactly to the contextual approach used by Felzenszwalb et al. [9]. Hence, we see that our contextual cue performs better than any of the ndvdual sources of context, ncludng ones that utlze sgnfcantly more amounts of labels. Smlar to [6], we analyze the performance of fusng varous sources of contextual nformaton. We use the same rescorng method as descrbed n Secton 3.4 for combnng the dfferent cues, snce we wsh to evaluate the nformaton captured by the cues, and not partcular contextual reasonng technques. Table 3 shows results of fusng dfferent combnatons of the above mentoned contextual cues. We append the hghest score for each contextual cue beng fused to re-score the OOI detecton (smlar to Equaton 9). We see that our proposed cue ndvdually performs better than the combnaton of both global and local context (Scene+EXO). Moreover, fusng our cue to these exstng sources of context provdes a further boost n performance, demonstratng the truly complementary nature of our cue that extracts contextual nformaton from unlabeled regons at adaptve granulartes. We now compare our approach that leverages unlabeled regons to learn contextual cues (nooi + ncmo, Equaton 9), to the baselne approach of [9] (nooi, Equaton 8) that only utlzes other labeled categores, as the number of labeled categores s vared. For each object category (OOI), we pck n categores wth hghest mutual nformaton (based on co-occurrence of categores wth the OOI category across tranng data) to provde contextual nformaton. Fgure 5 shows the trends. The green pont at n = 0

7 ISM [16] HOG-SVM [4] Part-based Model [9] OOI (mean AP) OOI+CMO (mean AP) Gan 23.8% 11.2% 8.4% Table 4. Detecton results wth and wthout the proposed CMO as addtonal contextual nformaton, by usng dfferent black-box detectors. [ ] ndcates t s our mplementaton of the reference work. gves the mean AP of the OOI detector usng no context. We can see that across the board, our approach leads to better AP than [9] whle usng the same amount of labeled data, demonstratng our effectve use of unlabeled regons. A break-down of the average AP across categores when usng all 20 labeled categores for both methods can be seen n Table 2. We see that ncorporatng our contextual cue that leverages unlabeled nformaton n addton to the 20 labeled object categores (20OOI+20CMO) provdes mprovements over 20OOI ([9]) n 19 out of 20 categores and matches the remanng category. We see that our contextual cue on average provdes a relatve mprovement of 12.05% over the state-of-the-art detector. We observe sgnfcant relatve mprovements of more than 20% n some categores such as brd, cat, dnng-table, and dog. Other detectors: As mentoned n Secton 3, our proposed contextual cue can be extracted va any object detector, and can nturn be used to enhance the performance of the detector for the OOI. We demonstrate that here. In addton to the sldng wndow part-based deformable model (Parts-based Model) by Felzenszwalb et al. [9] that we use n the above experments, we consder two other popular detectors: Implct Shape Model (ISM) [16] whch s houghtransform based, and the HOG-SVM detector [4] (also sldng wndow). These are used as the black-box modules to tran our CMO detectors, usng the procedure descrbed n Secton 3.3. Table 4 shows the mean AP across all 20 categores n the PASCAL 2007 dataset, acheved by our mplementaton of these detectors, wth and wthout the context provded by CMO. The results show that our proposed contextual cue CMO can be learnt va any detector, and consstently mproves the object detecton performance wth a sgnfcant gan. Note that we use the same detectors for both OOI and CMO, ndcatng that no addtonal technques are requred for achevng these performance gans Qualtatve results Fgure 6 shows some example CMO detectons usng the deformable parts-based model [9] as the detector. As quanttatvely demonstrated earler, we see that the CMO boundng-boxes contan a lot of unlabeled regons that are not labeled n the dataset, such as the sky regon for aeroplanes, road for bcycle, coffee table and wall pantngs for sofa, wndows for potted-plants, etc. Although our approach uses labeled data only from one object category to learn the contextual cues, we learn meanngful relatonshps among objects as well as object parts. For example, Fgure 7. The spatal maps ndcate the lkelhood of a person beng present gven the CMO detectons (red boxes). We see that the CMO provdes strong prmng for locatons of OOIs. the CMO detectons for bcycle n the 2nd column of Fgure 6 consstently nclude a person s body as parts of the CMO. In the 5th and 6th columns, we show detectons of two CMO components correspondng to the person category. We see that both components seem to capture two people, but whle the left-column models two people further apart from each other, the rght-column detects two people close together. We also see ntra-object CMO for cat n the last column. The false-postves shown n Fgure 6 clearly demonstrate that our learnt CMO models fre at contextually relevant regons n mages. Fnally, as seen n Fgure 4, these examples qualtatvely demonstrate that our cues can adaptvely learn contextual nformaton at dfferent granulartes n the scenes. Encouraged by the meanngful spatal nteractons observed n Fgure 6 (especally for the person category), we elaborate on that aspect a lttle further. As we demonstrated earler, our learnt CMO models often nclude objects from other labeled categores (e.g. the bcycle CMO often ncludes a person, a person CMO often ncludes another person, etc.). By examnng the CMO detectons on tranng mages, we can learn a dstrbuton of the locaton of each labeled category n the dataset relatve to our CMO models. These spatal dstrbutons can now be used as a pror to better gude the detecton of an OOI. We show a few examples of these spatal maps n Fgure 7. We also dsplay the HOG feature vsualzatons for the correspondng CMO. Whle n ths work we only leverage CMOs to provde co-occurrence based context, explorng the potental of our models to capture scene confguratons and provde explct spatal prmng for localzng objects s part of future work. 5. Concluson Regons that are not accounted for n the manually chosen lst of categores labeled n a dataset are often neglected by exstng works on contextual reasonng. In ths work, we explot these unlabeled regons to extract adaptve contextual cues for enhanced object detecton. We utlze the labeled object boundng-box as an anchor to algn scenes and learn spatally consstent and vsually dentfable contextual regons. The granularty of these regons s adap-

8 Fgure 6. Examples of the detected contextual meta-objects (CMO). Each column shows the CMO detectons for a specfc category. From left to rght: aeroplane, bcycle, sofa, potted-plant, person, person, cat. Each mage shows the hghest scorng CMO detecton. Red box ndcates the CMO boundng-box, whle the whte boxes represent the regon-flters wthn the CMO as learnt by the deformable parts-based model [9]. The frst four rows show true-postve detectons, and the last row shows false-postve detectons. Please refer to the authors webpages for more results. tvely and automatcally tuned for dfferent categores, and capture scene-level, nter-object as well as ntra-object nteractons. We cast the problem of learnng our proposed contextual cue nto that of learnng object models. Ths allows us to utlze any off-the-shelf object detector to learn our proposed contextual meta-objects. We present convncng quanttatve and qualtatve results on the challengng PASCAL VOC 2007 dataset, where we mprove on the performance of several object detectors, and compare favorably to exstng sources of context. The benefts of the adaptve granularty at whch we extract context, and the potental of our cue to provde complementary nformaton n addton to exstng cues are also demonstrated. These mprovements do not rely on advanced modelng technques or learnng algorthms, and ntellgently leverage exstng technology, makng them wdely accessble. [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] Acknowledgements: Ths research was supported n part by the Natonal Scence Foundaton under IIS [20] References [22] [1] MSRC 21-class Dataset. [2] M. Blaschko and C. Lampert. Object localzaton wth global and local context kernels. In BMVC, [3] M. Cho, J. Lm, A. Torralba, and A. Wllsky. Explotng herarchcal context on a large database of object categores. In CVPR, [4] N. Dalal and B. Trggs. Hstograms of orented gradents for human detecton. In CVPR, [5] C. Desa, D. Ramanan, and C. Fowlkes. Dscrmnatve models for mult-class object layout. In ICCV, [6] S. Dvvala, D. Hoem, J. Hays, A. Efros, and M. Hebert. An emprcal study of context n object detecton. In CVPR, [7] M. Everngham, L. Gool, C. Wllams, J. Wnn, and A. Zsserman. The Pascal Vsual Object Classes Challenge. [8] P. Felzenszwalb, R. Grshck, and D. McAllester. Dscrmnatvely traned [21] [23] [24] [25] [26] [27] [28] [29] [30] deformable part models, release 4. pff/latentrelease4/. P. Felzenszwalb, R. Grshck, D. McAllester, and D. Ramanan. Object detecton wth dscrmnatvely traned part-based models. PAMI, IEEE Transactons on, 32(9): , sep C. Gallegullos, B. McFee, S. Belonge, and G. Lanckret. Mult-class object localzaton by combnng local contextual nteractons. In CVPR, C. Gallegullos, A. Rabnovch, and S. Belonge. Object categorzaton usng co-occurrence, locaton and appearance. In CVPR, Anchorage, AK, G. Hetz and D. Koller. Learnng spatal context: Usng stuff to fnd thngs. In ECCV, D. Hoem, A. Efros, and M. Hebert. Puttng objects n perspectve. In CVPR, S. Kumar and M. Hebert. A herarchcal feld framework for unfed contextbased classfcaton. In ICCV, Y. Lee and K. Grauman. Object-graphs for context-aware category dscovery. In CVPR, B. Lebe, A. Leonards, and B. Schele. Robust object detecton wth nterleaved categorzaton and segmentaton. Int. J. Comput. Vson, 77: , May J. Lm, P. Arbel andez, C. Gu, and J. Malk. Context by regon ancestry. In ICCV, K. Murphy, A. Torralba, and W. Freeman. Usng the forest to see the trees: A graphcal model relatng features, objects, and scenes. In NIPS, D. Parkh, C. Ztnck, and T. Chen. From appearance to context-based recognton: Dense labelng n small mages. In CVPR, D. Parkh, C. Ztnck, and T. Chen. Unsupervsed learnng of herarchcal spatal structures n mages. In CVPR, D. Park, D. Ramanan, and C. Fowlkes. Multresoluton models for object detecton. In ECCV, A. Rabnovch, A. Vedald, C. Gallegullos, E. Wewora, and S. Belonge. Objects n context. In ICCV, N. Raswasa and N. Vasconcelos. Holstc context modelng usng semantc co-occurrences. In CVPR, M. Sadegh and A. Farhad. Recognton usng vsual phrases. In CVPR, E. Sudderth, A. Torralba, W. Freeman, and A. Wllsky. Learnng herarchcal models of scenes, objects, and parts. In ICCV, A. Torralba. Contextual prmng for object detecton. Int. J. Comput. Vson, 53(2): , A. Torralba, A. Olva, M. Castelhano, and J. Henderson. Contextual gudance of eye movements and attenton n real-world scenes: the role of global features n object search. Psychol Rev, 113(4), A. Vedald, V. Gulshan, M. Varma, and A. Zsserman. Multple kernels for object detecton. In ICCV, L. Wolf and S. Blesch. A crtcal vew of context. Int. J. Comput. Vson, 69: , August B. Yao and L. Fe-Fe. Modelng mutual context of object and human pose n human-object nteracton actvtes. In CVPR, 2010.

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1

Outline. Discriminative classifiers for image recognition. Where in the World? A nearest neighbor recognition example 4/14/2011. CS 376 Lecture 22 1 4/14/011 Outlne Dscrmnatve classfers for mage recognton Wednesday, Aprl 13 Krsten Grauman UT-Austn Last tme: wndow-based generc obect detecton basc ppelne face detecton wth boostng as case study Today:

More information

Lecture 5: Multilayer Perceptrons

Lecture 5: Multilayer Perceptrons Lecture 5: Multlayer Perceptrons Roger Grosse 1 Introducton So far, we ve only talked about lnear models: lnear regresson and lnear bnary classfers. We noted that there are functons that can t be represented

More information

TN348: Openlab Module - Colocalization

TN348: Openlab Module - Colocalization TN348: Openlab Module - Colocalzaton Topc The Colocalzaton module provdes the faclty to vsualze and quantfy colocalzaton between pars of mages. The Colocalzaton wndow contans a prevew of the two mages

More information

Learning the Kernel Parameters in Kernel Minimum Distance Classifier

Learning the Kernel Parameters in Kernel Minimum Distance Classifier Learnng the Kernel Parameters n Kernel Mnmum Dstance Classfer Daoqang Zhang 1,, Songcan Chen and Zh-Hua Zhou 1* 1 Natonal Laboratory for Novel Software Technology Nanjng Unversty, Nanjng 193, Chna Department

More information

Face Detection with Deep Learning

Face Detection with Deep Learning Face Detecton wth Deep Learnng Yu Shen Yus122@ucsd.edu A13227146 Kuan-We Chen kuc010@ucsd.edu A99045121 Yzhou Hao y3hao@ucsd.edu A98017773 Mn Hsuan Wu mhwu@ucsd.edu A92424998 Abstract The project here

More information

Combined Object Detection and Segmentation

Combined Object Detection and Segmentation Combned Object Detecton and Segmentaton Jarch Vansteenberge, Masayuk Mukunok, and Mchhko Mnoh Abstract We develop a method for combned object detecton and segmentaton n natural scene. In our approach segmentaton

More information

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching

A Fast Visual Tracking Algorithm Based on Circle Pixels Matching A Fast Vsual Trackng Algorthm Based on Crcle Pxels Matchng Zhqang Hou hou_zhq@sohu.com Chongzhao Han czhan@mal.xjtu.edu.cn Ln Zheng Abstract: A fast vsual trackng algorthm based on crcle pxels matchng

More information

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur

FEATURE EXTRACTION. Dr. K.Vijayarekha. Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur FEATURE EXTRACTION Dr. K.Vjayarekha Assocate Dean School of Electrcal and Electroncs Engneerng SASTRA Unversty, Thanjavur613 41 Jont Intatve of IITs and IISc Funded by MHRD Page 1 of 8 Table of Contents

More information

A Binarization Algorithm specialized on Document Images and Photos

A Binarization Algorithm specialized on Document Images and Photos A Bnarzaton Algorthm specalzed on Document mages and Photos Ergna Kavalleratou Dept. of nformaton and Communcaton Systems Engneerng Unversty of the Aegean kavalleratou@aegean.gr Abstract n ths paper, a

More information

Pictures at an Exhibition

Pictures at an Exhibition 1 Pctures at an Exhbton Stephane Kwan and Karen Zhu Department of Electrcal Engneerng Stanford Unversty, Stanford, CA 9405 Emal: {skwan1, kyzhu}@stanford.edu Abstract An mage processng algorthm s desgned

More information

CS 534: Computer Vision Model Fitting

CS 534: Computer Vision Model Fitting CS 534: Computer Vson Model Fttng Sprng 004 Ahmed Elgammal Dept of Computer Scence CS 534 Model Fttng - 1 Outlnes Model fttng s mportant Least-squares fttng Maxmum lkelhood estmaton MAP estmaton Robust

More information

Optimizing Document Scoring for Query Retrieval

Optimizing Document Scoring for Query Retrieval Optmzng Document Scorng for Query Retreval Brent Ellwen baellwe@cs.stanford.edu Abstract The goal of ths project was to automate the process of tunng a document query engne. Specfcally, I used machne learnng

More information

2 ZHENG et al.: ASSOCIATING GROUPS OF PEOPLE (a) Ambgutes from person re dentfcaton n solaton (b) Assocatng groups of people may reduce ambgutes n mat

2 ZHENG et al.: ASSOCIATING GROUPS OF PEOPLE (a) Ambgutes from person re dentfcaton n solaton (b) Assocatng groups of people may reduce ambgutes n mat ZHENG et al.: ASSOCIATING GROUPS OF PEOPLE 1 Assocatng Groups of People We-Sh Zheng jason@dcs.qmul.ac.uk Shaogang Gong sgg@dcs.qmul.ac.uk Tao Xang txang@dcs.qmul.ac.uk School of EECS, Queen Mary Unversty

More information

Collaboratively Regularized Nearest Points for Set Based Recognition

Collaboratively Regularized Nearest Points for Set Based Recognition Academc Center for Computng and Meda Studes, Kyoto Unversty Collaboratvely Regularzed Nearest Ponts for Set Based Recognton Yang Wu, Mchhko Mnoh, Masayuk Mukunok Kyoto Unversty 9/1/013 BMVC 013 @ Brstol,

More information

Reducing Frame Rate for Object Tracking

Reducing Frame Rate for Object Tracking Reducng Frame Rate for Object Trackng Pavel Korshunov 1 and We Tsang Oo 2 1 Natonal Unversty of Sngapore, Sngapore 11977, pavelkor@comp.nus.edu.sg 2 Natonal Unversty of Sngapore, Sngapore 11977, oowt@comp.nus.edu.sg

More information

S1 Note. Basis functions.

S1 Note. Basis functions. S1 Note. Bass functons. Contents Types of bass functons...1 The Fourer bass...2 B-splne bass...3 Power and type I error rates wth dfferent numbers of bass functons...4 Table S1. Smulaton results of type

More information

Online Detection and Classification of Moving Objects Using Progressively Improving Detectors

Online Detection and Classification of Moving Objects Using Progressively Improving Detectors Onlne Detecton and Classfcaton of Movng Objects Usng Progressvely Improvng Detectors Omar Javed Saad Al Mubarak Shah Computer Vson Lab School of Computer Scence Unversty of Central Florda Orlando, FL 32816

More information

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task

Term Weighting Classification System Using the Chi-square Statistic for the Classification Subtask at NTCIR-6 Patent Retrieval Task Proceedngs of NTCIR-6 Workshop Meetng, May 15-18, 2007, Tokyo, Japan Term Weghtng Classfcaton System Usng the Ch-square Statstc for the Classfcaton Subtask at NTCIR-6 Patent Retreval Task Kotaro Hashmoto

More information

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration

Improvement of Spatial Resolution Using BlockMatching Based Motion Estimation and Frame. Integration Improvement of Spatal Resoluton Usng BlockMatchng Based Moton Estmaton and Frame Integraton Danya Suga and Takayuk Hamamoto Graduate School of Engneerng, Tokyo Unversty of Scence, 6-3-1, Nuku, Katsuska-ku,

More information

Object-Based Techniques for Image Retrieval

Object-Based Techniques for Image Retrieval 54 Zhang, Gao, & Luo Chapter VII Object-Based Technques for Image Retreval Y. J. Zhang, Tsnghua Unversty, Chna Y. Y. Gao, Tsnghua Unversty, Chna Y. Luo, Tsnghua Unversty, Chna ABSTRACT To overcome the

More information

Combining Appearance Models and Markov Random Fields for Category Level Object Segmentation

Combining Appearance Models and Markov Random Fields for Category Level Object Segmentation Combnng Appearance Models and Markov Random Felds for Category Level Object Segmentaton Dane Larlus LEAR - INRIA, INPG Fre de rc Jure LEAR - INRIA, Caen Unversty dane.larlus@nralpes.fr frederc.jure@uncaen.fr

More information

User Authentication Based On Behavioral Mouse Dynamics Biometrics

User Authentication Based On Behavioral Mouse Dynamics Biometrics User Authentcaton Based On Behavoral Mouse Dynamcs Bometrcs Chee-Hyung Yoon Danel Donghyun Km Department of Computer Scence Department of Computer Scence Stanford Unversty Stanford Unversty Stanford, CA

More information

Efficient Inference for Fully-Connected CRFs with Stationarity

Efficient Inference for Fully-Connected CRFs with Stationarity Effcent Inference for Fully-onnected RFs wth Statonarty Ymeng Zhang Tsuhan hen School of Electrcal and omputer Engneerng, ornell Unversty {yz457,tsuhan}@cornell.edu Abstract The ondtonal Random Feld (RF)

More information

High Five: Recognising human interactions in TV shows

High Five: Recognising human interactions in TV shows PATRON-PEREZ ET AL.: RECOGNISING INTERACTIONS IN TV SHOWS 1 Hgh Fve: Recognsng human nteractons n TV shows Alonso Patron-Perez alonso@robots.ox.ac.uk Marcn Marszalek marcn@robots.ox.ac.uk Andrew Zsserman

More information

USING LINEAR REGRESSION FOR THE AUTOMATION OF SUPERVISED CLASSIFICATION IN MULTITEMPORAL IMAGES

USING LINEAR REGRESSION FOR THE AUTOMATION OF SUPERVISED CLASSIFICATION IN MULTITEMPORAL IMAGES USING LINEAR REGRESSION FOR THE AUTOMATION OF SUPERVISED CLASSIFICATION IN MULTITEMPORAL IMAGES 1 Fetosa, R.Q., 2 Merelles, M.S.P., 3 Blos, P. A. 1,3 Dept. of Electrcal Engneerng ; Catholc Unversty of

More information

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems

A Unified Framework for Semantics and Feature Based Relevance Feedback in Image Retrieval Systems A Unfed Framework for Semantcs and Feature Based Relevance Feedback n Image Retreval Systems Ye Lu *, Chunhu Hu 2, Xngquan Zhu 3*, HongJang Zhang 2, Qang Yang * School of Computng Scence Smon Fraser Unversty

More information

Parallelism for Nested Loops with Non-uniform and Flow Dependences

Parallelism for Nested Loops with Non-uniform and Flow Dependences Parallelsm for Nested Loops wth Non-unform and Flow Dependences Sam-Jn Jeong Dept. of Informaton & Communcaton Engneerng, Cheonan Unversty, 5, Anseo-dong, Cheonan, Chungnam, 330-80, Korea. seong@cheonan.ac.kr

More information

Categorizing objects: of appearance

Categorizing objects: of appearance Categorzng objects: global and part-based models of appearance UT Austn Generc categorzaton problem 1 Challenges: robustness Realstc scenes are crowded, cluttered, have overlappng objects. Generc category

More information

Query Clustering Using a Hybrid Query Similarity Measure

Query Clustering Using a Hybrid Query Similarity Measure Query clusterng usng a hybrd query smlarty measure Fu. L., Goh, D.H., & Foo, S. (2004). WSEAS Transacton on Computers, 3(3), 700-705. Query Clusterng Usng a Hybrd Query Smlarty Measure Ln Fu, Don Hoe-Lan

More information

A Background Subtraction for a Vision-based User Interface *

A Background Subtraction for a Vision-based User Interface * A Background Subtracton for a Vson-based User Interface * Dongpyo Hong and Woontack Woo KJIST U-VR Lab. {dhon wwoo}@kjst.ac.kr Abstract In ths paper, we propose a robust and effcent background subtracton

More information

Feature Reduction and Selection

Feature Reduction and Selection Feature Reducton and Selecton Dr. Shuang LIANG School of Software Engneerng TongJ Unversty Fall, 2012 Today s Topcs Introducton Problems of Dmensonalty Feature Reducton Statstc methods Prncpal Components

More information

3D vector computer graphics

3D vector computer graphics 3D vector computer graphcs Paolo Varagnolo: freelance engneer Padova Aprl 2016 Prvate Practce ----------------------------------- 1. Introducton Vector 3D model representaton n computer graphcs requres

More information

An Image Fusion Approach Based on Segmentation Region

An Image Fusion Approach Based on Segmentation Region Rong Wang, L-Qun Gao, Shu Yang, Yu-Hua Cha, and Yan-Chun Lu An Image Fuson Approach Based On Segmentaton Regon An Image Fuson Approach Based on Segmentaton Regon Rong Wang, L-Qun Gao, Shu Yang 3, Yu-Hua

More information

Efficient Segmentation and Classification of Remote Sensing Image Using Local Self Similarity

Efficient Segmentation and Classification of Remote Sensing Image Using Local Self Similarity ISSN(Onlne): 2320-9801 ISSN (Prnt): 2320-9798 Internatonal Journal of Innovatve Research n Computer and Communcaton Engneerng (An ISO 3297: 2007 Certfed Organzaton) Vol.2, Specal Issue 1, March 2014 Proceedngs

More information

Corner-Based Image Alignment using Pyramid Structure with Gradient Vector Similarity

Corner-Based Image Alignment using Pyramid Structure with Gradient Vector Similarity Journal of Sgnal and Informaton Processng, 013, 4, 114-119 do:10.436/jsp.013.43b00 Publshed Onlne August 013 (http://www.scrp.org/journal/jsp) Corner-Based Image Algnment usng Pyramd Structure wth Gradent

More information

Margin-Constrained Multiple Kernel Learning Based Multi-Modal Fusion for Affect Recognition

Margin-Constrained Multiple Kernel Learning Based Multi-Modal Fusion for Affect Recognition Margn-Constraned Multple Kernel Learnng Based Mult-Modal Fuson for Affect Recognton Shzh Chen and Yngl Tan Electrcal Engneerng epartment The Cty College of New Yor New Yor, NY USA {schen, ytan}@ccny.cuny.edu

More information

Fast Feature Value Searching for Face Detection

Fast Feature Value Searching for Face Detection Vol., No. 2 Computer and Informaton Scence Fast Feature Value Searchng for Face Detecton Yunyang Yan Department of Computer Engneerng Huayn Insttute of Technology Hua an 22300, Chna E-mal: areyyyke@63.com

More information

Robust Inlier Feature Tracking Method for Multiple Pedestrian Tracking

Robust Inlier Feature Tracking Method for Multiple Pedestrian Tracking 2011 Internatonal Conference on Informaton and Intellgent Computng IPCSIT vol.18 (2011) (2011) IACSIT Press, Sngapore Robust Inler Feature Trackng Method for Multple Pedestran Trackng Young-Chul Lm a*

More information

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields

A mathematical programming approach to the analysis, design and scheduling of offshore oilfields 17 th European Symposum on Computer Aded Process Engneerng ESCAPE17 V. Plesu and P.S. Agach (Edtors) 2007 Elsever B.V. All rghts reserved. 1 A mathematcal programmng approach to the analyss, desgn and

More information

A Novel Adaptive Descriptor Algorithm for Ternary Pattern Textures

A Novel Adaptive Descriptor Algorithm for Ternary Pattern Textures A Novel Adaptve Descrptor Algorthm for Ternary Pattern Textures Fahuan Hu 1,2, Guopng Lu 1 *, Zengwen Dong 1 1.School of Mechancal & Electrcal Engneerng, Nanchang Unversty, Nanchang, 330031, Chna; 2. School

More information

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz

Compiler Design. Spring Register Allocation. Sample Exercises and Solutions. Prof. Pedro C. Diniz Compler Desgn Sprng 2014 Regster Allocaton Sample Exercses and Solutons Prof. Pedro C. Dnz USC / Informaton Scences Insttute 4676 Admralty Way, Sute 1001 Marna del Rey, Calforna 90292 pedro@s.edu Regster

More information

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour

6.854 Advanced Algorithms Petar Maymounkov Problem Set 11 (November 23, 2005) With: Benjamin Rossman, Oren Weimann, and Pouya Kheradpour 6.854 Advanced Algorthms Petar Maymounkov Problem Set 11 (November 23, 2005) Wth: Benjamn Rossman, Oren Wemann, and Pouya Kheradpour Problem 1. We reduce vertex cover to MAX-SAT wth weghts, such that the

More information

Discriminative classifiers for object classification. Last time

Discriminative classifiers for object classification. Last time Dscrmnatve classfers for object classfcaton Thursday, Nov 12 Krsten Grauman UT Austn Last tme Supervsed classfcaton Loss and rsk, kbayes rule Skn color detecton example Sldng ndo detecton Classfers, boostng

More information

Image Alignment CSC 767

Image Alignment CSC 767 Image Algnment CSC 767 Image algnment Image from http://graphcs.cs.cmu.edu/courses/15-463/2010_fall/ Image algnment: Applcatons Panorama sttchng Image algnment: Applcatons Recognton of object nstances

More information

Factor Graphs for Region-based Whole-scene Classification

Factor Graphs for Region-based Whole-scene Classification Factor Graphs for Regon-based Whole-scene Classfcaton Matthew R. Boutell Jebo Luo Chrstopher M. Brown CSSE Dept. Res. and Dev. Labs Dept. of Computer Scence Rose-Hulman Inst. of Techn. Eastman Kodak Company

More information

Skew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach

Skew Angle Estimation and Correction of Hand Written, Textual and Large areas of Non-Textual Document Images: A Novel Approach Angle Estmaton and Correcton of Hand Wrtten, Textual and Large areas of Non-Textual Document Images: A Novel Approach D.R.Ramesh Babu Pyush M Kumat Mahesh D Dhannawat PES Insttute of Technology Research

More information

Histogram of Template for Pedestrian Detection

Histogram of Template for Pedestrian Detection PAPER IEICE TRANS. FUNDAMENTALS/COMMUN./ELECTRON./INF. & SYST., VOL. E85-A/B/C/D, No. xx JANUARY 20xx Hstogram of Template for Pedestran Detecton Shaopeng Tang, Non Member, Satosh Goto Fellow Summary In

More information

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance

Tsinghua University at TAC 2009: Summarizing Multi-documents by Information Distance Tsnghua Unversty at TAC 2009: Summarzng Mult-documents by Informaton Dstance Chong Long, Mnle Huang, Xaoyan Zhu State Key Laboratory of Intellgent Technology and Systems, Tsnghua Natonal Laboratory for

More information

A Gradient Difference based Technique for Video Text Detection

A Gradient Difference based Technique for Video Text Detection A Gradent Dfference based Technque for Vdeo Text Detecton Palaahnakote Shvakumara, Trung Quy Phan and Chew Lm Tan School of Computng, Natonal Unversty of Sngapore {shva, phanquyt, tancl }@comp.nus.edu.sg

More information

Hybrid Neuro-Bayesian Spatial Contextual Reasoning for Scene Content Understanding

Hybrid Neuro-Bayesian Spatial Contextual Reasoning for Scene Content Understanding 12th Internatonal Conference on Informaton Fuson Seattle, WA, USA, July 6-9, 2009 Hybrd Neuro-Bayesan Spatal Contextual Reasonng for Scene Content Understandng Dens Garagc *, Majd Zandpour *, Frank Stolle

More information

EYE CENTER LOCALIZATION ON A FACIAL IMAGE BASED ON MULTI-BLOCK LOCAL BINARY PATTERNS

EYE CENTER LOCALIZATION ON A FACIAL IMAGE BASED ON MULTI-BLOCK LOCAL BINARY PATTERNS P.G. Demdov Yaroslavl State Unversty Anatoly Ntn, Vladmr Khryashchev, Olga Stepanova, Igor Kostern EYE CENTER LOCALIZATION ON A FACIAL IMAGE BASED ON MULTI-BLOCK LOCAL BINARY PATTERNS Yaroslavl, 2015 Eye

More information

Local Quaternary Patterns and Feature Local Quaternary Patterns

Local Quaternary Patterns and Feature Local Quaternary Patterns Local Quaternary Patterns and Feature Local Quaternary Patterns Jayu Gu and Chengjun Lu The Department of Computer Scence, New Jersey Insttute of Technology, Newark, NJ 0102, USA Abstract - Ths paper presents

More information

Detection of an Object by using Principal Component Analysis

Detection of an Object by using Principal Component Analysis Detecton of an Object by usng Prncpal Component Analyss 1. G. Nagaven, 2. Dr. T. Sreenvasulu Reddy 1. M.Tech, Department of EEE, SVUCE, Trupath, Inda. 2. Assoc. Professor, Department of ECE, SVUCE, Trupath,

More information

A Gradient Difference based Technique for Video Text Detection

A Gradient Difference based Technique for Video Text Detection 2009 10th Internatonal Conference on Document Analyss and Recognton A Gradent Dfference based Technque for Vdeo Text Detecton Palaahnakote Shvakumara, Trung Quy Phan and Chew Lm Tan School of Computng,

More information

Semantic Image Retrieval Using Region Based Inverted File

Semantic Image Retrieval Using Region Based Inverted File Semantc Image Retreval Usng Regon Based Inverted Fle Dengsheng Zhang, Md Monrul Islam, Guoun Lu and Jn Hou 2 Gppsland School of Informaton Technology, Monash Unversty Churchll, VIC 3842, Australa E-mal:

More information

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION

MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION MULTISPECTRAL IMAGES CLASSIFICATION BASED ON KLT AND ATR AUTOMATIC TARGET RECOGNITION Paulo Quntlano 1 & Antono Santa-Rosa 1 Federal Polce Department, Brasla, Brazl. E-mals: quntlano.pqs@dpf.gov.br and

More information

Machine Learning 9. week

Machine Learning 9. week Machne Learnng 9. week Mappng Concept Radal Bass Functons (RBF) RBF Networks 1 Mappng It s probably the best scenaro for the classfcaton of two dataset s to separate them lnearly. As you see n the below

More information

Alignment Results of SOBOM for OAEI 2010

Alignment Results of SOBOM for OAEI 2010 Algnment Results of SOBOM for OAEI 2010 Pegang Xu, Yadong Wang, Lang Cheng, Tany Zang School of Computer Scence and Technology Harbn Insttute of Technology, Harbn, Chna pegang.xu@gmal.com, ydwang@ht.edu.cn,

More information

Joint Object and Part Segmentation using Deep Learned Potentials

Joint Object and Part Segmentation using Deep Learned Potentials Jont Object and Part Segmentaton usng Deep Learned Potentals Peng Wang 1 Xaohu Shen 2 Zhe Ln 2 Scott Cohen 2 Bran Prce 2 Alan Yulle 1 1 Unversty of Calforna, Los Angeles 2 Adobe Research Abstract Segmentng

More information

An Entropy-Based Approach to Integrated Information Needs Assessment

An Entropy-Based Approach to Integrated Information Needs Assessment Dstrbuton Statement A: Approved for publc release; dstrbuton s unlmted. An Entropy-Based Approach to ntegrated nformaton Needs Assessment June 8, 2004 Wllam J. Farrell Lockheed Martn Advanced Technology

More information

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1)

For instance, ; the five basic number-sets are increasingly more n A B & B A A = B (1) Secton 1.2 Subsets and the Boolean operatons on sets If every element of the set A s an element of the set B, we say that A s a subset of B, or that A s contaned n B, or that B contans A, and we wrte A

More information

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data

A Fast Content-Based Multimedia Retrieval Technique Using Compressed Data A Fast Content-Based Multmeda Retreval Technque Usng Compressed Data Borko Furht and Pornvt Saksobhavvat NSF Multmeda Laboratory Florda Atlantc Unversty, Boca Raton, Florda 3343 ABSTRACT In ths paper,

More information

CMPSCI 670: Computer Vision! Object detection continued. University of Massachusetts, Amherst November 10, 2014 Instructor: Subhransu Maji

CMPSCI 670: Computer Vision! Object detection continued. University of Massachusetts, Amherst November 10, 2014 Instructor: Subhransu Maji CMPSCI 670: Computer Vson! Object detecton contnued Unversty of Massachusetts, Amherst November 10, 2014 Instructor: Subhransu Maj No class on Wednesday Admnstrva Followng Tuesday s schedule ths Wednesday

More information

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification

12/2/2009. Announcements. Parametric / Non-parametric. Case-Based Reasoning. Nearest-Neighbor on Images. Nearest-Neighbor Classification Introducton to Artfcal Intellgence V22.0472-001 Fall 2009 Lecture 24: Nearest-Neghbors & Support Vector Machnes Rob Fergus Dept of Computer Scence, Courant Insttute, NYU Sldes from Danel Yeung, John DeNero

More information

Weakly supervised graph based semantic segmentation by learning communities of image-parts

Weakly supervised graph based semantic segmentation by learning communities of image-parts Weakly supervsed graph based semantc segmentaton by learnng communtes of mage-parts Nloufar Pouran, S. Karthkeyan, and B.S. Manjunath Department of Electrcal and Computer Engneerng, Unversty of Calforna,

More information

Large-scale Web Video Event Classification by use of Fisher Vectors

Large-scale Web Video Event Classification by use of Fisher Vectors Large-scale Web Vdeo Event Classfcaton by use of Fsher Vectors Chen Sun and Ram Nevata Unversty of Southern Calforna, Insttute for Robotcs and Intellgent Systems Los Angeles, CA 90089, USA {chensun nevata}@usc.org

More information

Manifold-Ranking Based Keyword Propagation for Image Retrieval *

Manifold-Ranking Based Keyword Propagation for Image Retrieval * Manfold-Rankng Based Keyword Propagaton for Image Retreval * Hanghang Tong,, Jngru He,, Mngjng L 2, We-Yng Ma 2, Hong-Jang Zhang 2 and Changshu Zhang 3,3 Department of Automaton, Tsnghua Unversty, Bejng

More information

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009.

Assignment # 2. Farrukh Jabeen Algorithms 510 Assignment #2 Due Date: June 15, 2009. Farrukh Jabeen Algorthms 51 Assgnment #2 Due Date: June 15, 29. Assgnment # 2 Chapter 3 Dscrete Fourer Transforms Implement the FFT for the DFT. Descrbed n sectons 3.1 and 3.2. Delverables: 1. Concse descrpton

More information

Edge Detection in Noisy Images Using the Support Vector Machines

Edge Detection in Noisy Images Using the Support Vector Machines Edge Detecton n Nosy Images Usng the Support Vector Machnes Hlaro Gómez-Moreno, Saturnno Maldonado-Bascón, Francsco López-Ferreras Sgnal Theory and Communcatons Department. Unversty of Alcalá Crta. Madrd-Barcelona

More information

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 25, NO. 4, APRIL

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 25, NO. 4, APRIL IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 25, NO. 4, APRIL 2016 1713 Weakly Supervsed Fne-Graned Categorzaton Wth Part-Based Image Representaton Yu Zhang, Xu-Shen We, Janxn Wu, Member, IEEE, Janfe Ca,

More information

Heterogeneous Visual Features Fusion via Sparse Multimodal Machine

Heterogeneous Visual Features Fusion via Sparse Multimodal Machine 013 IEEE Conference on Computer Vson and Pattern Recognton Heterogeneous Vsual Features Fuson va Sparse Multmodal Machne Hua ang, Fepng Ne, Heng Huang, Chrs Dng Department of Electrcal Engneerng and Computer

More information

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET

BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET 1 BOOSTING CLASSIFICATION ACCURACY WITH SAMPLES CHOSEN FROM A VALIDATION SET TZU-CHENG CHUANG School of Electrcal and Computer Engneerng, Purdue Unversty, West Lafayette, Indana 47907 SAUL B. GELFAND School

More information

Discrete-Continuous Depth Estimation from a Single Image

Discrete-Continuous Depth Estimation from a Single Image Dscrete-Contnuous Depth Estmaton from a Sngle Image Maomao Lu, Matheu Salzmann, Xumng He NICTA and CECS, ANU, Canberra {maomao.lu, matheu.salzmann, xumng.he}@ncta.com.au Abstract In ths paper, we tackle

More information

Classifier Selection Based on Data Complexity Measures *

Classifier Selection Based on Data Complexity Measures * Classfer Selecton Based on Data Complexty Measures * Edth Hernández-Reyes, J.A. Carrasco-Ochoa, and J.Fco. Martínez-Trndad Natonal Insttute for Astrophyscs, Optcs and Electroncs, Lus Enrque Erro No.1 Sta.

More information

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision

SLAM Summer School 2006 Practical 2: SLAM using Monocular Vision SLAM Summer School 2006 Practcal 2: SLAM usng Monocular Vson Javer Cvera, Unversty of Zaragoza Andrew J. Davson, Imperal College London J.M.M Montel, Unversty of Zaragoza. josemar@unzar.es, jcvera@unzar.es,

More information

arxiv: v2 [cs.cv] 9 Apr 2018

arxiv: v2 [cs.cv] 9 Apr 2018 Boundary-senstve Network for Portrat Segmentaton Xanzh Du 1, Xaolong Wang 2, Dawe L 2, Jngwen Zhu 2, Serafettn Tasc 2, Cameron Uprght 2, Stephen Walsh 2, Larry Davs 1 1 Computer Vson Lab, UMIACS, Unversty

More information

Support Vector Machines

Support Vector Machines /9/207 MIST.6060 Busness Intellgence and Data Mnng What are Support Vector Machnes? Support Vector Machnes Support Vector Machnes (SVMs) are supervsed learnng technques that analyze data and recognze patterns.

More information

Classifier Swarms for Human Detection in Infrared Imagery

Classifier Swarms for Human Detection in Infrared Imagery Classfer Swarms for Human Detecton n Infrared Imagery Yur Owechko, Swarup Medasan, and Narayan Srnvasa HRL Laboratores, LLC 3011 Malbu Canyon Road, Malbu, CA 90265 {owechko, smedasan, nsrnvasa}@hrl.com

More information

Fitting & Matching. Lecture 4 Prof. Bregler. Slides from: S. Lazebnik, S. Seitz, M. Pollefeys, A. Effros.

Fitting & Matching. Lecture 4 Prof. Bregler. Slides from: S. Lazebnik, S. Seitz, M. Pollefeys, A. Effros. Fttng & Matchng Lecture 4 Prof. Bregler Sldes from: S. Lazebnk, S. Setz, M. Pollefeys, A. Effros. How do we buld panorama? We need to match (algn) mages Matchng wth Features Detect feature ponts n both

More information

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law)

Machine Learning. Support Vector Machines. (contains material adapted from talks by Constantin F. Aliferis & Ioannis Tsamardinos, and Martin Law) Machne Learnng Support Vector Machnes (contans materal adapted from talks by Constantn F. Alfers & Ioanns Tsamardnos, and Martn Law) Bryan Pardo, Machne Learnng: EECS 349 Fall 2014 Support Vector Machnes

More information

Steps for Computing the Dissimilarity, Entropy, Herfindahl-Hirschman and. Accessibility (Gravity with Competition) Indices

Steps for Computing the Dissimilarity, Entropy, Herfindahl-Hirschman and. Accessibility (Gravity with Competition) Indices Steps for Computng the Dssmlarty, Entropy, Herfndahl-Hrschman and Accessblty (Gravty wth Competton) Indces I. Dssmlarty Index Measurement: The followng formula can be used to measure the evenness between

More information

Deep Spatial-Temporal Joint Feature Representation for Video Object Detection

Deep Spatial-Temporal Joint Feature Representation for Video Object Detection sensors Artcle Deep Spatal-Temporal Jont Feature Representaton for Vdeo Object Detecton Baojun Zhao 1,2, Boya Zhao 1,2 ID, Lnbo Tang 1,2, *, Yuq Han 1,2 and Wenzheng Wang 1,2 1 School of Informaton and

More information

An Evaluation of Divide-and-Combine Strategies for Image Categorization by Multi-Class Support Vector Machines

An Evaluation of Divide-and-Combine Strategies for Image Categorization by Multi-Class Support Vector Machines An Evaluaton of Dvde-and-Combne Strateges for Image Categorzaton by Mult-Class Support Vector Machnes C. Demrkesen¹ and H. Cherf¹, ² 1: Insttue of Scence and Engneerng 2: Faculté des Scences Mrande Galatasaray

More information

Complex Numbers. Now we also saw that if a and b were both positive then ab = a b. For a second let s forget that restriction and do the following.

Complex Numbers. Now we also saw that if a and b were both positive then ab = a b. For a second let s forget that restriction and do the following. Complex Numbers The last topc n ths secton s not really related to most of what we ve done n ths chapter, although t s somewhat related to the radcals secton as we wll see. We also won t need the materal

More information

Improving Web Image Search using Meta Re-rankers

Improving Web Image Search using Meta Re-rankers VOLUME-1, ISSUE-V (Aug-Sep 2013) IS NOW AVAILABLE AT: www.dcst.com Improvng Web Image Search usng Meta Re-rankers B.Kavtha 1, N. Suata 2 1 Department of Computer Scence and Engneerng, Chtanya Bharath Insttute

More information

Smoothing Spline ANOVA for variable screening

Smoothing Spline ANOVA for variable screening Smoothng Splne ANOVA for varable screenng a useful tool for metamodels tranng and mult-objectve optmzaton L. Rcco, E. Rgon, A. Turco Outlne RSM Introducton Possble couplng Test case MOO MOO wth Game Theory

More information

A PATTERN RECOGNITION APPROACH TO IMAGE SEGMENTATION

A PATTERN RECOGNITION APPROACH TO IMAGE SEGMENTATION 1 THE PUBLISHING HOUSE PROCEEDINGS OF THE ROMANIAN ACADEMY, Seres A, OF THE ROMANIAN ACADEMY Volume 4, Number 2/2003, pp.000-000 A PATTERN RECOGNITION APPROACH TO IMAGE SEGMENTATION Tudor BARBU Insttute

More information

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes

R s s f. m y s. SPH3UW Unit 7.3 Spherical Concave Mirrors Page 1 of 12. Notes SPH3UW Unt 7.3 Sphercal Concave Mrrors Page 1 of 1 Notes Physcs Tool box Concave Mrror If the reflectng surface takes place on the nner surface of the sphercal shape so that the centre of the mrror bulges

More information

The Research of Support Vector Machine in Agricultural Data Classification

The Research of Support Vector Machine in Agricultural Data Classification The Research of Support Vector Machne n Agrcultural Data Classfcaton Le Sh, Qguo Duan, Xnmng Ma, Me Weng College of Informaton and Management Scence, HeNan Agrcultural Unversty, Zhengzhou 45000 Chna Zhengzhou

More information

Performance Evaluation of Information Retrieval Systems

Performance Evaluation of Information Retrieval Systems Why System Evaluaton? Performance Evaluaton of Informaton Retreval Systems Many sldes n ths secton are adapted from Prof. Joydeep Ghosh (UT ECE) who n turn adapted them from Prof. Dk Lee (Unv. of Scence

More information

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points;

Subspace clustering. Clustering. Fundamental to all clustering techniques is the choice of distance measure between data points; Subspace clusterng Clusterng Fundamental to all clusterng technques s the choce of dstance measure between data ponts; D q ( ) ( ) 2 x x = x x, j k = 1 k jk Squared Eucldean dstance Assumpton: All features

More information

Color Image Segmentation Using Multispectral Random Field Texture Model & Color Content Features

Color Image Segmentation Using Multispectral Random Field Texture Model & Color Content Features Color Image Segmentaton Usng Multspectral Random Feld Texture Model & Color Content Features Orlando J. Hernandez E-mal: hernande@tcnj.edu Department Electrcal & Computer Engneerng, The College of New

More information

Recognizing Faces. Outline

Recognizing Faces. Outline Recognzng Faces Drk Colbry Outlne Introducton and Motvaton Defnng a feature vector Prncpal Component Analyss Lnear Dscrmnate Analyss !"" #$""% http://www.nfotech.oulu.f/annual/2004 + &'()*) '+)* 2 ! &

More information

An Optimal Algorithm for Prufer Codes *

An Optimal Algorithm for Prufer Codes * J. Software Engneerng & Applcatons, 2009, 2: 111-115 do:10.4236/jsea.2009.22016 Publshed Onlne July 2009 (www.scrp.org/journal/jsea) An Optmal Algorthm for Prufer Codes * Xaodong Wang 1, 2, Le Wang 3,

More information

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning

Outline. Type of Machine Learning. Examples of Application. Unsupervised Learning Outlne Artfcal Intellgence and ts applcatons Lecture 8 Unsupervsed Learnng Professor Danel Yeung danyeung@eee.org Dr. Patrck Chan patrckchan@eee.org South Chna Unversty of Technology, Chna Introducton

More information

What is Object Detection? Face Detection using AdaBoost. Detection as Classification. Principle of Boosting (Schapire 90)

What is Object Detection? Face Detection using AdaBoost. Detection as Classification. Principle of Boosting (Schapire 90) CIS 5543 Coputer Vson Object Detecton What s Object Detecton? Locate an object n an nput age Habn Lng Extensons Vola & Jones, 2004 Dalal & Trggs, 2005 one or ultple objects Object segentaton Object detecton

More information

Adaptive Transfer Learning

Adaptive Transfer Learning Adaptve Transfer Learnng Bn Cao, Snno Jaln Pan, Yu Zhang, Dt-Yan Yeung, Qang Yang Hong Kong Unversty of Scence and Technology Clear Water Bay, Kowloon, Hong Kong {caobn,snnopan,zhangyu,dyyeung,qyang}@cse.ust.hk

More information

CROWD Analysis is fundamental to solving many realword

CROWD Analysis is fundamental to solving many realword 1986 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 37, NO. 10, OCTOBER 2015 Detectng Humans n Dense Crowds Usng Locally-Consstent Scale Pror and Global Occluson Reasonng Haroon Idrees,

More information

Multi-View Face Alignment Using 3D Shape Model for View Estimation

Multi-View Face Alignment Using 3D Shape Model for View Estimation Mult-Vew Face Algnment Usng 3D Shape Model for Vew Estmaton Yanchao Su 1, Hazhou A 1, Shhong Lao 1 Computer Scence and Technology Department, Tsnghua Unversty Core Technology Center, Omron Corporaton ahz@mal.tsnghua.edu.cn

More information

Feature Kernel Functions: Improving SVMs Using High-level Knowledge

Feature Kernel Functions: Improving SVMs Using High-level Knowledge Feature Kernel Functons: Improvng SVMs Usng Hgh-level Knowledge Qang Sun, Gerald DeJong Department of Computer Scence, Unversty of Illnos at Urbana-Champagn qangsun@uuc.edu, dejong@cs.uuc.edu Abstract

More information