Multiple cosegmentation
|
|
- Wilfrid Small
- 5 years ago
- Views:
Transcription
1 Armand Joulin, Francis Bach and Jean Ponce. INRIA -Ecole Normale Supérieure April 25, 2012
2 Segmentation Introduction Segmentation Supervised and weakly-supervised segmentation Cosegmentation Segmentation is classical and fundamental vision problem.
3 Segmentation Introduction Segmentation Supervised and weakly-supervised segmentation Cosegmentation Segmentation is classical and fundamental vision problem. Problem: Many possible solutions.
4 Existing solutions Introduction Segmentation Supervised and weakly-supervised segmentation Cosegmentation Supervised Segmentation: Need ground truth for every class of object Cannot deal with an unknown object. P. Krähenbühl and V. Koltun (NIPS 11) Interactive segmentation (scribbles or a bounding box) Need human interaction for each image. GrabCut
5 Cosegmentation Introduction Segmentation Supervised and weakly-supervised segmentation Cosegmentation Dividing one images.
6 Cosegmentation Introduction Segmentation Supervised and weakly-supervised segmentation Cosegmentation Dividing a set of images by using shared information.
7 Cosegmentation Introduction Segmentation Supervised and weakly-supervised segmentation Cosegmentation Dividing a set of images by using shared information. No prior information. But: common foreground and different background.
8 Cosegmentation Introduction Segmentation Supervised and weakly-supervised segmentation Cosegmentation Previous existing methods (Rother et al. 2006, Singh and Hochbaum 2009,...) only work with 2 images and the exact same object. The first presented method works on multiple images and on an object class. The second one extends it to multiple images and multiple object classes.
9 Segmentation Supervised and weakly-supervised segmentation Cosegmentation...Cosegmentation is also a ill-posed problem In natural images, objects are link with their environement...the background is also common to all the images.
10 Segmentation Supervised and weakly-supervised segmentation Cosegmentation...Cosegmentation is also a ill-posed problem In natural images, objects are link with their environement...the background is also common to all the images. Solutions: Use user interaction on some images,
11 Segmentation Supervised and weakly-supervised segmentation Cosegmentation...Cosegmentation is also a ill-posed problem In natural images, objects are link with their environement...the background is also common to all the images. Solutions: Use user interaction on some images, Segment the background into meaningful regions.
12 The goals of our approach Goal of our approach Notations Our method should: Handle multiple images.
13 The goals of our approach Goal of our approach Notations Our method should: Handle multiple images. Works on any kind of object/stuff.
14 The goals of our approach Goal of our approach Notations Our method should: Handle multiple images. Works on any kind of object/stuff. Segments the background into meaningful regions.
15 The goals of our approach Goal of our approach Notations Our method should: Handle multiple images. Works on any kind of object/stuff. Segments the background into meaningful regions. Uses no prior information but can be easily extended to interactive cosegmentation.
16 Method goals Introduction Goal of our approach Notations Local consistency Maximizing spatial consistency within a particular image. Figure: Image space.
17 Method goals Introduction Goal of our approach Notations Local consistency Maximizing spatial consistency within a particular image. Separation of the classes Maximizing the separability of K classes between different images Our framework: Unsupervised discriminative clustering. Figure: Image space. Figure: Feature space.
18 Problem Notations Introduction Goal of our approach Notations Each image i is reduced to a subsampled grid of pixels. For the n-th pixel, we denote by: x n its d-dimensional feature vector. y n the K-vector such as y nk = 1 if the n-th pixel is in the k-class and 0 otherwise.
19 Figure: Image space. Normalized Cut (Shi and Malik, 2000): The similarty between two pixels is mesured by the rbf distance between their position p n and their color c n.
20 Figure: Image space. Normalized Cut (Shi and Malik, 2000): The similarty between two pixels is mesured by the rbf distance between their position p n and their color c n. For an image i, our similarity matrix is: W i nm = exp( λ p p n p m 2 2 λ c c n c m 2 ).
21 Figure: Image space. Normalized Cut (Shi and Malik, 2000): The Laplacian matrix is L = I D 1/2 WD 1/2 where D the diagonal matrix composed of the row sums of W
22 Figure: Image space. Normalized Cut (Shi and Malik, 2000): The Laplacian matrix is L = I D 1/2 WD 1/2 where D the diagonal matrix composed of the row sums of W We thus have the following in our cost function: E B (y) = µ N tr(yt Ly).
23 Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Figure: Feature space. Discriminative classifier: given the labels y, we solve the following problem: 1 E U (y) = min A IR K d, N b IR K N n=1 l(y n,aφ(x n )+b)+ λ 2K A 2 F, Notations φ a non-linear mapping of the feature, l is a cost function.
24 Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Figure: Feature space. Discriminative classifier: given the labels y, we solve the following problem: 1 E U (y) = min A IR K d, N b IR K N n=1 l(y n,aφ(x n )+b)+ λ 2K A 2 F, Notations φ a non-linear mapping of the feature, l is a cost function.
25 Mapping approximation Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Our discriminative clustering framework works with positive definite kernels
26 Mapping approximation Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Our discriminative clustering framework works with positive definite kernels We use the χ 2 kernel matrix K: ( D K nm = exp λ h d=1 (x nd x md ) 2 ), x nd +x md
27 Mapping approximation Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Our discriminative clustering framework works with positive definite kernels We use the χ 2 kernel matrix K: ( D K nm = exp λ h d=1 (x nd x md ) 2 ), x nd +x md Equivalent to apply a mapping φ from the feature space to a high-dimensional Hilbert space F, such that: K nm = φ(x n ), φ(x m )
28 Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Figure: Feature space. Discriminative classifier: given the labels y, we solve the following problem: 1 E U (y) = min A IR K d, N b IR K N n=1 l(y n,aφ(x n )+b)+ λ 2K A 2 F, Notations φ a non-linear mapping of the feature, l is a cost function.
29 Loss function Introduction Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation We choose the soft-max loss function because it is suited for multiclass and is related to probabilistic models: l(y n,aφ(x n )+b) = K k=1 ( exp(a T ) y nk log k φ(x n )+b k ) K, l=1 exp(at l φ(x n )+b l )
30 Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Find the set of labels y which leads to the best data separation into K classes: min y {0,1} N K, y1 K =1 N min E U(y,A,b) A IR K d,b IR K
31 Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Find the set of labels y which leads to the best data separation into K classes: min y {0,1} N K, y1 K =1 N min E U(y,A,b) A IR K d,b IR K Problem: Same label for all the pixels perfect separation
32 Cluster size balancing Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Two solutions: adding linear constraints on the number of elements per class Encourage the proportion of points per class to be uniform
33 Cluster size balancing Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Two solutions: adding linear constraints on the number of elements per class Encourage the proportion of points per class to be uniform We choose the second: No additional parameters and have a probabilistic interpretation. H(y) = i I K k=1 ( 1 N n N i y nk ) ( 1 log N n N i y nk where i is an image, and N i the number of pixels in i Note: In a weakly supervised setting (e.g., interactive segmentation), this term can be modify to take into account prior knowledge. ).
34 Overall problem Introduction Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Combining the unary and binary term with the class balancing term, we obtain the following problem: [ ] min y {0,1} N K, y1 K =1 N min E U (y,a,b) +E B (y) H(y). A IR d K, b IR K
35 Probabilistic interpretation Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation We introduce t n in {0,1} I indicating to which image n belongs and z n in {1,...,M} giving for each pixel n some observable information The label y is a latent variable of the observable information z given x (x y z t) inducing an explain away phenomenon: the label y n and the variable t n compete to explain the observable information z n.
36 Probabilistic interpretation Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation More precisely, we suppose a bilinear model: P(z nm = 1 t ni = 1, y nk = 1) = y nk G ik mt ni, where N m=1 Gik m = 1
37 Probabilistic interpretation Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation More precisely, we suppose a bilinear model: P(z nm = 1 t ni = 1, y nk = 1) = y nk G ik mt ni, where N m=1 Gik m = 1 and a exponential family model for Y = (y 1,...,y N ) given X = (x 1,...,x N ) with unary parameters (A,b) and binary parameters L.
38 Probabilistic interpretation Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Our cost function is the mean-field variational approximation of the following (regularized) negative conditional log-likelihood of Z = (z 1,...,z N ) given X and T = (t 1,...,t N ) for our model: min A IR d K,b IR K, G IR N K I, G T 1 N =1, G 0 1 N N n=1 log ( p(z n x n, t n ) ) + λ 2K A 2 2.
39 Probabilistic interpretation Formulation Mapping approximation Loss function Cluster size balancing Overall problem Probabilistic interpretation Our cost function is the mean-field variational approximation of the following (regularized) negative conditional log-likelihood of Z = (z 1,...,z N ) given X and T = (t 1,...,t N ) for our model: min A IR d K,b IR K, G IR N K I, G T 1 N =1, G 0 1 N N n=1 log ( p(z n x n, t n ) ) + λ 2K A 2 2. Z can encode must-link and must-not-link constraints between pixels (e.g., superpixels).
40 EM procedure Introduction [ min y {0,1} N K, y1 K =1 N min A IR d K, b IR K ] E U (y,a,b) +E B (y) H(y).
41 EM procedure Introduction [ min y {0,1} N K, y1 K =1 N min A IR d K, b IR K ] E U (y,a,b) +E B (y) H(y). This cost function is not jointly convex in y and (A,b).
42 EM procedure Introduction [ min y {0,1} N K, y1 K =1 N min A IR d K, b IR K ] E U (y,a,b) +E B (y) H(y). This cost function is not jointly convex in y and (A,b). However it is convex in both independently.
43 EM procedure Introduction [ min y {0,1} N K, y1 K =1 N min A IR d K, b IR K ] E U (y,a,b) +E B (y) H(y). This cost function is not jointly convex in y and (A,b). However it is convex in both independently. We alternatively optimize over each variable while fixing the other: We use L-BFGS for (A,b) We use a projected gradient descent for y.
44 The initialization Since our problem is not convex, a good initialization is crucial We propose a quadratic convex approximation related to Joulin et al. (CVPR 10). Quadratic function may lead to poor solutions, thus we also use random initializations.
45 Initialization: Quadratic approximation The second-order Taylor expansion of our cost function is: J(y) = K [ tr(yy T C)+ 2µ 2 NK tr(yyt L) 1 ] N tr(yyt Π I ), where C = 1 N Π N(I Φ(NλI K +Φ T Π N Φ) 1 Φ T )Π N is related to the reweighted ridge regression classifier (Joulin et al. CVPR 10).
46 Initialization: Quadratic approximation The second-order Taylor expansion of our cost function is: J(y) = K [ tr(yy T C)+ 2µ 2 NK tr(yyt L) 1 ] N tr(yyt Π I ), where C = 1 N Π N(I Φ(NλI K +Φ T Π N Φ) 1 Φ T )Π N is related to the reweighted ridge regression classifier (Joulin et al. CVPR 10). This is not convex because of the last term which can be replaced by the following linear constraints: y nk 0.9N i ; y nk 0.1(N N i ). n N i n N j j I\i we obtain a formulation similar to Joulin et al. (CVPR 10).
47 Introduction Binary segmentation (foreground/background) on MSRC: High variability in foreground and background, around 30 images per classes, We use SIFT features. Multiclass cosegmentation on icoseg: Low variability in the image, same illumination... around 10 images per classes, We use color histograms. Some extensions: Grabcut. weakly supervised problem video key frames.
48 Binary cosegmentation
49 Binary cosegmentation
50 Binary cosegmentation class Ours Kim et al. (ICCV 11) Joulin et al. (CVPR 10) Bike Bird Car Cat Chair Cow Dog Face Flower House Plane Sheep Sign Tree Average
51
52
53 class K Ours Joulin et al. CVPR 10 Kim et al ICCV 11 Baseball player Brown bear Elephant Ferrari Football player Helicopter Kite Panda Monk Panda Skating Stonehedge Plane Face Average
54 Extensions Introduction grabcut: Weakly supervised learning with image tags ({ plane, sheep, sky, grass}). Video shot segmentation:
55 Limitations Introduction Number of classes: Each class must be in each image (because of the entropy). Running time: About half an hour to one hour (MATLAB implementation).
56 Thank you.
Discriminative Clustering for Image Co-segmentation
Discriminative Clustering for Image Co-segmentation Armand Joulin 1,2 Francis Bach 1,2 Jean Ponce 2 1 INRIA 2 Ecole Normale Supérieure January 15, 2010 Introduction Introduction problem: dividing simultaneously
More informationDiscriminative Clustering for Image Co-segmentation
Discriminative Clustering for Image Co-segmentation Armand Joulin Francis Bach Jean Ponce INRIA Ecole Normale Supérieure, Paris January 2010 Introduction Introduction Task: dividing simultaneously q images
More informationMulti-Class Cosegmentation
Multi-Class Cosegmentation Armand Joulin 1,2,3 Francis Bach 1,4 JeanPonce 2,3 1 INRIA 23 avenue d Italie, 75214 Paris, France. 2 EcoleNormale Supérieure 45rue d Ulm 75005 Paris, France. Abstract Bottom-up,
More informationDiscriminative Clustering for Image Co-Segmentation
Discriminative Clustering for Image Co-Segmentation Joulin, A.; Bach, F.; Ponce, J. (CVPR. 2010) Iretiayo Akinola Josh Tennefoss Outline Why Co-segmentation? Previous Work Problem Formulation Experimental
More informationUNSUPERVISED CO-SEGMENTATION BASED ON A NEW GLOBAL GMM CONSTRAINT IN MRF. Hongkai Yu, Min Xian, and Xiaojun Qi
UNSUPERVISED CO-SEGMENTATION BASED ON A NEW GLOBAL GMM CONSTRAINT IN MRF Hongkai Yu, Min Xian, and Xiaojun Qi Department of Computer Science, Utah State University, Logan, UT 84322-4205 hongkai.yu@aggiemail.usu.edu,
More informationAn Efficient Image Co-segmentation Algorithm based on Active Contour and Image Saliency
An Efficient Image Co-segmentation Algorithm based on Active Contour and Image Saliency Zhizhi Zhang 1, Xiabi Liu 1, Nouman Qadeer Soomro 2 1 Beijing Laboratory of Intelligent Information Technology, School
More informationUNSUPERVISED COSEGMENTATION BASED ON SUPERPIXEL MATCHING AND FASTGRABCUT. Hongkai Yu and Xiaojun Qi
UNSUPERVISED COSEGMENTATION BASED ON SUPERPIXEL MATCHING AND FASTGRABCUT Hongkai Yu and Xiaojun Qi Department of Computer Science, Utah State University, Logan, UT 84322-4205 hongkai.yu@aggiemail.usu.edu
More informationImage Co-Segmentation via Consistent Functional Maps
Image Co-Segmentation via Consistent Functional Maps Fan Wang Stanford University fanw@stanford.edu Qixing Huang Stanford University huangqx@stanford.edu Leonidas J. Guibas Stanford University guibas@cs.stanford.edu
More informationAnalysis: TextonBoost and Semantic Texton Forests. Daniel Munoz Februrary 9, 2009
Analysis: TextonBoost and Semantic Texton Forests Daniel Munoz 16-721 Februrary 9, 2009 Papers [shotton-eccv-06] J. Shotton, J. Winn, C. Rother, A. Criminisi, TextonBoost: Joint Appearance, Shape and Context
More informationGenerative and discriminative classification techniques
Generative and discriminative classification techniques Machine Learning and Category Representation 013-014 Jakob Verbeek, December 13+0, 013 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.13.14
More informationSegmentation: Clustering, Graph Cut and EM
Segmentation: Clustering, Graph Cut and EM Ying Wu Electrical Engineering and Computer Science Northwestern University, Evanston, IL 60208 yingwu@northwestern.edu http://www.eecs.northwestern.edu/~yingwu
More informationCRFs for Image Classification
CRFs for Image Classification Devi Parikh and Dhruv Batra Carnegie Mellon University Pittsburgh, PA 15213 {dparikh,dbatra}@ece.cmu.edu Abstract We use Conditional Random Fields (CRFs) to classify regions
More informationA Hierarchical Image Clustering Cosegmentation Framework
A Hierarchical Image Clustering Cosegmentation Framework Edward Kim, Hongsheng Li, Xiaolei Huang Department of Computer Science and Engineering Lehigh University, PA 18015 {edk208,h.li,xih206}@lehigh.edu
More informationDiscriminative clustering for image co-segmentation
Discriminative clustering for image co-segmentation Armand Joulin 1,2,3 Francis Bach 1,3 JeanPonce 2,3 1 INRIA 23 avenue d Italie, 75214 Paris, France. 2 EcoleNormale Supérieure 45rue d Ulm 75005 Paris,
More informationTagProp: Discriminative Metric Learning in Nearest Neighbor Models for Image Annotation
TagProp: Discriminative Metric Learning in Nearest Neighbor Models for Image Annotation Matthieu Guillaumin, Thomas Mensink, Jakob Verbeek, Cordelia Schmid LEAR team, INRIA Rhône-Alpes, Grenoble, France
More informationLecture 13 Segmentation and Scene Understanding Chris Choy, Ph.D. candidate Stanford Vision and Learning Lab (SVL)
Lecture 13 Segmentation and Scene Understanding Chris Choy, Ph.D. candidate Stanford Vision and Learning Lab (SVL) http://chrischoy.org Stanford CS231A 1 Understanding a Scene Objects Chairs, Cups, Tables,
More informationConstrained Convolutional Neural Networks for Weakly Supervised Segmentation. Deepak Pathak, Philipp Krähenbühl and Trevor Darrell
Constrained Convolutional Neural Networks for Weakly Supervised Segmentation Deepak Pathak, Philipp Krähenbühl and Trevor Darrell 1 Multi-class Image Segmentation Assign a class label to each pixel in
More informationMarkov Networks in Computer Vision
Markov Networks in Computer Vision Sargur Srihari srihari@cedar.buffalo.edu 1 Markov Networks for Computer Vision Some applications: 1. Image segmentation 2. Removal of blur/noise 3. Stereo reconstruction
More informationMarkov Networks in Computer Vision. Sargur Srihari
Markov Networks in Computer Vision Sargur srihari@cedar.buffalo.edu 1 Markov Networks for Computer Vision Important application area for MNs 1. Image segmentation 2. Removal of blur/noise 3. Stereo reconstruction
More informationAUTOMATIC IMAGE CO-SEGMENTATION USING GEOMETRIC MEAN SALIENCY
AUTOMATIC IMAGE CO-SEGMENTATION USING GEOMETRIC MEAN SALIENCY Koteswar Rao Jerripothula Jianfei Cai Fanman Meng Junsong Yuan ROSE Lab, Interdisciplinary Graduate School, Nanyang Technological University,
More informationCPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2017
CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2017 Assignment 3: 2 late days to hand in tonight. Admin Assignment 4: Due Friday of next week. Last Time: MAP Estimation MAP
More informationMetric learning approaches! for image annotation! and face recognition!
Metric learning approaches! for image annotation! and face recognition! Jakob Verbeek" LEAR Team, INRIA Grenoble, France! Joint work with :"!Matthieu Guillaumin"!!Thomas Mensink"!!!Cordelia Schmid! 1 2
More informationData Analysis 3. Support Vector Machines. Jan Platoš October 30, 2017
Data Analysis 3 Support Vector Machines Jan Platoš October 30, 2017 Department of Computer Science Faculty of Electrical Engineering and Computer Science VŠB - Technical University of Ostrava Table of
More informationCS4670 / 5670: Computer Vision Noah Snavely
{ { 11/26/2013 CS4670 / 5670: Computer Vision Noah Snavely Graph-Based Image Segmentation Stereo as a minimization problem match cost Want each pixel to find a good match in the other image smoothness
More informationUnsupervised Multi-Class Joint Image Segmentation
Unsupervised Multi-Class Joint Image Segmentation Fan Wang Stanford University fanw@stanford.edu Qixing Huang Stanford University huangqx@stanford.edu Leonidas J. Guibas Stanford University guibas@cs.stanford.edu
More informationAnnouncements. Image Segmentation. From images to objects. Extracting objects. Status reports next Thursday ~5min presentations in class
Image Segmentation Announcements Status reports next Thursday ~5min presentations in class Project voting From Sandlot Science Today s Readings Forsyth & Ponce, Chapter 1 (plus lots of optional references
More informationMarkov Random Fields and Segmentation with Graph Cuts
Markov Random Fields and Segmentation with Graph Cuts Computer Vision Jia-Bin Huang, Virginia Tech Many slides from D. Hoiem Administrative stuffs Final project Proposal due Oct 27 (Thursday) HW 4 is out
More informationK-Means and Gaussian Mixture Models
K-Means and Gaussian Mixture Models David Rosenberg New York University June 15, 2015 David Rosenberg (New York University) DS-GA 1003 June 15, 2015 1 / 43 K-Means Clustering Example: Old Faithful Geyser
More informationImproving Image Segmentation Quality Via Graph Theory
International Symposium on Computers & Informatics (ISCI 05) Improving Image Segmentation Quality Via Graph Theory Xiangxiang Li, Songhao Zhu School of Automatic, Nanjing University of Post and Telecommunications,
More information3 October, 2013 MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos
Machine Learning for Computer Vision 1 3 October, 2013 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos Iasonas.kokkinos@ecp.fr Department of Applied Mathematics Ecole Centrale
More informationCPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016
CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2016 A2/Midterm: Admin Grades/solutions will be posted after class. Assignment 4: Posted, due November 14. Extra office hours:
More informationApplied Bayesian Nonparametrics 5. Spatial Models via Gaussian Processes, not MRFs Tutorial at CVPR 2012 Erik Sudderth Brown University
Applied Bayesian Nonparametrics 5. Spatial Models via Gaussian Processes, not MRFs Tutorial at CVPR 2012 Erik Sudderth Brown University NIPS 2008: E. Sudderth & M. Jordan, Shared Segmentation of Natural
More informationImage Segmentation Using Iterated Graph Cuts Based on Multi-scale Smoothing
Image Segmentation Using Iterated Graph Cuts Based on Multi-scale Smoothing Tomoyuki Nagahashi 1, Hironobu Fujiyoshi 1, and Takeo Kanade 2 1 Dept. of Computer Science, Chubu University. Matsumoto 1200,
More informationEnergy Minimization for Segmentation in Computer Vision
S * = arg S min E(S) Energy Minimization for Segmentation in Computer Vision Meng Tang, Dmitrii Marin, Ismail Ben Ayed, Yuri Boykov Outline Clustering/segmentation methods K-means, GrabCut, Normalized
More informationCLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS
CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of
More informationComputer Vision. Exercise Session 10 Image Categorization
Computer Vision Exercise Session 10 Image Categorization Object Categorization Task Description Given a small number of training images of a category, recognize a-priori unknown instances of that category
More informationMachine Learning : Clustering, Self-Organizing Maps
Machine Learning Clustering, Self-Organizing Maps 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps Clustering The task: partition a set of objects into meaningful subsets (clusters). The
More informationon learned visual embedding patrick pérez Allegro Workshop Inria Rhônes-Alpes 22 July 2015
on learned visual embedding patrick pérez Allegro Workshop Inria Rhônes-Alpes 22 July 2015 Vector visual representation Fixed-size image representation High-dim (100 100,000) Generic, unsupervised: BoW,
More informationPart 5: Structured Support Vector Machines
Part 5: Structured Support Vector Machines Sebastian Nowozin and Christoph H. Lampert Providence, 21st June 2012 1 / 34 Problem (Loss-Minimizing Parameter Learning) Let d(x, y) be the (unknown) true data
More informationCosegmentation Revisited: Models and Optimization
Cosegmentation Revisited: Models and Optimization Sara Vicente 1, Vladimir Kolmogorov 1, and Carsten Rother 2 1 University College London 2 Microsoft Research Cambridge Abstract. The problem of cosegmentation
More informationExpectation Maximization (EM) and Gaussian Mixture Models
Expectation Maximization (EM) and Gaussian Mixture Models Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 2 3 4 5 6 7 8 Unsupervised Learning Motivation
More informationDiscrete Optimization of Ray Potentials for Semantic 3D Reconstruction
Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction Marc Pollefeys Joined work with Nikolay Savinov, Christian Haene, Lubor Ladicky 2 Comparison to Volumetric Fusion Higher-order ray
More informationCRF Based Point Cloud Segmentation Jonathan Nation
CRF Based Point Cloud Segmentation Jonathan Nation jsnation@stanford.edu 1. INTRODUCTION The goal of the project is to use the recently proposed fully connected conditional random field (CRF) model to
More information( ) =cov X Y = W PRINCIPAL COMPONENT ANALYSIS. Eigenvectors of the covariance matrix are the principal components
Review Lecture 14 ! PRINCIPAL COMPONENT ANALYSIS Eigenvectors of the covariance matrix are the principal components 1. =cov X Top K principal components are the eigenvectors with K largest eigenvalues
More information9.1. K-means Clustering
424 9. MIXTURE MODELS AND EM Section 9.2 Section 9.3 Section 9.4 view of mixture distributions in which the discrete latent variables can be interpreted as defining assignments of data points to specific
More informationImage Segmentation Using Iterated Graph Cuts BasedonMulti-scaleSmoothing
Image Segmentation Using Iterated Graph Cuts BasedonMulti-scaleSmoothing Tomoyuki Nagahashi 1, Hironobu Fujiyoshi 1, and Takeo Kanade 2 1 Dept. of Computer Science, Chubu University. Matsumoto 1200, Kasugai,
More informationSeparating Objects and Clutter in Indoor Scenes
Separating Objects and Clutter in Indoor Scenes Salman H. Khan School of Computer Science & Software Engineering, The University of Western Australia Co-authors: Xuming He, Mohammed Bennamoun, Ferdous
More informationLearning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009
Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer
More informationJoint Inference in Image Databases via Dense Correspondence. Michael Rubinstein MIT CSAIL (while interning at Microsoft Research)
Joint Inference in Image Databases via Dense Correspondence Michael Rubinstein MIT CSAIL (while interning at Microsoft Research) My work Throughout the year (and my PhD thesis): Temporal Video Analysis
More informationLearning Representations for Visual Object Class Recognition
Learning Representations for Visual Object Class Recognition Marcin Marszałek Cordelia Schmid Hedi Harzallah Joost van de Weijer LEAR, INRIA Grenoble, Rhône-Alpes, France October 15th, 2007 Bag-of-Features
More informationSupport Vector Machines.
Support Vector Machines srihari@buffalo.edu SVM Discussion Overview 1. Overview of SVMs 2. Margin Geometry 3. SVM Optimization 4. Overlapping Distributions 5. Relationship to Logistic Regression 6. Dealing
More informationSupervised texture detection in images
Supervised texture detection in images Branislav Mičušík and Allan Hanbury Pattern Recognition and Image Processing Group, Institute of Computer Aided Automation, Vienna University of Technology Favoritenstraße
More informationObject Segmentation and Tracking in 3D Video With Sparse Depth Information Using a Fully Connected CRF Model
Object Segmentation and Tracking in 3D Video With Sparse Depth Information Using a Fully Connected CRF Model Ido Ofir Computer Science Department Stanford University December 17, 2011 Abstract This project
More informationSpatial Latent Dirichlet Allocation
Spatial Latent Dirichlet Allocation Xiaogang Wang and Eric Grimson Computer Science and Computer Science and Artificial Intelligence Lab Massachusetts Tnstitute of Technology, Cambridge, MA, 02139, USA
More informationSegmentation with non-linear constraints on appearance, complexity, and geometry
IPAM February 2013 Western Univesity Segmentation with non-linear constraints on appearance, complexity, and geometry Yuri Boykov Andrew Delong Lena Gorelick Hossam Isack Anton Osokin Frank Schmidt Olga
More informationBilevel Sparse Coding
Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional
More informationClustering. So far in the course. Clustering. Clustering. Subhransu Maji. CMPSCI 689: Machine Learning. dist(x, y) = x y 2 2
So far in the course Clustering Subhransu Maji : Machine Learning 2 April 2015 7 April 2015 Supervised learning: learning with a teacher You had training data which was (feature, label) pairs and the goal
More informationGradient of the lower bound
Weakly Supervised with Latent PhD advisor: Dr. Ambedkar Dukkipati Department of Computer Science and Automation gaurav.pandey@csa.iisc.ernet.in Objective Given a training set that comprises image and image-level
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Clustering and EM Barnabás Póczos & Aarti Singh Contents Clustering K-means Mixture of Gaussians Expectation Maximization Variational Methods 2 Clustering 3 K-
More informationhuman vision: grouping k-means clustering graph-theoretic clustering Hough transform line fitting RANSAC
COS 429: COMPUTER VISON Segmentation human vision: grouping k-means clustering graph-theoretic clustering Hough transform line fitting RANSAC Reading: Chapters 14, 15 Some of the slides are credited to:
More informationLecture 11: E-M and MeanShift. CAP 5415 Fall 2007
Lecture 11: E-M and MeanShift CAP 5415 Fall 2007 Review on Segmentation by Clustering Each Pixel Data Vector Example (From Comanciu and Meer) Review of k-means Let's find three clusters in this data These
More informationSparse Coding and Dictionary Learning for Image Analysis
Sparse Coding and Dictionary Learning for Image Analysis Part IV: Recent Advances in Computer Vision and New Models Francis Bach, Julien Mairal, Jean Ponce and Guillermo Sapiro CVPR 10 tutorial, San Francisco,
More informationThe goals of segmentation
Image segmentation The goals of segmentation Group together similar-looking pixels for efficiency of further processing Bottom-up process Unsupervised superpixels X. Ren and J. Malik. Learning a classification
More informationHuman detection using histogram of oriented gradients. Srikumar Ramalingam School of Computing University of Utah
Human detection using histogram of oriented gradients Srikumar Ramalingam School of Computing University of Utah Reference Navneet Dalal and Bill Triggs, Histograms of Oriented Gradients for Human Detection,
More informationCS4670: Computer Vision
CS4670: Computer Vision Noah Snavely Lecture 34: Segmentation From Sandlot Science Announcements In-class exam this Friday, December 3 Review session in class on Wednesday Final projects: Slides due: Sunday,
More informationA Taxonomy of Semi-Supervised Learning Algorithms
A Taxonomy of Semi-Supervised Learning Algorithms Olivier Chapelle Max Planck Institute for Biological Cybernetics December 2005 Outline 1 Introduction 2 Generative models 3 Low density separation 4 Graph
More informationPart III: Affinity Functions for Image Segmentation
Part III: Affinity Functions for Image Segmentation Charless Fowlkes joint work with David Martin and Jitendra Malik at University of California at Berkeley 1 Q: What measurements should we use for constructing
More information4/13/ Introduction. 1. Introduction. 2. Formulation. 2. Formulation. 2. Formulation
1. Introduction Motivation: Beijing Jiaotong University 1 Lotus Hill Research Institute University of California, Los Angeles 3 CO 3 for Ultra-fast and Accurate Interactive Image Segmentation This paper
More informationImage Segmentation continued Graph Based Methods
Image Segmentation continued Graph Based Methods Previously Images as graphs Fully-connected graph node (vertex) for every pixel link between every pair of pixels, p,q affinity weight w pq for each link
More informationBag-of-features. Cordelia Schmid
Bag-of-features for category classification Cordelia Schmid Visual search Particular objects and scenes, large databases Category recognition Image classification: assigning a class label to the image
More informationClustering. Subhransu Maji. CMPSCI 689: Machine Learning. 2 April April 2015
Clustering Subhransu Maji CMPSCI 689: Machine Learning 2 April 2015 7 April 2015 So far in the course Supervised learning: learning with a teacher You had training data which was (feature, label) pairs
More informationSelection of Scale-Invariant Parts for Object Class Recognition
Selection of Scale-Invariant Parts for Object Class Recognition Gy. Dorkó and C. Schmid INRIA Rhône-Alpes, GRAVIR-CNRS 655, av. de l Europe, 3833 Montbonnot, France fdorko,schmidg@inrialpes.fr Abstract
More informationCS 534: Computer Vision Segmentation and Perceptual Grouping
CS 534: Computer Vision Segmentation and Perceptual Grouping Ahmed Elgammal Dept of Computer Science CS 534 Segmentation - 1 Outlines Mid-level vision What is segmentation Perceptual Grouping Segmentation
More informationEE 701 ROBOT VISION. Segmentation
EE 701 ROBOT VISION Regions and Image Segmentation Histogram-based Segmentation Automatic Thresholding K-means Clustering Spatial Coherence Merging and Splitting Graph Theoretic Segmentation Region Growing
More informationCIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]
CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, 2015. 11:59pm, PDF to Canvas [100 points] Instructions. Please write up your responses to the following problems clearly and concisely.
More informationMachine Learning: Think Big and Parallel
Day 1 Inderjit S. Dhillon Dept of Computer Science UT Austin CS395T: Topics in Multicore Programming Oct 1, 2013 Outline Scikit-learn: Machine Learning in Python Supervised Learning day1 Regression: Least
More informationCo-segmentation of Non-homogeneous Image Sets
1 Co-segmentation of Non-homogeneous Image Sets Avik Hati Subhasis Chaudhuri Rajbabu Velmurugan Abstract In this paper, we formulate image co-segmentation as a classification problem in an unsupervised
More informationBlind Image Deblurring Using Dark Channel Prior
Blind Image Deblurring Using Dark Channel Prior Jinshan Pan 1,2,3, Deqing Sun 2,4, Hanspeter Pfister 2, and Ming-Hsuan Yang 3 1 Dalian University of Technology 2 Harvard University 3 UC Merced 4 NVIDIA
More informationMetric Learning for Large Scale Image Classification:
Metric Learning for Large Scale Image Classification: Generalizing to New Classes at Near-Zero Cost Thomas Mensink 1,2 Jakob Verbeek 2 Florent Perronnin 1 Gabriela Csurka 1 1 TVPA - Xerox Research Centre
More information08 An Introduction to Dense Continuous Robotic Mapping
NAVARCH/EECS 568, ROB 530 - Winter 2018 08 An Introduction to Dense Continuous Robotic Mapping Maani Ghaffari March 14, 2018 Previously: Occupancy Grid Maps Pose SLAM graph and its associated dense occupancy
More informationImage Segmentation continued Graph Based Methods. Some slides: courtesy of O. Capms, Penn State, J.Ponce and D. Fortsyth, Computer Vision Book
Image Segmentation continued Graph Based Methods Some slides: courtesy of O. Capms, Penn State, J.Ponce and D. Fortsyth, Computer Vision Book Previously Binary segmentation Segmentation by thresholding
More information570 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 2, FEBRUARY 2011
570 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 2, FEBRUARY 2011 Contextual Object Localization With Multiple Kernel Nearest Neighbor Brian McFee, Student Member, IEEE, Carolina Galleguillos, Student
More informationDeep Generative Models Variational Autoencoders
Deep Generative Models Variational Autoencoders Sudeshna Sarkar 5 April 2017 Generative Nets Generative models that represent probability distributions over multiple variables in some way. Directed Generative
More informationString distance for automatic image classification
String distance for automatic image classification Nguyen Hong Thinh*, Le Vu Ha*, Barat Cecile** and Ducottet Christophe** *University of Engineering and Technology, Vietnam National University of HaNoi,
More informationRobotics Programming Laboratory
Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car
More informationLogistic Regression
Logistic Regression ddebarr@uw.edu 2016-05-26 Agenda Model Specification Model Fitting Bayesian Logistic Regression Online Learning and Stochastic Optimization Generative versus Discriminative Classifiers
More informationAll lecture slides will be available at CSC2515_Winter15.html
CSC2515 Fall 2015 Introduc3on to Machine Learning Lecture 9: Support Vector Machines All lecture slides will be available at http://www.cs.toronto.edu/~urtasun/courses/csc2515/ CSC2515_Winter15.html Many
More informationMixture Models and EM
Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering
More informationEstimating Human Pose in Images. Navraj Singh December 11, 2009
Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks
More informationTri-modal Human Body Segmentation
Tri-modal Human Body Segmentation Master of Science Thesis Cristina Palmero Cantariño Advisor: Sergio Escalera Guerrero February 6, 2014 Outline 1 Introduction 2 Tri-modal dataset 3 Proposed baseline 4
More informationCS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas
CS839: Probabilistic Graphical Models Lecture 10: Learning with Partially Observed Data Theo Rekatsinas 1 Partially Observed GMs Speech recognition 2 Partially Observed GMs Evolution 3 Partially Observed
More informationMulti-Class Segmentation with Relative Location Prior
Multi-Class Segmentation with Relative Location Prior Stephen Gould, Jim Rodgers, David Cohen, Gal Elidan, Daphne Koller Department of Computer Science, Stanford University International Journal of Computer
More informationIntroduction to object recognition. Slides adapted from Fei-Fei Li, Rob Fergus, Antonio Torralba, and others
Introduction to object recognition Slides adapted from Fei-Fei Li, Rob Fergus, Antonio Torralba, and others Overview Basic recognition tasks A statistical learning approach Traditional or shallow recognition
More informationintro, applications MRF, labeling... how it can be computed at all? Applications in segmentation: GraphCut, GrabCut, demos
Image as Markov Random Field and Applications 1 Tomáš Svoboda, svoboda@cmp.felk.cvut.cz Czech Technical University in Prague, Center for Machine Perception http://cmp.felk.cvut.cz Talk Outline Last update:
More informationVisual Recognition: Examples of Graphical Models
Visual Recognition: Examples of Graphical Models Raquel Urtasun TTI Chicago March 6, 2012 Raquel Urtasun (TTI-C) Visual Recognition March 6, 2012 1 / 64 Graphical models Applications Representation Inference
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 19: Graph Cuts source S sink T Readings Szeliski, Chapter 11.2 11.5 Stereo results with window search problems in areas of uniform texture Window-based matching
More informationIMPROVED FINE STRUCTURE MODELING VIA GUIDED STOCHASTIC CLIQUE FORMATION IN FULLY CONNECTED CONDITIONAL RANDOM FIELDS
IMPROVED FINE STRUCTURE MODELING VIA GUIDED STOCHASTIC CLIQUE FORMATION IN FULLY CONNECTED CONDITIONAL RANDOM FIELDS M. J. Shafiee, A. G. Chung, A. Wong, and P. Fieguth Vision & Image Processing Lab, Systems
More informationInteractive Image Segmentation with GrabCut
Interactive Image Segmentation with GrabCut Bryan Anenberg Stanford University anenberg@stanford.edu Michela Meister Stanford University mmeister@stanford.edu Abstract We implement GrabCut and experiment
More informationUnsupervised discovery of category and object models. The task
Unsupervised discovery of category and object models Martial Hebert The task 1 Common ingredients 1. Generate candidate segments 2. Estimate similarity between candidate segments 3. Prune resulting (implicit)
More informationSupport Vector Machines.
Support Vector Machines srihari@buffalo.edu SVM Discussion Overview. Importance of SVMs. Overview of Mathematical Techniques Employed 3. Margin Geometry 4. SVM Training Methodology 5. Overlapping Distributions
More information