Experts-Shift: Learning Active Spatial Classification Experts for Keyframe-based Video Segmentation
|
|
- Philip McKinney
- 6 years ago
- Views:
Transcription
1 Experts-Shift: Learning Active Spatial Classification Experts for Keyframe-based Video Segmentation Yibiao Zhao 1,3, Yanbiao Duan 2,3, Xiaohan Nie 2,3, Yaping Huang 1, Siwei Luo 1 1 Beijing Jiaotong University, 2 Beijing Institute of Technology, 3 Lotus Hill Institute {ybzhao.lhi,ybduan.lhi,xhnie.lhi}@gmail.com,{yphuang,swluo}@bjtu.edu.cn Abstract Experts-Shift is a novel statistical framework for keyframe-based video segmentation. Compared to existing video segmentation techniques with simple color models, our method proposes a probability mixture model coupling strong image classifiers (experts) with latent spatial configuration. In order to propagate image labels to the successive frames, our algorithm track all experts jointly by a efficient MCMC sampler with their relations modeled by MRFs. This algorithm is capable to handle overlapping color distribution, ambiguous image boundaries, large displacement in challenging scenario with a solid foundation of both generative modeling and discriminative learning. Experiment shows our algorithm achieves high quality results and need less supervision than previous work. Figure 1. Probabilistic Graphical Model representing the Mixture of Classification Experts. 1. Introduction Keyframe-based video segmentation, the process to propagate segmentation label from manually labeled keyframe to successive frames, is an essential technique in many video applications. Several existing approaches consider video as image sequence [9, 8] or 3D volume [2, ], and attempt to solve the problem by modeling foreground vs. background distributions in the color space. In some challenging video sequences, unfortunately, foreground and background regions contain the similar colors or very ambiguous image patterns, which will confuse algorithms relying only on the global color statistics (Fig.2(a)). Video Snapcut [3] first addressed this problem, and learned foreground/background color models in localized features along object boundary. The SIFT matching together with optical flow algorithm is applied to build correspondence between successive frames. In this paper, we aim to model the local image pattern and the spatial position simultaneously. By introducing a location-sensitive variable, the latent relation between target label and spatial position are explicitly established. The Figure 2. The learned experts (color ellipses) seem to be very meaningful in that they represent identifiable spatial areas and handle the local ambiguities (eg. area in red and violet ellipses) quite well. mixture model thus acts as an ensemble of localized classifiers, namely experts, with each of which in charge of a spatial domain of Gaussian (illustrated by color ellipses in Fig.2(b)). It naturally embraces the generative nature of spatial configuration and the discriminative power of local image feature. The mixture model is estimated in EM fashion: experts compete for the ownership of pixels at E step; //$26. 2 IEEE 622
2 and Boosting algorithm is applied to distinguish the subtle difference in their domination area at M step. Moreover, we design an efficient model-adapting scheme based on MCMC sampler: Markov Random Field model captures pairwise relations between experts, which enables reliable multiple-target tracking and maintains a stable topological relationship between experts (illustrated in Fig.2(b)). 2. Mixture Model of Classification Experts In our method, instead of learning the complicated distribution with local image features x and spatial position s jointly, we introduce a location-sensitive latent variable z for each pixel n = 1,, N to infer the latent relations between spatial position s and target variable y on the keyframe t. z is a K-dimensional binary random variable having a 1-of-K representation in which a particular element z k is equal to 1 and all other elements are equal to. The target distribution p(y t x t, s t ) (Considering a single frame, we omit the superscript t in the rest of this section for simplification) is obtained by marginalization of joint distribution p(y, z x, s) over the latent variables z, which allows complicated distributions to be formed from simpler components. p(y x, s) = k K p(y, z k x, s) = k K p(y x, z k )p(z k s) The model is thus formulated as an ensemble of K discriminative classifiers p(y x, z k ). Each classifier, known as an expert, is an independent discriminative model associated to the location-sensitive latent variable z k. The latent relation between observed variable s and target variable y are explicitly established by z k. (the corresponding graphical model representation is depicted in Fig.1) Each expert is learned by a discriminative classifier of Boosting, p(y x, z k = 1) = exp{ 2 y F (x)} where F (x) = i λ if i (x) is a strong classifier, namely an ensemble of weak classifiers f i (x) with coefficients λ i. The classifiers with strong discriminative power enable using rich image features without additional effort on modeling the observed data distribution p(x) in large feature space. To obtain good generalization ability, we explicitly formulate a generative distribution of spatial position for each expert as normal distribution, p(s z k = 1) = N(s µ k, Σ k ). With Bayesian rules, we get the posterior by π k N(s µ k, Σ k ) p(z k = 1 s) = j K π jn(s µ j, Σ j ) (1) (2) (3) which π k = P (z k = 1) is mixing coefficients of each Gaussian component. A intuitive example is shown in Fig.2. We visualize the seven experts by ellipses indicating their latent spatial distribution. The learned experts seem to be very meaningful in that they represent identifiable spatial areas and handle the local ambiguities quite well. Model Estimation with the EM The standard procedure for maximum likelihood estimation in latent variable models is the Expectation Maximization (EM) algorithm, which alternates two steps: (a) E-step is an expectation step where posterior probabilities are computed for the latent variables, p(y x, z k )p(z k s) γ(z k ) = p(z k y, x, s) = j K p(y x, z j)p(z j s) (b) M-step is a maximization step where parameters are updated by taking the expectation of the complete-data log likelihood with respect to the posterior distribution of the latent variables, (4) Θ k = arg max n N γ(z k ) ln(p(y z k, x)p(z k s)) (5) The parameters π k, µ k, Σ k for the spatial model p(z k s) are estimated same as original Gaussian Mixture Model. Meanwhile, the GentleBoost algorithm [6] is performed, and it optimizes the local image feature model with an exponential criterion exp( yf (x)) which to second order is proved to be equivalent to the binomial log-likelihood criterion [6]. 3. Joint Experts-Shift with MRF Having learned classification experts on the keyframe, p(y t x t, s t ; Θ t ), in this section, we focus on tracking the expert Θ t to the next frame Θ t+1 according to local image evidence x t, x t+1 between two successive frames. Benefiting from the Gaussian model of spatial position N(s; µ k, Σ k ) for each expert, the mixture model is very flexible to be explicitly updated. Inspired by [7], we model the pair-wise relation between experts by a Markov Random Field model. The MRF defines a graph structure (V, E) (shown in Fig.3(d)) with undirected edges E between nodes V where the joint probability is factored as a product of local potential functions on each clique, p(θ t+1 Θ t ) φ(θ t+1 i, Θ t i) ψ(θ t+1 i, Θ t+1 j ) (6) i i,j E where φ(θ t+1 i, Θ t i ) are transition models between successive frames capturing individual expert s motion, and ψ(θ t i, Θt j ) are pairwise interaction potentials proportion to, exp{ min(n(s µ i, Σ i ), N(s µ j, Σ j ))dxdy}. (7) 623
3 Figure 3. Experts (top), probability map (bottom left) and error map (bottom right) with different learning strategies (a-c). (d) MRF graph structure. where min(, )dxdy is the intersection distance between two distributions. An efficient MCMC sampling step of Metropolis- Hastings (MH) algorithm is applied to optimize the MRF model as introduced in [7]. In this paper, we do not elaborate it further. Incorporating motion prior After determined the position of each expert, we can estimate how much does the entire label image shift between two frames. A very simple intuition is that the probability of a pixel on frame t + 1 to be a segmentation boundary will be very high if there are boundary pixel nearby on the last frame t. p(y t+1 y t 1 ) = 1 + exp{ yp t+1 min d(s, v)}, v Bt+1 p yp t+1 (s) = y t (τ(s) + s) (8) 624
4 where we formulate the motion prior as logistic function of signed distance between the pixel s and its nearest point on the boundary B t+1 p, τ(s) is the total shifting of all experts weight on the position of a pixel. As is shown in Fig. 3, the shade confused the single classifier(a), and was solved by introducing the mixture of experts to a great extent (b). The prior forcefully suppresses the ambiguity in the inner regions(c), making the experts focus on the local difference near the object boundary. This is implemented by simply adjusting the initial weight of data according to the prior. We perform the learning process with different strategies Fig.3. The shades in both foreground and background confused the single classifier but is easily solved by mixture of experts. The prior deals with the main component of the foreground and background, leaving the area near the boundary to the experts, and improve the performance greatly. 4. Experiments We evaluate our algorithm on several popular experimental video sequences. Videos in Fig.4 (1-4) and Fig.5 are selected from [8, 9, 3], and LHI dataset [11]. For all experiment, we use eight experts. We apply a recently proposed interactive image segmentation algorithm of CO3 [12] to segment each keyframe. Following the algorithm of CO3, we use Bregman iteration to refine the classification map obtained from learned experts. The speed of post-processing process of Bregman iteration is generally 6 frames per second, and the experts can test 2.3 frames for each second on a common PC. Fig.4 shows some representative results of our algorithm on several challenging video sequences. The keyframes with user interaction are annotated on the time axis. Fig.5 further display algorithm comparison on the fish sequence. For this challenging data, Geodesic Matting [2], Video Cutout [] only based on a simple global color model requires intensive user interaction. Video Snapcut [3] with local color model performs better than the formers. Our method, without any user interaction, successively propagates the desired segment label to the last frame. 5. Conclusion In this paper, we present a statistical framework, Experts-Shift, to deal with complicated video sequence in keyframe-based video segmentation task. We first propose a mixture model to learn local image classifiers (experts), and infer their latent spatial configuration in the EM fashion. Moreover, an efficient joint tracking scheme is designed based on MRF model of all experts, which enables reliable multiple-target tracking and maintains a stable topological relationship. Experiment shows our algorithm achieves high quality results and need less user interaction than previous work. 6. Acknowledgments This work at Beijing Jiaotong University is supported by China 863 Program 27AA1Z168, NSF China grants , 69258, 68541, , , Beijing Natural Science Foundation 9233, and Doctoral Foundations of Ministry of Education of China And the work at the Lotus Hill Institute is supported by China 863 Program 27AA1Z3, 29AA1Z331 and NSF China grants , References [1] A. Agarwala, A. Hertzmann, D. Salesin, and S. Seitz. Keyframe-based tracking for rotoscoping and animation. ACM Transactions on Graphics, pages , 24. [2] X. Bai and G. Sapiro. Geodesic matting: A framework for fast interactive image and video segmentation and matting. International Journal of Computer Vision, 82(2): , 29. [3] X. Bai, J. Wang, D. Simons, and G. Sapiro. Video snapcut: robust video object cutout using localized classifiers. ACM Transactions on Graphics, pages 1 11, 29. [4] X. Bai, J. Wang, D. Simons, and G. Sapiro. Dynamic Color Flow: A Motion-Adaptive Color Model for Object Segmentation in Video. In European Conference on Computer Vision (ECCV), 2. [5] C. M. Bishop. Pattern Recognition and Machine Learning. Springer-Verlag New York, Inc., 26. [6] Y. Y. Boykov and M. P. Jolly. Interactive graph cuts for optimal boundary and region segmentation of objects in n-d images. In Proceeding of International Conference on Computer Vision., volume 1, pages vol.1, 21. [7] J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical view of boosting. Annals of Statatistics, 28:2, 2. [8] Z. Khan, T. Balch, and F. Dellaert. An MCMC-based Particle Filter For Tracking Multiple Interacting Targets. In European Conference on Computer Vision. 23. pages, pages [9] Y. Li, J. Sun, and H. Shum. Video object cut and paste. ACM Transactions on Graphics, pages 595 6, 25. [] B. Price, B. Morse, and S. Cohen. Livecut: Learningbased interactive video segmentation by evaluation of multiple propagated cues. In International Conference on Computer Vision. 29. [11] J. Wang, P. Bhat, R. Colburn, M. Agrawala, and M. Cohen. Interactive video cutout. ACM Transactions on Graphics, pages , 25. [12] B. Yao, X. Yang, and S.C. Zhu. Introduction to a large-scale general purpose ground truth database: Methodology, annotation tool and benchmarks. In EMMCVPR., pages , 27. [13] Y. Zhao, S.C. Zhu and S. Luo, CO3 for Ultra-fast and Accurate Interactive Segmentation. In ACM Multimedia
5 Frame 4 2 Frame Frame Frame Frame Frame 45 Frame 31 5 Frame 6 Frame 28 Frame 5 Frame 38 Frame 19 Frame 37 Frame Figure 4. Representative results of proposed algorithm. Red dots indicates keyframes with user interaction
6 Experts-Shift Geodesic 7 Cutout 5 Snapcut 9 Figure 5. Results of different methods. Red dots indicates keyframes with user interaction. 627
A Benchmark for Interactive Image Segmentation Algorithms
A Benchmark for Interactive Image Segmentation Algorithms Yibiao Zhao 1,3, Xiaohan Nie 2,3, Yanbiao Duan 2,3, Yaping Huang 1, Siwei Luo 1 1 Beijing Jiaotong University, 2 Beijing Institute of Technology,
More information4/13/ Introduction. 1. Introduction. 2. Formulation. 2. Formulation. 2. Formulation
1. Introduction Motivation: Beijing Jiaotong University 1 Lotus Hill Research Institute University of California, Los Angeles 3 CO 3 for Ultra-fast and Accurate Interactive Image Segmentation This paper
More informationIMA Preprint Series # 2153
DISTANCECUT: INTERACTIVE REAL-TIME SEGMENTATION AND MATTING OF IMAGES AND VIDEOS By Xue Bai and Guillermo Sapiro IMA Preprint Series # 2153 ( January 2007 ) INSTITUTE FOR MATHEMATICS AND ITS APPLICATIONS
More informationCO3 for Ultra-fast and Accurate Interactive Segmentation
CO3 for Ultra-fast and Accurate Interactive Segmentation Yibiao Zhao Beijing Jiaotong University; Lotus Hill Research Institute; University of California, Los Angeles. ybzhao@ucla.edu Song-Chun Zhu Lotus
More informationLearning and Recognizing Visual Object Categories Without First Detecting Features
Learning and Recognizing Visual Object Categories Without First Detecting Features Daniel Huttenlocher 2007 Joint work with D. Crandall and P. Felzenszwalb Object Category Recognition Generic classes rather
More informationImage Segmentation Using Iterated Graph Cuts Based on Multi-scale Smoothing
Image Segmentation Using Iterated Graph Cuts Based on Multi-scale Smoothing Tomoyuki Nagahashi 1, Hironobu Fujiyoshi 1, and Takeo Kanade 2 1 Dept. of Computer Science, Chubu University. Matsumoto 1200,
More informationMarkov Random Fields and Gibbs Sampling for Image Denoising
Markov Random Fields and Gibbs Sampling for Image Denoising Chang Yue Electrical Engineering Stanford University changyue@stanfoed.edu Abstract This project applies Gibbs Sampling based on different Markov
More informationSegmentation. Bottom up Segmentation Semantic Segmentation
Segmentation Bottom up Segmentation Semantic Segmentation Semantic Labeling of Street Scenes Ground Truth Labels 11 classes, almost all occur simultaneously, large changes in viewpoint, scale sky, road,
More informationImage Segmentation Using Iterated Graph Cuts BasedonMulti-scaleSmoothing
Image Segmentation Using Iterated Graph Cuts BasedonMulti-scaleSmoothing Tomoyuki Nagahashi 1, Hironobu Fujiyoshi 1, and Takeo Kanade 2 1 Dept. of Computer Science, Chubu University. Matsumoto 1200, Kasugai,
More informationSupervised Learning for Image Segmentation
Supervised Learning for Image Segmentation Raphael Meier 06.10.2016 Raphael Meier MIA 2016 06.10.2016 1 / 52 References A. Ng, Machine Learning lecture, Stanford University. A. Criminisi, J. Shotton, E.
More informationSupervised texture detection in images
Supervised texture detection in images Branislav Mičušík and Allan Hanbury Pattern Recognition and Image Processing Group, Institute of Computer Aided Automation, Vienna University of Technology Favoritenstraße
More informationScene Grammars, Factor Graphs, and Belief Propagation
Scene Grammars, Factor Graphs, and Belief Propagation Pedro Felzenszwalb Brown University Joint work with Jeroen Chua Probabilistic Scene Grammars General purpose framework for image understanding and
More informationGraph-Based Superpixel Labeling for Enhancement of Online Video Segmentation
Graph-Based Superpixel Labeling for Enhancement of Online Video Segmentation Alaa E. Abdel-Hakim Electrical Engineering Department Assiut University Assiut, Egypt alaa.aly@eng.au.edu.eg Mostafa Izz Cairo
More informationImage analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis
7 Computer Vision and Classification 413 / 458 Computer Vision and Classification The k-nearest-neighbor method The k-nearest-neighbor (knn) procedure has been used in data analysis and machine learning
More informationScene Grammars, Factor Graphs, and Belief Propagation
Scene Grammars, Factor Graphs, and Belief Propagation Pedro Felzenszwalb Brown University Joint work with Jeroen Chua Probabilistic Scene Grammars General purpose framework for image understanding and
More informationA Feature Point Matching Based Approach for Video Objects Segmentation
A Feature Point Matching Based Approach for Video Objects Segmentation Yan Zhang, Zhong Zhou, Wei Wu State Key Laboratory of Virtual Reality Technology and Systems, Beijing, P.R. China School of Computer
More informationCS395T paper review. Indoor Segmentation and Support Inference from RGBD Images. Chao Jia Sep
CS395T paper review Indoor Segmentation and Support Inference from RGBD Images Chao Jia Sep 28 2012 Introduction What do we want -- Indoor scene parsing Segmentation and labeling Support relationships
More informationApplied Bayesian Nonparametrics 5. Spatial Models via Gaussian Processes, not MRFs Tutorial at CVPR 2012 Erik Sudderth Brown University
Applied Bayesian Nonparametrics 5. Spatial Models via Gaussian Processes, not MRFs Tutorial at CVPR 2012 Erik Sudderth Brown University NIPS 2008: E. Sudderth & M. Jordan, Shared Segmentation of Natural
More informationAdaptive Feature Selection via Boosting-like Sparsity Regularization
Adaptive Feature Selection via Boosting-like Sparsity Regularization Libin Wang, Zhenan Sun, Tieniu Tan Center for Research on Intelligent Perception and Computing NLPR, Beijing, China Email: {lbwang,
More informationUndirected Graphical Models. Raul Queiroz Feitosa
Undirected Graphical Models Raul Queiroz Feitosa Pros and Cons Advantages of UGMs over DGMs UGMs are more natural for some domains (e.g. context-dependent entities) Discriminative UGMs (CRF) are better
More informationTri-modal Human Body Segmentation
Tri-modal Human Body Segmentation Master of Science Thesis Cristina Palmero Cantariño Advisor: Sergio Escalera Guerrero February 6, 2014 Outline 1 Introduction 2 Tri-modal dataset 3 Proposed baseline 4
More informationSupport Vector Machines
Support Vector Machines About the Name... A Support Vector A training sample used to define classification boundaries in SVMs located near class boundaries Support Vector Machines Binary classifiers whose
More informationEstimating Human Pose in Images. Navraj Singh December 11, 2009
Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks
More informationDynamic Color Flow: A Motion-Adaptive Color Model for Object Segmentation in Video
Dynamic Color Flow: A Motion-Adaptive Color Model for Object Segmentation in Video Xue Bai 1, Jue Wang 2, and Guillermo Sapiro 1 1 University of Minnesota, Minneapolis, MN 55455, USA 2 Adobe Systems, Seattle,
More informationIMPROVED FINE STRUCTURE MODELING VIA GUIDED STOCHASTIC CLIQUE FORMATION IN FULLY CONNECTED CONDITIONAL RANDOM FIELDS
IMPROVED FINE STRUCTURE MODELING VIA GUIDED STOCHASTIC CLIQUE FORMATION IN FULLY CONNECTED CONDITIONAL RANDOM FIELDS M. J. Shafiee, A. G. Chung, A. Wong, and P. Fieguth Vision & Image Processing Lab, Systems
More informationNeighbourhood-consensus message passing and its potentials in image processing applications
Neighbourhood-consensus message passing and its potentials in image processing applications Tijana Ružić a, Aleksandra Pižurica a and Wilfried Philips a a Ghent University, TELIN-IPI-IBBT, Sint-Pietersnieuwstraat
More informationCHAPTER 6 PERCEPTUAL ORGANIZATION BASED ON TEMPORAL DYNAMICS
CHAPTER 6 PERCEPTUAL ORGANIZATION BASED ON TEMPORAL DYNAMICS This chapter presents a computational model for perceptual organization. A figure-ground segregation network is proposed based on a novel boundary
More informationMachine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme
Machine Learning B. Unsupervised Learning B.1 Cluster Analysis Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University of Hildesheim, Germany
More informationMarkov Random Fields and Segmentation with Graph Cuts
Markov Random Fields and Segmentation with Graph Cuts Computer Vision Jia-Bin Huang, Virginia Tech Many slides from D. Hoiem Administrative stuffs Final project Proposal due Oct 27 (Thursday) HW 4 is out
More informationExpectation Maximization (EM) and Gaussian Mixture Models
Expectation Maximization (EM) and Gaussian Mixture Models Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 2 3 4 5 6 7 8 Unsupervised Learning Motivation
More informationPattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition
Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant
More informationCS 664 Flexible Templates. Daniel Huttenlocher
CS 664 Flexible Templates Daniel Huttenlocher Flexible Template Matching Pictorial structures Parts connected by springs and appearance models for each part Used for human bodies, faces Fischler&Elschlager,
More informationImproving Recognition through Object Sub-categorization
Improving Recognition through Object Sub-categorization Al Mansur and Yoshinori Kuno Graduate School of Science and Engineering, Saitama University, 255 Shimo-Okubo, Sakura-ku, Saitama-shi, Saitama 338-8570,
More informationGenerative and discriminative classification techniques
Generative and discriminative classification techniques Machine Learning and Category Representation 2014-2015 Jakob Verbeek, November 28, 2014 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.14.15
More informationAn Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework
IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXX 23 An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework Ji Won Yoon arxiv:37.99v [cs.lg] 3 Jul 23 Abstract In order to cluster
More informationAnalysis: TextonBoost and Semantic Texton Forests. Daniel Munoz Februrary 9, 2009
Analysis: TextonBoost and Semantic Texton Forests Daniel Munoz 16-721 Februrary 9, 2009 Papers [shotton-eccv-06] J. Shotton, J. Winn, C. Rother, A. Criminisi, TextonBoost: Joint Appearance, Shape and Context
More informationCRFs for Image Classification
CRFs for Image Classification Devi Parikh and Dhruv Batra Carnegie Mellon University Pittsburgh, PA 15213 {dparikh,dbatra}@ece.cmu.edu Abstract We use Conditional Random Fields (CRFs) to classify regions
More informationInference and Representation
Inference and Representation Rachel Hodos New York University Lecture 5, October 6, 2015 Rachel Hodos Lecture 5: Inference and Representation Today: Learning with hidden variables Outline: Unsupervised
More informationGenerative and discriminative classification techniques
Generative and discriminative classification techniques Machine Learning and Category Representation 013-014 Jakob Verbeek, December 13+0, 013 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.13.14
More informationHigh-Resolution Image Classification Integrating Spectral-Spatial-Location Cues by Conditional Random Fields
IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 25, NO. 9, SEPTEMBER 2016 4033 High-Resolution Image Classification Integrating Spectral-Spatial-Location Cues by Conditional Random Fields Ji Zhao, Student
More informationIMA Preprint Series # 2171
A GEODESIC FRAMEWORK FOR FAST INTERACTIVE IMAGE AND VIDEO SEGMENTATION AND MATTING By Xue Bai and Guillermo Sapiro IMA Preprint Series # 2171 ( August 2007 ) INSTITUTE FOR MATHEMATICS AND ITS APPLICATIONS
More informationA Taxonomy of Semi-Supervised Learning Algorithms
A Taxonomy of Semi-Supervised Learning Algorithms Olivier Chapelle Max Planck Institute for Biological Cybernetics December 2005 Outline 1 Introduction 2 Generative models 3 Low density separation 4 Graph
More information08 An Introduction to Dense Continuous Robotic Mapping
NAVARCH/EECS 568, ROB 530 - Winter 2018 08 An Introduction to Dense Continuous Robotic Mapping Maani Ghaffari March 14, 2018 Previously: Occupancy Grid Maps Pose SLAM graph and its associated dense occupancy
More informationImproving Image Segmentation Quality Via Graph Theory
International Symposium on Computers & Informatics (ISCI 05) Improving Image Segmentation Quality Via Graph Theory Xiangxiang Li, Songhao Zhu School of Automatic, Nanjing University of Post and Telecommunications,
More informationMachine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme, Nicolas Schilling
Machine Learning B. Unsupervised Learning B.1 Cluster Analysis Lars Schmidt-Thieme, Nicolas Schilling Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University of Hildesheim,
More informationRegularization and Markov Random Fields (MRF) CS 664 Spring 2008
Regularization and Markov Random Fields (MRF) CS 664 Spring 2008 Regularization in Low Level Vision Low level vision problems concerned with estimating some quantity at each pixel Visual motion (u(x,y),v(x,y))
More informationLearning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009
Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer
More informationCost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling
[DOI: 10.2197/ipsjtcva.7.99] Express Paper Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling Takayoshi Yamashita 1,a) Takaya Nakamura 1 Hiroshi Fukui 1,b) Yuji
More information3 October, 2013 MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos
Machine Learning for Computer Vision 1 3 October, 2013 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos Iasonas.kokkinos@ecp.fr Department of Applied Mathematics Ecole Centrale
More informationMRFs and Segmentation with Graph Cuts
02/24/10 MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Today s class Finish up EM MRFs w ij i Segmentation with Graph Cuts j EM Algorithm: Recap
More informationD-Separation. b) the arrows meet head-to-head at the node, and neither the node, nor any of its descendants, are in the set C.
D-Separation Say: A, B, and C are non-intersecting subsets of nodes in a directed graph. A path from A to B is blocked by C if it contains a node such that either a) the arrows on the path meet either
More informationClustering and The Expectation-Maximization Algorithm
Clustering and The Expectation-Maximization Algorithm Unsupervised Learning Marek Petrik 3/7 Some of the figures in this presentation are taken from An Introduction to Statistical Learning, with applications
More informationLearning Objects and Parts in Images
Learning Objects and Parts in Images Chris Williams E H U N I V E R S I T T Y O H F R G E D I N B U School of Informatics, University of Edinburgh, UK Learning multiple objects and parts from images (joint
More informationAlgorithms for Markov Random Fields in Computer Vision
Algorithms for Markov Random Fields in Computer Vision Dan Huttenlocher November, 2003 (Joint work with Pedro Felzenszwalb) Random Field Broadly applicable stochastic model Collection of n sites S Hidden
More informationFUSION OF MULTITEMPORAL AND MULTIRESOLUTION REMOTE SENSING DATA AND APPLICATION TO NATURAL DISASTERS
FUSION OF MULTITEMPORAL AND MULTIRESOLUTION REMOTE SENSING DATA AND APPLICATION TO NATURAL DISASTERS Ihsen HEDHLI, Josiane ZERUBIA INRIA Sophia Antipolis Méditerranée (France), Ayin team, in collaboration
More informationEfficient Particle Filter-Based Tracking of Multiple Interacting Targets Using an MRF-based Motion Model
Efficient Particle Filter-Based Tracking of Multiple Interacting Targets Using an MRF-based Motion Model Zia Khan, Tucker Balch, and Frank Dellaert {zkhan,tucker,dellaert}@cc.gatech.edu College of Computing,
More informationHuman Head-Shoulder Segmentation
Human Head-Shoulder Segmentation Hai Xin, Haizhou Ai Computer Science and Technology Tsinghua University Beijing, China ahz@mail.tsinghua.edu.cn Hui Chao, Daniel Tretter Hewlett-Packard Labs 1501 Page
More informationEfficient Feature Learning Using Perturb-and-MAP
Efficient Feature Learning Using Perturb-and-MAP Ke Li, Kevin Swersky, Richard Zemel Dept. of Computer Science, University of Toronto {keli,kswersky,zemel}@cs.toronto.edu Abstract Perturb-and-MAP [1] is
More informationintro, applications MRF, labeling... how it can be computed at all? Applications in segmentation: GraphCut, GrabCut, demos
Image as Markov Random Field and Applications 1 Tomáš Svoboda, svoboda@cmp.felk.cvut.cz Czech Technical University in Prague, Center for Machine Perception http://cmp.felk.cvut.cz Talk Outline Last update:
More informationHuman Upper Body Pose Estimation in Static Images
1. Research Team Human Upper Body Pose Estimation in Static Images Project Leader: Graduate Students: Prof. Isaac Cohen, Computer Science Mun Wai Lee 2. Statement of Project Goals This goal of this project
More informationObject Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision
Object Recognition Using Pictorial Structures Daniel Huttenlocher Computer Science Department Joint work with Pedro Felzenszwalb, MIT AI Lab In This Talk Object recognition in computer vision Brief definition
More informationParticle Filtering. CS6240 Multimedia Analysis. Leow Wee Kheng. Department of Computer Science School of Computing National University of Singapore
Particle Filtering CS6240 Multimedia Analysis Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore (CS6240) Particle Filtering 1 / 28 Introduction Introduction
More informationSelf Lane Assignment Using Smart Mobile Camera For Intelligent GPS Navigation and Traffic Interpretation
For Intelligent GPS Navigation and Traffic Interpretation Tianshi Gao Stanford University tianshig@stanford.edu 1. Introduction Imagine that you are driving on the highway at 70 mph and trying to figure
More informationCS 534: Computer Vision Segmentation and Perceptual Grouping
CS 534: Computer Vision Segmentation and Perceptual Grouping Ahmed Elgammal Dept of Computer Science CS 534 Segmentation - 1 Outlines Mid-level vision What is segmentation Perceptual Grouping Segmentation
More informationComputer Vision Group Prof. Daniel Cremers. 4a. Inference in Graphical Models
Group Prof. Daniel Cremers 4a. Inference in Graphical Models Inference on a Chain (Rep.) The first values of µ α and µ β are: The partition function can be computed at any node: Overall, we have O(NK 2
More informationClustering web search results
Clustering K-means Machine Learning CSE546 Emily Fox University of Washington November 4, 2013 1 Clustering images Set of Images [Goldberger et al.] 2 1 Clustering web search results 3 Some Data 4 2 K-means
More informationMachine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves
Machine Learning A 708.064 11W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence I [2 P] a) [1 P] Give an example for a probability distribution P (A, B, C) that disproves
More informationGraphical Models, Bayesian Method, Sampling, and Variational Inference
Graphical Models, Bayesian Method, Sampling, and Variational Inference With Application in Function MRI Analysis and Other Imaging Problems Wei Liu Scientific Computing and Imaging Institute University
More informationAn efficient face recognition algorithm based on multi-kernel regularization learning
Acta Technica 61, No. 4A/2016, 75 84 c 2017 Institute of Thermomechanics CAS, v.v.i. An efficient face recognition algorithm based on multi-kernel regularization learning Bi Rongrong 1 Abstract. A novel
More informationContent-based image and video analysis. Machine learning
Content-based image and video analysis Machine learning for multimedia retrieval 04.05.2009 What is machine learning? Some problems are very hard to solve by writing a computer program by hand Almost all
More informationECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov
ECE521: Week 11, Lecture 20 27 March 2017: HMM learning/inference With thanks to Russ Salakhutdinov Examples of other perspectives Murphy 17.4 End of Russell & Norvig 15.2 (Artificial Intelligence: A Modern
More informationMixture Models and EM
Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering
More informationMixture Models and EM
Table of Content Chapter 9 Mixture Models and EM -means Clustering Gaussian Mixture Models (GMM) Expectation Maximiation (EM) for Mixture Parameter Estimation Introduction Mixture models allows Complex
More informationVisual Motion Analysis and Tracking Part II
Visual Motion Analysis and Tracking Part II David J Fleet and Allan D Jepson CIAR NCAP Summer School July 12-16, 16, 2005 Outline Optical Flow and Tracking: Optical flow estimation (robust, iterative refinement,
More informationDetection of Man-made Structures in Natural Images
Detection of Man-made Structures in Natural Images Tim Rees December 17, 2004 Abstract Object detection in images is a very active research topic in many disciplines. Probabilistic methods have been applied
More informationIterated Graph Cuts for Image Segmentation
Iterated Graph Cuts for Image Segmentation Bo Peng 1, Lei Zhang 1, and Jian Yang 2 1 Department of Computing, The Hong Kong Polytechnic University, Kowloon, Hong Kong, China. 2 School of Computer Science
More informationIntroduction to Pattern Recognition Part II. Selim Aksoy Bilkent University Department of Computer Engineering
Introduction to Pattern Recognition Part II Selim Aksoy Bilkent University Department of Computer Engineering saksoy@cs.bilkent.edu.tr RETINA Pattern Recognition Tutorial, Summer 2005 Overview Statistical
More informationRandom projection for non-gaussian mixture models
Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,
More informationCS 664 Segmentation. Daniel Huttenlocher
CS 664 Segmentation Daniel Huttenlocher Grouping Perceptual Organization Structural relationships between tokens Parallelism, symmetry, alignment Similarity of token properties Often strong psychophysical
More informationMachine Learning. Chao Lan
Machine Learning Chao Lan Machine Learning Prediction Models Regression Model - linear regression (least square, ridge regression, Lasso) Classification Model - naive Bayes, logistic regression, Gaussian
More informationPROBABILISTIC IMAGE SEGMENTATION FOR LOW-LEVEL MAP BUILDING IN ROBOT NAVIGATION. Timo Kostiainen, Jouko Lampinen
PROBABILISTIC IMAGE SEGMENTATION FOR LOW-LEVEL MAP BUILDING IN ROBOT NAVIGATION Timo Kostiainen, Jouko Lampinen Helsinki University of Technology Laboratory of Computational Engineering P.O.Box 9203 FIN-02015,
More informationProbabilistic Index Histogram for Robust Object Tracking
Probabilistic Index Histogram for Robust Object Tracking Wei Li 1, Xiaoqin Zhang 2, Nianhua Xie 1, Weiming Hu 1, Wenhan Luo 1, Haibin Ling 3 1 National Lab of Pattern Recognition, Institute of Automation,CAS,
More informationSegmentation with non-linear constraints on appearance, complexity, and geometry
IPAM February 2013 Western Univesity Segmentation with non-linear constraints on appearance, complexity, and geometry Yuri Boykov Andrew Delong Lena Gorelick Hossam Isack Anton Osokin Frank Schmidt Olga
More informationPSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D.
PSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D. Rhodes 5/10/17 What is Machine Learning? Machine learning
More informationProbabilistic Graphical Models
Overview of Part Two Probabilistic Graphical Models Part Two: Inference and Learning Christopher M. Bishop Exact inference and the junction tree MCMC Variational methods and EM Example General variational
More informationA Hierarchical Compositional System for Rapid Object Detection
A Hierarchical Compositional System for Rapid Object Detection Long Zhu and Alan Yuille Department of Statistics University of California at Los Angeles Los Angeles, CA 90095 {lzhu,yuille}@stat.ucla.edu
More informationCS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas
CS839: Probabilistic Graphical Models Lecture 10: Learning with Partially Observed Data Theo Rekatsinas 1 Partially Observed GMs Speech recognition 2 Partially Observed GMs Evolution 3 Partially Observed
More informationA Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data
A Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data Seyoung Kim, Padhraic Smyth, and Hal Stern Bren School of Information and Computer Sciences University of California,
More informationMulti-Instance Hidden Markov Model For Facial Expression Recognition
Multi-Instance Hidden Markov Model For Facial Expression Recognition Chongliang Wu 1, Shangfei Wang 1 and Qiang Ji 2 1 School of Computer Science and Technology, University of Science and Technology of
More informationSimultaneous Appearance Modeling and Segmentation for Matching People under Occlusion
Simultaneous Appearance Modeling and Segmentation for Matching People under Occlusion Zhe Lin, Larry S. Davis, David Doermann, and Daniel DeMenthon Institute for Advanced Computer Studies University of
More informationRecovering Intrinsic Images from a Single Image
Recovering Intrinsic Images from a Single Image Marshall F Tappen William T Freeman Edward H Adelson MIT Artificial Intelligence Laboratory Cambridge, MA 02139 mtappen@ai.mit.edu, wtf@ai.mit.edu, adelson@ai.mit.edu
More informationLecture 21 : A Hybrid: Deep Learning and Graphical Models
10-708: Probabilistic Graphical Models, Spring 2018 Lecture 21 : A Hybrid: Deep Learning and Graphical Models Lecturer: Kayhan Batmanghelich Scribes: Paul Liang, Anirudha Rayasam 1 Introduction and Motivation
More informationGeneric Face Alignment Using an Improved Active Shape Model
Generic Face Alignment Using an Improved Active Shape Model Liting Wang, Xiaoqing Ding, Chi Fang Electronic Engineering Department, Tsinghua University, Beijing, China {wanglt, dxq, fangchi} @ocrserv.ee.tsinghua.edu.cn
More informationLocally Adaptive Learning for Translation-Variant MRF Image Priors
Locally Adaptive Learning for Translation-Variant MRF Image Priors Masayuki Tanaka and Masatoshi Okutomi Tokyo Institute of Technology 2-12-1 O-okayama, Meguro-ku, Tokyo, JAPAN mtanaka@ok.ctrl.titech.ac.p,
More informationEstimating Labels from Label Proportions
Estimating Labels from Label Proportions Novi Quadrianto Novi.Quad@gmail.com The Australian National University, Australia NICTA, Statistical Machine Learning Program, Australia Joint work with Alex Smola,
More informationMarkov/Conditional Random Fields, Graph Cut, and applications in Computer Vision
Markov/Conditional Random Fields, Graph Cut, and applications in Computer Vision Fuxin Li Slides and materials from Le Song, Tucker Hermans, Pawan Kumar, Carsten Rother, Peter Orchard, and others Recap:
More informationMarkov Random Fields for Recognizing Textures modeled by Feature Vectors
Markov Random Fields for Recognizing Textures modeled by Feature Vectors Juliette Blanchet, Florence Forbes, and Cordelia Schmid Inria Rhône-Alpes, 655 avenue de l Europe, Montbonnot, 38334 Saint Ismier
More informationOverview Citation. ML Introduction. Overview Schedule. ML Intro Dataset. Introduction to Semi-Supervised Learning Review 10/4/2010
INFORMATICS SEMINAR SEPT. 27 & OCT. 4, 2010 Introduction to Semi-Supervised Learning Review 2 Overview Citation X. Zhu and A.B. Goldberg, Introduction to Semi- Supervised Learning, Morgan & Claypool Publishers,
More informationLesion Segmentation and Bias Correction in Breast Ultrasound B-mode Images Including Elastography Information
Lesion Segmentation and Bias Correction in Breast Ultrasound B-mode Images Including Elastography Information Gerard Pons a, Joan Martí a, Robert Martí a, Mariano Cabezas a, Andrew di Battista b, and J.
More informationNon-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines
Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2007 c 2007,
More information