Parallel constraint optimization algorithms for higher-order discrete graphical models: applications in hyperspectral imaging
|
|
- Roderick Goodwin
- 5 years ago
- Views:
Transcription
1 Parallel constraint optimization algorithms for higher-order discrete graphical models: applications in hyperspectral imaging MSc Billy Braithwaite Supervisors: Prof. Pekka Neittaanmäki and Phd Ilkka Pölönen Department of Mathematical Information Technology Spectral Imaging Laboratory (SLI) University of Jyväskylä 26 April 2016
2 Introduction: Background BSc and MSc in Computer Science from University of Eastern Finland (UEF) in 2012 and 2015 respectively. Research assistant ( ): Speaker Evaluation ( ); Computer vision ( ). Machine learning, pattern recognition, computer vision, algorithms. Bachelor s and Master s thesis on metric indexing (similarity search). PhD studies at JYU, MIT, Spectral Imaging Laboratory November 2015.
3 Current PhD research proposal Designing and analyzing parallel algorithms for structured prediction using higher-order discrete probabilistic graphical models. Solving higher-order graphical models viewed as solving (valued) constraint satisfaction problems. Proximal methods used as solvers. Modeling and analysis of hyperspectral images using discrete probabilistic graphical models.
4 Primer on graphical models A graphical model is a realization of state spaces of individual vertices in a graph. Definition Let G = (V, E) be a graph, where V and E are sets of vertices and edges respectively. Also let i V be vertices, and let X = (X i ) i V be random joint variables and x = (x i ) i V X = i V X i be the realization of X. Then the graphical is the tuple (G, X). Two (main) types of graphical models: Markov Random Fields and Bayesian Belief Networks. Additional types: Factor Graphs and Coniditional Random Fields.
5 Markov Random Fields Markov Random Fields (MRFs) is an undirected graphical model, satisfying the following conditions: Local Markov property: i V, X i X j X c C, where c C a set of cliques of G. p(x) = 1 Z θ c (x c ). (1) c C Global Markov property: If V 1, V 2, V 3 V there exists a path from i V 1 to j V 2 which includes at least one k V 3, then X V1 X V2 X V3.
6 Markov Random Fields Inferece in graphical models are posed as optimization problems: solving p(x) as a maximum a posteriori (MAP) problem: x = arg max x X p(x) (2) p(x) = 1 Z exp( E(x)), Z = i V exp( E(x i )), (3) E(x) = c C θ c (x c ). (4) Equation (2) is equivalent to x = arg min x X E(x). (5)
7 Factor Graphs A Factor Graph (FG) is a more unified representation of a MRF, which uses an additional vertices coined factor vertices. The associated joint probability of a FG is a product of its factors: p(x) = 1 Z θ f (x f ). (6) f F FGs are bipartite graphs which gives a clearer (and more explicit) picture of the underlyingh factorization of the distribution of the graphical model.
8 Discrete probabilistic graphical models in vision Markovian Factor Graph x 1 x 2 x 3 x 1 x 2 x 3 x 6 x 4 x 5 x 7 f 1,2,4 f 1,3,5 x 5 x 6 x 4 x 7 x 8 x 9 f 4,6,8 f 5,7,9 x 8 x 9
9 Vision models using graphical models Pairwise models: E(x) = θ i (x i ) + θ c (x c ), C 2. (7) i V c C w i,j (1 δ(x i x j )) θ c (x c ) = min{k i,j, x i x j } min{k i,j, (x i x j ) 2 } (Potts models) (truncated absolute distance) (truncated quadratic) Higher-order models: E(x) = θ i (x i ) + θ c (x c ), C > 2 (8) i V c C
10 Graph-cuts Discrete optimization methods based on the Ford-Fulkerson max-flow/min-cut algorithm: C(S, T ) = i S,j T,{i,j} E st c(i, j). The aim is to construct a directed graph, and perform the s t-cut by partitioning the graph into two disjoint sets. Can be applied to multilabeling scenarios. Exact optimization possible when θ c ( ) is submodular: f (S) + f (T ) f (S T ) + f (S T )
11 Belief propagation First proposed for tree structures. Computes the marginal distribution of a graphical model via max-product: p(x) = 1 θ i,j (x i, x j ) θ i (x i ) (9) Z {i,j} E i V Standard Belief propagation update rule: m i,j (x j ) α θ i,j (x i, x j )θ i (x i ) m k,i (x i ) x i k N i \j b i (x i ) αθ i (x i ) k N i m k,i (x i ) (10a) (10b)
12 Dual-methods The MAP inference reformulated as an integer linear programming problem: E(θ, τ) = θ, τ = θ i (a)τ i (a) + i V a X i {i,j} E (a,b) X i X j θ i,j (a, b)τ i,j (a, b) a X i τ i (a) = 1 i inv, a X τ τ G = i τ i,j (a, b) = τ j (b) {i, j} E, b X j, τ i (a) {0, 1} i V, a X i, τ i,j (a, b) {0, 1} {i, j} E, (a, b) X i X j τ τ G is minimized over a relaxed domain ˆτ G (local marginal polytope).
13 Valued constraint satisfaction problem (VCSP) Solving E( ) in equation (8) can be viewed as an instance of constraint programming. This is known as VCSP. Linear Programming relaxation of VSCP: θ S (x)µ S (x) min, (11a) S H x D S µ S (y) = µ i (x), i S H, x D, (11b) y D S y i =x µ i (x) = 1, i V, (11c) x D µ S (x), S H, x D S, (11d) µ i (x) 0, i V, x D. (11e)
14 Challenges graphical models: higher-order cliques Higher-order interactions offers better modeling of the structures within an image. Challenges arise when C > 2: E(x) = i V θ i (x i ) + c C θ c (x c ) (12) Decision making/optimization becomes intractable = NP-Completness.
15 Order-reduction of higher-order graphical models E(x) = i V θ i (x i ) + {i,j} E θ i,j + c S P n model: Let l L = {0, 1, 2,... k} be labels. { 0, if x i = l k, i c, θ c (x c ) = θ 1 c θα, else Robust P n model: θ c (x c ) = { N i (x c ) 1 Q γ max, if N i (x c Q, γ max, else, θ c (x c ). (13) where N i (x c ) = min k ( c n k (x c )) and γ max = c θα (θ 1 + θ 2 M(c)) and M(c) a quality measure.
16 Split Bregman iteration For solving l 1 -regularized problems. The general form: min u Φ(u) + H(u) (14) The Split Bregman iteration decouples the l 1 and l 2 portions of equation (14): min u,d d + H(u), such that d = Φ(u). (15) Convert Equation (15) into an unconstrained optimization problem: min u,d d + H(u) + λ 2 d Φ(u) 2 2. (16)
17 Split Bregman iteration The Bregman distance: D J p (u, v) = J(u) J(v) < p, u v >, (17) Two step iteration for constrained optimization: { uk+1 min u (D J p k (u, u k ) + λh(u)), µ > 0, p k+1 p k H(u k+1 ).
18 Split Bregman iteration Apply the Bregman distance and the above iteration scheme to equation (16): (u k+1, d k+1 ) min u (D J (u, u p k k, d, d k ) + λ 2 d Φ(u) ), µ > 0, pk+1 u pu k λ( Φ)T (Φ(u k+1 ) d k+1 ) pk+1 d pu k λ(d k+1 Φ(u k+1 )) The Split Bregman iteration { (u k+1, d k+1 ) min u,d d + H(u) + λ 2 d Φ(u) b k 2 2 b k+1 b k + (Φ(u k+1 ) d k+1 )
19 Parallel computing approaches Both multithread and GPU approaches considered. The Split Bregman Iteration (also proximal methods in general) are well-suited for parallel computing. For example using Domain Decomposition on Equation (5), each subproblem ( slave problem) can be solved independently with the Split Bregman Iteration via saddle-point formulation of the Lagrangian multipliers. Using parallel computing for computing promising clique configurations for a graphical model.
20 Application: Dense CRFs for image segmentation Give a random field X of random variables and a graph G with a discrete label set L = {l 1, l 2,..., l k }, the CRF (I, X) is characterized by a Gibbs distribution: p(x I ) = 1 Z exp( c C θ c(x I )). The corresponding Gibb s energy: E(x) = θ u(x i ) + θ p(x i, x j ) i V i<j Pairwise potentials: θ p(x i, x j ) = µ(x i, x j ) k(f i, f j ) = w 1 exp( p i p j 2 2θ 2 α K w m k m (f i, f j ), m=1 I i I j 2 2θ 2 α ) + w 2 exp( p i p j 2 ) 2θ 2 γ
21 Application: Dense CRFs for image segmentation (a) Input image. (b) Priors. (c) Output. (a) Input image. (b) Priors. (c) Output.
22 Applications in proposed PhD research Applying optimization and machine learning methods in bio-medical related computer vision applications. For example, texture analysis and segmentation. Present research plan is to apply hyperspectral imaging techniques to bio-medical imaging applications: Analyzing histology image samples from Central Hospital of Jyäskylä Modeling and simulation(?) of cell structures and diseases from cell cultivations obtained from FICAM Skin cancer prognosis/analysis from imaging samples from Helsinki University Central Hospital (HUS)
23 Current status 60/60 ECTS acquired during December of First journal article Quicker range-and-k-nn joins in metric spaces published in Used as an auxilary data structure for computing distance oralces. Second journal article Split Bregman solvers for higher-order random fields in progress. Estimated submission: June 2016 to International Journal of Computer Vision. Third jourrnal article Non-submodular function minimization via Split Bregman iterators in preparation.
Models for grids. Computer vision: models, learning and inference. Multi label Denoising. Binary Denoising. Denoising Goal.
Models for grids Computer vision: models, learning and inference Chapter 9 Graphical Models Consider models where one unknown world state at each pixel in the image takes the form of a grid. Loops in the
More informationMarkov Networks in Computer Vision. Sargur Srihari
Markov Networks in Computer Vision Sargur srihari@cedar.buffalo.edu 1 Markov Networks for Computer Vision Important application area for MNs 1. Image segmentation 2. Removal of blur/noise 3. Stereo reconstruction
More informationExpectation Propagation
Expectation Propagation Erik Sudderth 6.975 Week 11 Presentation November 20, 2002 Introduction Goal: Efficiently approximate intractable distributions Features of Expectation Propagation (EP): Deterministic,
More informationintro, applications MRF, labeling... how it can be computed at all? Applications in segmentation: GraphCut, GrabCut, demos
Image as Markov Random Field and Applications 1 Tomáš Svoboda, svoboda@cmp.felk.cvut.cz Czech Technical University in Prague, Center for Machine Perception http://cmp.felk.cvut.cz Talk Outline Last update:
More informationMarkov Networks in Computer Vision
Markov Networks in Computer Vision Sargur Srihari srihari@cedar.buffalo.edu 1 Markov Networks for Computer Vision Some applications: 1. Image segmentation 2. Removal of blur/noise 3. Stereo reconstruction
More information3 : Representation of Undirected GMs
0-708: Probabilistic Graphical Models 0-708, Spring 202 3 : Representation of Undirected GMs Lecturer: Eric P. Xing Scribes: Nicole Rafidi, Kirstin Early Last Time In the last lecture, we discussed directed
More informationComputer Vision Group Prof. Daniel Cremers. 4. Probabilistic Graphical Models Directed Models
Prof. Daniel Cremers 4. Probabilistic Graphical Models Directed Models The Bayes Filter (Rep.) (Bayes) (Markov) (Tot. prob.) (Markov) (Markov) 2 Graphical Representation (Rep.) We can describe the overall
More informationConditional Random Fields and beyond D A N I E L K H A S H A B I C S U I U C,
Conditional Random Fields and beyond D A N I E L K H A S H A B I C S 5 4 6 U I U C, 2 0 1 3 Outline Modeling Inference Training Applications Outline Modeling Problem definition Discriminative vs. Generative
More informationRegularization and Markov Random Fields (MRF) CS 664 Spring 2008
Regularization and Markov Random Fields (MRF) CS 664 Spring 2008 Regularization in Low Level Vision Low level vision problems concerned with estimating some quantity at each pixel Visual motion (u(x,y),v(x,y))
More informationComputer Vision Group Prof. Daniel Cremers. 4a. Inference in Graphical Models
Group Prof. Daniel Cremers 4a. Inference in Graphical Models Inference on a Chain (Rep.) The first values of µ α and µ β are: The partition function can be computed at any node: Overall, we have O(NK 2
More informationProbabilistic Graphical Models
School of Computer Science Probabilistic Graphical Models Theory of Variational Inference: Inner and Outer Approximation Eric Xing Lecture 14, February 29, 2016 Reading: W & J Book Chapters Eric Xing @
More informationD-Separation. b) the arrows meet head-to-head at the node, and neither the node, nor any of its descendants, are in the set C.
D-Separation Say: A, B, and C are non-intersecting subsets of nodes in a directed graph. A path from A to B is blocked by C if it contains a node such that either a) the arrows on the path meet either
More informationQuadratic Pseudo-Boolean Optimization(QPBO): Theory and Applications At-a-Glance
Quadratic Pseudo-Boolean Optimization(QPBO): Theory and Applications At-a-Glance Presented By: Ahmad Al-Kabbany Under the Supervision of: Prof.Eric Dubois 12 June 2012 Outline Introduction The limitations
More informationAlgorithms for Markov Random Fields in Computer Vision
Algorithms for Markov Random Fields in Computer Vision Dan Huttenlocher November, 2003 (Joint work with Pedro Felzenszwalb) Random Field Broadly applicable stochastic model Collection of n sites S Hidden
More informationComparison of Graph Cuts with Belief Propagation for Stereo, using Identical MRF Parameters
Comparison of Graph Cuts with Belief Propagation for Stereo, using Identical MRF Parameters Marshall F. Tappen William T. Freeman Computer Science and Artificial Intelligence Laboratory Massachusetts Institute
More informationUndirected Graphical Models. Raul Queiroz Feitosa
Undirected Graphical Models Raul Queiroz Feitosa Pros and Cons Advantages of UGMs over DGMs UGMs are more natural for some domains (e.g. context-dependent entities) Discriminative UGMs (CRF) are better
More information2. Graphical Models. Undirected pairwise graphical models. Factor graphs. Bayesian networks. Conversion between graphical models. Graphical Models 2-1
Graphical Models 2-1 2. Graphical Models Undirected pairwise graphical models Factor graphs Bayesian networks Conversion between graphical models Graphical Models 2-2 Graphical models Families of graphical
More informationInter and Intra-Modal Deformable Registration:
Inter and Intra-Modal Deformable Registration: Continuous Deformations Meet Efficient Optimal Linear Programming Ben Glocker 1,2, Nikos Komodakis 1,3, Nikos Paragios 1, Georgios Tziritas 3, Nassir Navab
More informationCollective classification in network data
1 / 50 Collective classification in network data Seminar on graphs, UCSB 2009 Outline 2 / 50 1 Problem 2 Methods Local methods Global methods 3 Experiments Outline 3 / 50 1 Problem 2 Methods Local methods
More informationECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning
ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning Topics Bayes Nets: Inference (Finish) Variable Elimination Graph-view of VE: Fill-edges, induced width
More informationStatistical and Learning Techniques in Computer Vision Lecture 1: Markov Random Fields Jens Rittscher and Chuck Stewart
Statistical and Learning Techniques in Computer Vision Lecture 1: Markov Random Fields Jens Rittscher and Chuck Stewart 1 Motivation Up to now we have considered distributions of a single random variable
More informationComputer Vision Group Prof. Daniel Cremers. 4. Probabilistic Graphical Models Directed Models
Prof. Daniel Cremers 4. Probabilistic Graphical Models Directed Models The Bayes Filter (Rep.) (Bayes) (Markov) (Tot. prob.) (Markov) (Markov) 2 Graphical Representation (Rep.) We can describe the overall
More informationObject Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision
Object Recognition Using Pictorial Structures Daniel Huttenlocher Computer Science Department Joint work with Pedro Felzenszwalb, MIT AI Lab In This Talk Object recognition in computer vision Brief definition
More informationMarkov/Conditional Random Fields, Graph Cut, and applications in Computer Vision
Markov/Conditional Random Fields, Graph Cut, and applications in Computer Vision Fuxin Li Slides and materials from Le Song, Tucker Hermans, Pawan Kumar, Carsten Rother, Peter Orchard, and others Recap:
More informationGraphical Models. Pradeep Ravikumar Department of Computer Science The University of Texas at Austin
Graphical Models Pradeep Ravikumar Department of Computer Science The University of Texas at Austin Useful References Graphical models, exponential families, and variational inference. M. J. Wainwright
More information2. Graphical Models. Undirected graphical models. Factor graphs. Bayesian networks. Conversion between graphical models. Graphical Models 2-1
Graphical Models 2-1 2. Graphical Models Undirected graphical models Factor graphs Bayesian networks Conversion between graphical models Graphical Models 2-2 Graphical models There are three families of
More informationCS242: Probabilistic Graphical Models Lecture 2B: Loopy Belief Propagation & Junction Trees
CS242: Probabilistic Graphical Models Lecture 2B: Loopy Belief Propagation & Junction Trees Professor Erik Sudderth Brown University Computer Science September 22, 2016 Some figures and materials courtesy
More informationCS242: Probabilistic Graphical Models Lecture 3: Factor Graphs & Variable Elimination
CS242: Probabilistic Graphical Models Lecture 3: Factor Graphs & Variable Elimination Instructor: Erik Sudderth Brown University Computer Science September 11, 2014 Some figures and materials courtesy
More informationApplication of MRF s to Segmentation
EE641 Digital Image Processing II: Purdue University VISE - November 14, 2012 1 Application of MRF s to Segmentation Topics to be covered: The Model Bayesian Estimation MAP Optimization Parameter Estimation
More informationAn Introduction to LP Relaxations for MAP Inference
An Introduction to LP Relaxations for MAP Inference Adrian Weller MLSALT4 Lecture Feb 27, 2017 With thanks to David Sontag (NYU) for use of some of his slides and illustrations For more information, see
More informationImage analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis
7 Computer Vision and Classification 413 / 458 Computer Vision and Classification The k-nearest-neighbor method The k-nearest-neighbor (knn) procedure has been used in data analysis and machine learning
More informationDetection of Man-made Structures in Natural Images
Detection of Man-made Structures in Natural Images Tim Rees December 17, 2004 Abstract Object detection in images is a very active research topic in many disciplines. Probabilistic methods have been applied
More informationDigital Image Processing Laboratory: Markov Random Fields and MAP Image Segmentation
Purdue University: Digital Image Processing Laboratories Digital Image Processing Laboratory: Markov Random Fields and MAP Image Segmentation December, 205 Introduction This laboratory explores the use
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Bayesian Networks Directed Acyclic Graph (DAG) Bayesian Networks General Factorization Bayesian Curve Fitting (1) Polynomial Bayesian
More informationDigital Image Processing Laboratory: MAP Image Restoration
Purdue University: Digital Image Processing Laboratories 1 Digital Image Processing Laboratory: MAP Image Restoration October, 015 1 Introduction This laboratory explores the use of maximum a posteriori
More informationInformation Processing Letters
Information Processing Letters 112 (2012) 449 456 Contents lists available at SciVerse ScienceDirect Information Processing Letters www.elsevier.com/locate/ipl Recursive sum product algorithm for generalized
More informationIMAGE FUSION WITH SIMULTANEOUS CARTOON AND TEXTURE DECOMPOSITION MAHDI DODANGEH, ISABEL NARRA FIGUEIREDO AND GIL GONÇALVES
Pré-Publicações do Departamento de Matemática Universidade de Coimbra Preprint Number 15 14 IMAGE FUSION WITH SIMULTANEOUS CARTOON AND TEXTURE DECOMPOSITION MAHDI DODANGEH, ISABEL NARRA FIGUEIREDO AND
More informationLecture 13 Segmentation and Scene Understanding Chris Choy, Ph.D. candidate Stanford Vision and Learning Lab (SVL)
Lecture 13 Segmentation and Scene Understanding Chris Choy, Ph.D. candidate Stanford Vision and Learning Lab (SVL) http://chrischoy.org Stanford CS231A 1 Understanding a Scene Objects Chairs, Cups, Tables,
More informationGraph Coloring via Constraint Programming-based Column Generation
Graph Coloring via Constraint Programming-based Column Generation Stefano Gualandi Federico Malucelli Dipartimento di Elettronica e Informatica, Politecnico di Milano Viale Ponzio 24/A, 20133, Milan, Italy
More informationThe Alternating Direction Method of Multipliers
The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein October 8, 2015 1 / 30 Introduction Presentation Outline 1 Convex
More informationDynamic Planar-Cuts: Efficient Computation of Min-Marginals for Outer-Planar Models
Dynamic Planar-Cuts: Efficient Computation of Min-Marginals for Outer-Planar Models Dhruv Batra 1 1 Carnegie Mellon University www.ece.cmu.edu/ dbatra Tsuhan Chen 2,1 2 Cornell University tsuhan@ece.cornell.edu
More informationSegmentation. Bottom up Segmentation Semantic Segmentation
Segmentation Bottom up Segmentation Semantic Segmentation Semantic Labeling of Street Scenes Ground Truth Labels 11 classes, almost all occur simultaneously, large changes in viewpoint, scale sky, road,
More informationBiomedical Image Analysis Using Markov Random Fields & Efficient Linear Programing
Biomedical Image Analysis Using Markov Random Fields & Efficient Linear Programing Nikos Komodakis Ahmed Besbes Ben Glocker Nikos Paragios Abstract Computer-aided diagnosis through biomedical image analysis
More informationSemantic Segmentation. Zhongang Qi
Semantic Segmentation Zhongang Qi qiz@oregonstate.edu Semantic Segmentation "Two men riding on a bike in front of a building on the road. And there is a car." Idea: recognizing, understanding what's in
More informationPart II. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS
Part II C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Converting Directed to Undirected Graphs (1) Converting Directed to Undirected Graphs (2) Add extra links between
More informationDiscrete Optimization Algorithms in Computer Vision
Discrete Optimization Algorithms in Computer Vision Pedro Felzenszwalb and Ramin Zabih Abstract Optimization is a powerful paradigm for expressing and solving problems in a wide range of areas, and has
More informationHigher-order models in Computer Vision
Chapter 1 Higher-order models in Computer Vision PUSHMEET KOHLI Machine Learning and Perception Microsoft Research Cambridge, UK Email: pkohli@microsoft.com CARSTEN ROTHER Machine Learning and Perception
More informationMassachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms for Inference Fall 2014
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms for Inference Fall 2014 1 Course Overview This course is about performing inference in complex
More informationA primal-dual framework for mixtures of regularizers
A primal-dual framework for mixtures of regularizers Baran Gözcü baran.goezcue@epfl.ch Laboratory for Information and Inference Systems (LIONS) École Polytechnique Fédérale de Lausanne (EPFL) Switzerland
More informationJoint optimization of segmentation and appearance models November Vicente, 26, Kolmogorov, / Rothe 14
Joint optimization of segmentation and appearance models Vicente, Kolmogorov, Rother Chris Russell November 26, 2009 Joint optimization of segmentation and appearance models November Vicente, 26, Kolmogorov,
More informationMeasuring Uncertainty in Graph Cut Solutions - Efficiently Computing Min-marginal Energies using Dynamic Graph Cuts
Measuring Uncertainty in Graph Cut Solutions - Efficiently Computing Min-marginal Energies using Dynamic Graph Cuts Pushmeet Kohli Philip H.S. Torr Department of Computing, Oxford Brookes University, Oxford.
More informationMarkov Random Fields and Gibbs Sampling for Image Denoising
Markov Random Fields and Gibbs Sampling for Image Denoising Chang Yue Electrical Engineering Stanford University changyue@stanfoed.edu Abstract This project applies Gibbs Sampling based on different Markov
More informationDecomposition of log-linear models
Graphical Models, Lecture 5, Michaelmas Term 2009 October 27, 2009 Generating class Dependence graph of log-linear model Conformal graphical models Factor graphs A density f factorizes w.r.t. A if there
More informationMathematics in Image Processing
Mathematics in Image Processing Michal Šorel Department of Image Processing Institute of Information Theory and Automation (ÚTIA) Academy of Sciences of the Czech Republic http://zoi.utia.cas.cz/ Mathematics
More informationChapter 8 of Bishop's Book: Graphical Models
Chapter 8 of Bishop's Book: Graphical Models Review of Probability Probability density over possible values of x Used to find probability of x falling in some range For continuous variables, the probability
More informationConditional Random Fields as Recurrent Neural Networks
BIL722 - Deep Learning for Computer Vision Conditional Random Fields as Recurrent Neural Networks S. Zheng, S. Jayasumana, B. Romera-Paredes V. Vineet, Z. Su, D. Du, C. Huang, P.H.S. Torr Introduction
More informationComplex Prediction Problems
Problems A novel approach to multiple Structured Output Prediction Max-Planck Institute ECML HLIE08 Information Extraction Extract structured information from unstructured data Typical subtasks Named Entity
More informationReduce, Reuse & Recycle: Efficiently Solving Multi-Label MRFs
Reduce, Reuse & Recycle: Efficiently Solving Multi-Label MRFs Karteek Alahari 1 Pushmeet Kohli 2 Philip H. S. Torr 1 1 Oxford Brookes University 2 Microsoft Research, Cambridge Abstract In this paper,
More informationMarkov Random Fields and Segmentation with Graph Cuts
Markov Random Fields and Segmentation with Graph Cuts Computer Vision Jia-Bin Huang, Virginia Tech Many slides from D. Hoiem Administrative stuffs Final project Proposal due Oct 27 (Thursday) HW 4 is out
More informationMesh segmentation. Florent Lafarge Inria Sophia Antipolis - Mediterranee
Mesh segmentation Florent Lafarge Inria Sophia Antipolis - Mediterranee Outline What is mesh segmentation? M = {V,E,F} is a mesh S is either V, E or F (usually F) A Segmentation is a set of sub-meshes
More informationMean Field and Variational Methods finishing off
Readings: K&F: 10.1, 10.5 Mean Field and Variational Methods finishing off Graphical Models 10708 Carlos Guestrin Carnegie Mellon University November 5 th, 2008 10-708 Carlos Guestrin 2006-2008 1 10-708
More informationGeometric Registration for Deformable Shapes 3.3 Advanced Global Matching
Geometric Registration for Deformable Shapes 3.3 Advanced Global Matching Correlated Correspondences [ASP*04] A Complete Registration System [HAW*08] In this session Advanced Global Matching Some practical
More information6.801/866. Segmentation and Line Fitting. T. Darrell
6.801/866 Segmentation and Line Fitting T. Darrell Segmentation and Line Fitting Gestalt grouping Background subtraction K-Means Graph cuts Hough transform Iterative fitting (Next time: Probabilistic segmentation)
More informationThe More the Merrier: Parameter Learning for Graphical Models with Multiple MAPs
The More the Merrier: Parameter Learning for Graphical Models with Multiple MAPs Franziska Meier fmeier@usc.edu U. of Southern California, 94 Bloom Walk, Los Angeles, CA 989 USA Amir Globerson amir.globerson@mail.huji.ac.il
More informationImage Restoration using Markov Random Fields
Image Restoration using Markov Random Fields Based on the paper Stochastic Relaxation, Gibbs Distributions and Bayesian Restoration of Images, PAMI, 1984, Geman and Geman. and the book Markov Random Field
More informationA discriminative view of MRF pre-processing algorithms supplementary material
A discriminative view of MRF pre-processing algorithms supplementary material Chen Wang 1,2 Charles Herrmann 2 Ramin Zabih 1,2 1 Google Research 2 Cornell University {chenwang, cih, rdz}@cs.cornell.edu
More informationComputer vision: models, learning and inference. Chapter 10 Graphical Models
Computer vision: models, learning and inference Chapter 10 Graphical Models Independence Two variables x 1 and x 2 are independent if their joint probability distribution factorizes as Pr(x 1, x 2 )=Pr(x
More information5/3/2010Z:\ jeh\self\notes.doc\7 Chapter 7 Graphical models and belief propagation Graphical models and belief propagation
//00Z:\ jeh\self\notes.doc\7 Chapter 7 Graphical models and belief propagation 7. Graphical models and belief propagation Outline graphical models Bayesian networks pair wise Markov random fields factor
More informationAnalysis: TextonBoost and Semantic Texton Forests. Daniel Munoz Februrary 9, 2009
Analysis: TextonBoost and Semantic Texton Forests Daniel Munoz 16-721 Februrary 9, 2009 Papers [shotton-eccv-06] J. Shotton, J. Winn, C. Rother, A. Criminisi, TextonBoost: Joint Appearance, Shape and Context
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Raquel Urtasun and Tamir Hazan TTI Chicago April 8, 2011 Raquel Urtasun and Tamir Hazan (TTI-C) Graphical Models April 8, 2011 1 / 19 Factor Graphs H does not reveal the
More informationA Randomized Algorithm for Minimizing User Disturbance Due to Changes in Cellular Technology
A Randomized Algorithm for Minimizing User Disturbance Due to Changes in Cellular Technology Carlos A. S. OLIVEIRA CAO Lab, Dept. of ISE, University of Florida Gainesville, FL 32611, USA David PAOLINI
More informationClustering: Classic Methods and Modern Views
Clustering: Classic Methods and Modern Views Marina Meilă University of Washington mmp@stat.washington.edu June 22, 2015 Lorentz Center Workshop on Clusters, Games and Axioms Outline Paradigms for clustering
More informationAN ANALYSIS ON MARKOV RANDOM FIELDS (MRFs) USING CYCLE GRAPHS
Volume 8 No. 0 208, -20 ISSN: 3-8080 (printed version); ISSN: 34-3395 (on-line version) url: http://www.ijpam.eu doi: 0.2732/ijpam.v8i0.54 ijpam.eu AN ANALYSIS ON MARKOV RANDOM FIELDS (MRFs) USING CYCLE
More informationA MRF Shape Prior for Facade Parsing with Occlusions Supplementary Material
A MRF Shape Prior for Facade Parsing with Occlusions Supplementary Material Mateusz Koziński, Raghudeep Gadde, Sergey Zagoruyko, Guillaume Obozinski and Renaud Marlet Université Paris-Est, LIGM (UMR CNRS
More informationECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning
ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning Topics Markov Random Fields: Inference Exact: VE Exact+Approximate: BP Readings: Barber 5 Dhruv Batra
More informationIntroduction to Machine Learning
Introduction to Machine Learning Maximum Margin Methods Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574
More informationFrom the Jungle to the Garden: Growing Trees for Markov Chain Monte Carlo Inference in Undirected Graphical Models
From the Jungle to the Garden: Growing Trees for Markov Chain Monte Carlo Inference in Undirected Graphical Models by Jean-Noël Rivasseau, M.Sc., Ecole Polytechnique, 2003 A THESIS SUBMITTED IN PARTIAL
More informationA Weighted Kernel PCA Approach to Graph-Based Image Segmentation
A Weighted Kernel PCA Approach to Graph-Based Image Segmentation Carlos Alzate Johan A. K. Suykens ESAT-SCD-SISTA Katholieke Universiteit Leuven Leuven, Belgium January 25, 2007 International Conference
More informationIntegrating Probabilistic Reasoning with Constraint Satisfaction
Integrating Probabilistic Reasoning with Constraint Satisfaction IJCAI Tutorial #7 Instructor: Eric I. Hsu July 17, 2011 http://www.cs.toronto.edu/~eihsu/tutorial7 Getting Started Discursive Remarks. Organizational
More informationTightening MRF Relaxations with Planar Subproblems
Tightening MRF Relaxations with Planar Subproblems Julian Yarkony, Ragib Morshed, Alexander T. Ihler, Charless C. Fowlkes Department of Computer Science University of California, Irvine {jyarkony,rmorshed,ihler,fowlkes}@ics.uci.edu
More information6 : Factor Graphs, Message Passing and Junction Trees
10-708: Probabilistic Graphical Models 10-708, Spring 2018 6 : Factor Graphs, Message Passing and Junction Trees Lecturer: Kayhan Batmanghelich Scribes: Sarthak Garg 1 Factor Graphs Factor Graphs are graphical
More informationShuheng Zhou. Annotated Bibliography
Shuheng Zhou Annotated Bibliography High-dimensional Statistical Inference S. Zhou, J. Lafferty and L. Wasserman, Compressed Regression, in Advances in Neural Information Processing Systems 20 (NIPS 2007).
More informationJunction tree propagation - BNDG 4-4.6
Junction tree propagation - BNDG 4-4. Finn V. Jensen and Thomas D. Nielsen Junction tree propagation p. 1/2 Exact Inference Message Passing in Join Trees More sophisticated inference technique; used in
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 19: Graph Cuts source S sink T Readings Szeliski, Chapter 11.2 11.5 Stereo results with window search problems in areas of uniform texture Window-based matching
More informationRevisiting MAP Estimation, Message Passing and Perfect Graphs
James R. Foulds Nicholas Navaroli Padhraic Smyth Alexander Ihler Department of Computer Science University of California, Irvine {jfoulds,nnavarol,smyth,ihler}@ics.uci.edu Abstract Given a graphical model,
More informationRandomized rounding of semidefinite programs and primal-dual method for integer linear programming. Reza Moosavi Dr. Saeedeh Parsaeefard Dec.
Randomized rounding of semidefinite programs and primal-dual method for integer linear programming Dr. Saeedeh Parsaeefard 1 2 3 4 Semidefinite Programming () 1 Integer Programming integer programming
More informationA Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems
A Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems Keely L. Croxton Fisher College of Business The Ohio State University Bernard Gendron Département
More informationContextual Classification with Functional Max-Margin Markov Networks
Contextual Classification with Functional Max-Margin Markov Networks Dan Munoz Nicolas Vandapel Drew Bagnell Martial Hebert Geometry Estimation (Hoiem et al.) Sky Problem 3-D Point Cloud Classification
More informationCS 534: Computer Vision Segmentation and Perceptual Grouping
CS 534: Computer Vision Segmentation and Perceptual Grouping Ahmed Elgammal Dept of Computer Science CS 534 Segmentation - 1 Outlines Mid-level vision What is segmentation Perceptual Grouping Segmentation
More informationEnergy Minimization for Segmentation in Computer Vision
S * = arg S min E(S) Energy Minimization for Segmentation in Computer Vision Meng Tang, Dmitrii Marin, Ismail Ben Ayed, Yuri Boykov Outline Clustering/segmentation methods K-means, GrabCut, Normalized
More informationA Principled Deep Random Field Model for Image Segmentation
A Principled Deep Random Field Model for Image Segmentation Pushmeet Kohli Microsoft Research Cambridge, UK Anton Osokin Moscow State University Moscow, Russia Stefanie Jegelka UC Berkeley Berkeley, CA,
More informationIntroduction to Optimization
Introduction to Optimization Second Order Optimization Methods Marc Toussaint U Stuttgart Planned Outline Gradient-based optimization (1st order methods) plain grad., steepest descent, conjugate grad.,
More informationPRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction
PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING KELLER VANDEBOGERT AND CHARLES LANNING 1. Introduction Interior point methods are, put simply, a technique of optimization where, given a problem
More informationModeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA
Modeling and Reasoning with Bayesian Networks Adnan Darwiche University of California Los Angeles, CA darwiche@cs.ucla.edu June 24, 2008 Contents Preface 1 1 Introduction 1 1.1 Automated Reasoning........................
More informationMachine Learning
Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University April 1, 2019 Today: Inference in graphical models Learning graphical models Readings: Bishop chapter 8 Bayesian
More informationAn LP View of the M-best MAP problem
An LP View of the M-best MAP problem Menachem Fromer Amir Globerson School of Computer Science and Engineering The Hebrew University of Jerusalem {fromer,gamir}@cs.huji.ac.il Abstract We consider the problem
More informationLoopy Belief Propagation
Loopy Belief Propagation Research Exam Kristin Branson September 29, 2003 Loopy Belief Propagation p.1/73 Problem Formalization Reasoning about any real-world problem requires assumptions about the structure
More informationOn Iteratively Constraining the Marginal Polytope for Approximate Inference and MAP
On Iteratively Constraining the Marginal Polytope for Approximate Inference and MAP David Sontag CSAIL Massachusetts Institute of Technology Cambridge, MA 02139 Tommi Jaakkola CSAIL Massachusetts Institute
More informationGraphical Models for Computer Vision
Graphical Models for Computer Vision Pedro F Felzenszwalb Brown University Joint work with Dan Huttenlocher, Joshua Schwartz, Ross Girshick, David McAllester, Deva Ramanan, Allie Shapiro, John Oberlin
More informationFinding Non-overlapping Clusters for Generalized Inference Over Graphical Models
1 Finding Non-overlapping Clusters for Generalized Inference Over Graphical Models Divyanshu Vats and José M. F. Moura arxiv:1107.4067v2 [stat.ml] 18 Mar 2012 Abstract Graphical models use graphs to compactly
More information