Markov Networks in Computer Vision

Similar documents
Markov Networks in Computer Vision. Sargur Srihari

Models for grids. Computer vision: models, learning and inference. Multi label Denoising. Binary Denoising. Denoising Goal.

Segmentation. Bottom up Segmentation Semantic Segmentation

Regularization and Markov Random Fields (MRF) CS 664 Spring 2008

Algorithms for Markov Random Fields in Computer Vision

Undirected Graphical Models. Raul Queiroz Feitosa

Markov/Conditional Random Fields, Graph Cut, and applications in Computer Vision

Graphical Models for Computer Vision

CS 664 Segmentation. Daniel Huttenlocher

Image Restoration using Markov Random Fields

CS6670: Computer Vision

SPM-BP: Sped-up PatchMatch Belief Propagation for Continuous MRFs. Yu Li, Dongbo Min, Michael S. Brown, Minh N. Do, Jiangbo Lu

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Markov Network Structure Learning

Collective classification in network data

D-Separation. b) the arrows meet head-to-head at the node, and neither the node, nor any of its descendants, are in the set C.

Markov Random Fields and Segmentation with Graph Cuts

Multi-view Stereo. Ivo Boyadzhiev CS7670: September 13, 2011

Machine Problem 8 - Mean Field Inference on Boltzman Machine

Seminar Heidelberg University

Analysis: TextonBoost and Semantic Texton Forests. Daniel Munoz Februrary 9, 2009

Markov Random Fields and Gibbs Sampling for Image Denoising

Robotics Programming Laboratory

3 : Representation of Undirected GMs

Computer Vision Group Prof. Daniel Cremers. 4a. Inference in Graphical Models

Semantic Segmentation. Zhongang Qi

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Statistical and Learning Techniques in Computer Vision Lecture 1: Markov Random Fields Jens Rittscher and Chuck Stewart

Geometric Registration for Deformable Shapes 3.3 Advanced Global Matching

10708 Graphical Models: Homework 4

Multiple cosegmentation

Topics to be Covered in the Rest of the Semester. CSci 4968 and 6270 Computational Vision Lecture 15 Overview of Remainder of the Semester

Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction

Learning Undirected Models with Missing Data

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Object Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision

EE795: Computer Vision and Intelligent Systems

Digital Image Processing Laboratory: MAP Image Restoration

Contextual Classification with Functional Max-Margin Markov Networks

MR IMAGE SEGMENTATION

Multi-Class Segmentation with Relative Location Prior

Comparison of Graph Cuts with Belief Propagation for Stereo, using Identical MRF Parameters

Problem 1: Complexity of Update Rules for Logistic Regression

Bayesian Methods in Vision: MAP Estimation, MRFs, Optimization

A MULTI-RESOLUTION APPROACH TO DEPTH FIELD ESTIMATION IN DENSE IMAGE ARRAYS F. Battisti, M. Brizzi, M. Carli, A. Neri

Discrete Optimization Algorithms in Computer Vision

Generative and discriminative classification techniques

Last update: May 4, Vision. CMSC 421: Chapter 24. CMSC 421: Chapter 24 1

CRFs for Image Classification

Fast Approximate Energy Minimization via Graph Cuts

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?

What is Computer Vision?

Motion Estimation (II) Ce Liu Microsoft Research New England

Nearest Neighbor with KD Trees

Graphical Models, Bayesian Method, Sampling, and Variational Inference

Parallel constraint optimization algorithms for higher-order discrete graphical models: applications in hyperspectral imaging

Dr. Ulas Bagci

Region-based Segmentation and Object Detection

Quadratic Pseudo-Boolean Optimization(QPBO): Theory and Applications At-a-Glance

Image analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis

intro, applications MRF, labeling... how it can be computed at all? Applications in segmentation: GraphCut, GrabCut, demos

GRID WARPING IN TOTAL VARIATION IMAGE ENHANCEMENT METHODS. Andrey Nasonov, and Andrey Krylov

Visual Recognition: Examples of Graphical Models

MRF-based Algorithms for Segmentation of SAR Images

Part 5: Structured Support Vector Machines

Convex Optimization MLSS 2015

Image Segmentation continued Graph Based Methods. Some slides: courtesy of O. Capms, Penn State, J.Ponce and D. Fortsyth, Computer Vision Book

Object Extraction Using Image Segmentation and Adaptive Constraint Propagation

Digital Image Processing Laboratory: Markov Random Fields and MAP Image Segmentation

(University Improving of Montreal) Generative Adversarial Networks with Denoising Feature Matching / 17

EE 701 ROBOT VISION. Segmentation

Clustering. Mihaela van der Schaar. January 27, Department of Engineering Science University of Oxford

Super Resolution Using Graph-cut

6.801/866. Segmentation and Line Fitting. T. Darrell

Data Analysis 3. Support Vector Machines. Jan Platoš October 30, 2017

Data-driven Depth Inference from a Single Still Image

What have we leaned so far?

Lecture 13 Segmentation and Scene Understanding Chris Choy, Ph.D. candidate Stanford Vision and Learning Lab (SVL)

Computer Vision Group Prof. Daniel Cremers. 4. Probabilistic Graphical Models Directed Models

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms for Inference Fall 2014

Approximate Parameter Learning in Conditional Random Fields: An Empirical Investigation

CS 664 Slides #11 Image Segmentation. Prof. Dan Huttenlocher Fall 2003

Project Updates Short lecture Volumetric Modeling +2 papers

Computer Vision Group Prof. Daniel Cremers. 4. Probabilistic Graphical Models Directed Models

Discrete Optimization Methods in Computer Vision CSE 6389 Slides by: Boykov Modified and Presented by: Mostafa Parchami Basic overview of graph cuts

Combinatorial optimization and its applications in image Processing. Filip Malmberg

The Surprising Power of Belief Propagation

Interactive segmentation, Combinatorial optimization. Filip Malmberg

Generative and discriminative classification techniques

STRUCTURAL EDGE LEARNING FOR 3-D RECONSTRUCTION FROM A SINGLE STILL IMAGE. Nan Hu. Stanford University Electrical Engineering

Input. Output. Problem Definition. Rectified stereo image pair All correspondences lie in same scan lines

Deep Generative Models Variational Autoencoders

Convexization in Markov Chain Monte Carlo

Detection of Man-made Structures in Natural Images

Motion Estimation. There are three main types (or applications) of motion estimation:

Bilevel Sparse Coding

CS 534: Computer Vision Segmentation and Perceptual Grouping

Segmentation & Clustering

Probabilistic Graphical Models

Segmentation: Clustering, Graph Cut and EM

Transcription:

Markov Networks in Computer Vision Sargur Srihari srihari@cedar.buffalo.edu 1

Markov Networks for Computer Vision Some applications: 1. Image segmentation 2. Removal of blur/noise 3. Stereo reconstruction 4. Object recognition Typically called MRFs in vision community 2

1. Image Segmentation Task Partition the image pixels into regions corresponding to distinct parts of scene Different variants of segmentation task Many formulated as a Markov network Multiclass segmentation Each variable X i has a domain {1,..,K} pixels Value of X i represents region assignment for pixel i, e.g., grass, water, sky, car Classifying each pixel is expensive Oversegment image into superpixels (coherent regions) and classify each superpixels All pixels within region are assigned same value 3

Variables in Computer Vision X i : Pixels or Super-pixels Joint probability distribution over an image Log-linear model: P(X 1,..X n ;θ) = 1 k Z(θ) exp θ f (D ) i i i i=1 ( ) = exp θ i f i ( ξ ) Z θ ξ i F={f i }: Features between variables A,B ε D i : f a (a,b) = I { a = a 0 0 b }I b = b 0 0 { } 4

Network Structure In most applications structure is pairwise Variables correspond to pixels Edges (factors) correspond to interactions between adjacent pixels in grid on image Each interior pixel has exactly four neighbors Value space of variables and exact form of factors depend on task Usually formulated: Factors in terms of energies Negative log potentials Values represent penalties:» lower value = higher probability 5

Three Examples from Computer Vision 1. Image Segmentation Partition image pixels into regions 2. Image Denoising Restore true value of all pixels 3. Stereo Reconstruction Reconstruct depth disparity of each pixel 6

Model Edge potential between every pair of superpixels X i, X j Encodes a contiguity preference With a penalty λ whenever X i X j Model can be im[roved by making penalty depend on presence of an image gradient between pixels Even better model: Non default values for class pairs Tigers adjacent to vegetation, water below vegetation 7

Graph from Superpixels A node for each superpixel Edge between nodes if regions are adjacent This defines a distribution in terms of this graph 8

Features for Image Segmentation Features extracted for each superpixel Statistics over color, texture, location Features either clustered or input to local classifiers to reduce dimensionality Node potential is a function of these features Factors depend upon pixels in image Each image defines a different probability distribution over segment labels for pixels or superpixels Model in effect is a Conditional Random Field 9

Importance of Modeling Correlations between superpixels building car road cow Original image Oversegmented image-superpixels Each superpixel is a random variable Classification using node potentials alone-each superpixel classified independently grass (a) (b) (c) Segmentation (d) using pairwise Markov Network encoding interactions between adjacent 10 superpixels

Metric MRFs Class of MRFs used for labeling Graph of nodes X 1,..X n related by set of edges E Wish to assign to each X i a label in space V={v 1,..v k } Each node, taken in isolation, has its preference among possible labels Also need to impose a soft smoothness constraint that neighboring nodes should take similar values 11

Encoding preferences Node preferences are node potentials in pairwise MRF Smoothness preferences are edge potentials Traditional to encode these models in negative log-space using energy functions With MAP objective we can ignore the partition function 12

Energy Function Energy function E(x 1,..x n ) = ε i (x i ) + ε ij (x i x j ) i {i,j } Goal is to minimize the energy arg min E(x x 1,..x 1,..x n ) n 13

Smoothness definition Slight variant of Ising model ε ij (x 1,x j ) = 0 λ i,j x i = x j x i x j Obtain lowest possible pairwise energy (0) when neighboring nodes X i,x j take the same value and a higher energy λ i,j when they do not 14

Generalizations Potts model extends it to more than two labels Distance function on labels Prefer neighboring nodes to have labels smaller distance apart 15

Metric definition A function µ: V Và [0, ) is a metric if it satisfies Reflexivity: µ(v k,v l )=0 if and only if k=l Symmetry:µ(v k,v l )=µ(v l,v k ); Triangle Inequality:µ(v k,v l )+µ(v l,v m ) µ(v k,v m ) µ is a semi-metric if it satisfies first two Metric MRF is defined by defining ε i,j (v k,v j ) = µ(v k,v l ) A common metric: ε(x i,x j ) = min(c x i -x j, dist max ) 16

2. Image denoising Task: Restore true value given noisy pixel values Node potential ϕ(x i ) for each pixel X i penalize large discrepancies from observed pixel value y i Edge potential Encode continuity between adjacent pixel values Penalize cases where inferred value of X i is too far from inferred value of neighbor X j Important not to over-penalize true edge disparities (edges between objects or regions) Leads to oversmoothing of image Solution: Bound the penalty using a truncated norm ε(x i,x j ) = min(c x i -x j p, dist max ) for p ε {1,2} 17

Binary Image de-noising Noise removal from binary image Observed noisy image Binary pixel values y i {-1,+1}, i=1,..,d Unknown noise-free image Binary pixel values x i {-1,+1}, i=1,..,d Noisy image assumed to randomly flip sign of pixels with small probability 18

Markov Random Field Model x i = unknown noise-free pixel y i = known noisy pixel Known Strong correlation between x i and y i Neighbor pixels x i and x j are strongly correlated Prior knowledge captured using MRF 19

Energy Functions Graph has two types of cliques {x i,y i } expresses correlation between variables Choose simple energy function η x i y i Lower energy (higher probability) when x i and y i have same sign {x i,x j } which are neighboring pixels Choose β x i x j 20

Potential Function Complete energy function of model E(x, y) = h xi β xx i j η xiyi i {, i j} i The hx i term biases towards pixel values that have one particular sign Which defines a joint distribution over x and y given by 1 p(x,y) = exp{ E(x,y)} Z 21

De-noising problem statement We fix y to observed pixels in the noisy image p(x y) is a conditional distribution over all noise-free images Called Ising model in statistical physics We wish to find an image x that has a high probability 22

De-noising algorithm Gradient ascent Set x i =y i for all i Take one node x j at a time evaluate total energy for states x i =+1 and x i =-1 keeping all other node variable fixed Set x j to value for which energy is lower This is a local computation which can be done efficiently Repeat for all pixels until a stopping criterion is met Nodes updated systematically by raster scan or randomly Finds a local maximum (which need not be global) Algorithm is called Iterated Conditional Modes (ICM) 23

Image Restoration Results Parameters β=1.0, η=2.1, h=0 Noise Free image Noisy image where 10% of pixels are corrupted Result of ICM Global maximum obtained by Graph Cut algorithm 24

3. Stereo Reconstruction Reconstruct depth disparity of each pixel in the image Variables represent discretized version of depth dimension (more finely for discretized for close to camera and coarse when away) Node potential: a computer vision technique to estimate depth disparity Edge potential: a truncated metric Inversely proportional to image gradient between pixels Smaller penalty to large gradient suggesting occlusion 25