Markov Random Fields

Similar documents
Probabilistic Graphical Models

STAT 598L Probabilistic Graphical Models. Instructor: Sergey Kirshner. Exact Inference

Machine Learning Lecture 16

Lecture 5: Exact inference. Queries. Complexity of inference. Queries (continued) Bayesian networks can answer questions about the underlying

CS242: Probabilistic Graphical Models Lecture 3: Factor Graphs & Variable Elimination

Lecture 5: Exact inference

ALGORITHMS FOR DECISION SUPPORT

Probabilistic Graphical Models

These notes present some properties of chordal graphs, a set of undirected graphs that are important for undirected graphical models.

An Effective Upperbound on Treewidth Using Partial Fill-in of Separators

Ch9: Exact Inference: Variable Elimination. Shimi Salant, Barak Sternberg

Bayesian Networks and Decision Graphs

6 : Factor Graphs, Message Passing and Junction Trees

Lecture 4: Undirected Graphical Models

CS242: Probabilistic Graphical Models Lecture 2B: Loopy Belief Propagation & Junction Trees

Junction tree propagation - BNDG 4-4.6

Bayesian Networks Inference

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014

Part II. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS

Chapter 8 of Bishop's Book: Graphical Models

Learning decomposable models with a bounded clique size

Lecture 9: Undirected Graphical Models Machine Learning

Learning Bounded Treewidth Bayesian Networks

Computer Vision Group Prof. Daniel Cremers. 4. Probabilistic Graphical Models Directed Models

Probabilistic Graphical Models

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning

Computer Vision Group Prof. Daniel Cremers. 4. Probabilistic Graphical Models Directed Models

Treewidth: Preprocessing and Kernelization

Computer Vision Group Prof. Daniel Cremers. 4a. Inference in Graphical Models

Search Algorithms for Solving Queries on Graphical Models & the Importance of Pseudo-trees in their Complexity.

Belief propagation in a bucket-tree. Handouts, 275B Fall Rina Dechter. November 1, 2000

STA 4273H: Statistical Machine Learning

EE512 Graphical Models Fall 2009

2. Graphical Models. Undirected graphical models. Factor graphs. Bayesian networks. Conversion between graphical models. Graphical Models 2-1

Machine Learning. Sourangshu Bhattacharya

Part I: Sum Product Algorithm and (Loopy) Belief Propagation. What s wrong with VarElim. Forwards algorithm (filtering) Forwards-backwards algorithm

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS

Probabilistic Graphical Models

Our Graphs Become Larger

Finding Non-overlapping Clusters for Generalized Inference Over Graphical Models

Treewidth and graph minors

Treewidth. Kai Wang, Zheng Lu, and John Hicks. April 1 st, 2015

Introduction to Triangulated Graphs. Tandy Warnow

Loopy Belief Propagation

3 : Representation of Undirected GMs

FMA901F: Machine Learning Lecture 6: Graphical Models. Cristian Sminchisescu

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning

Recitation 4: Elimination algorithm, reconstituted graph, triangulation

Semi-Independent Partitioning: A Method for Bounding the Solution to COP s

Necessary edges in k-chordalizations of graphs

COS 513: Foundations of Probabilistic Modeling. Lecture 5

D-Separation. b) the arrows meet head-to-head at the node, and neither the node, nor any of its descendants, are in the set C.

CS 441 Discrete Mathematics for CS Lecture 26. Graphs. CS 441 Discrete mathematics for CS. Final exam

Graphical Models. Pradeep Ravikumar Department of Computer Science The University of Texas at Austin

2. Graphical Models. Undirected pairwise graphical models. Factor graphs. Bayesian networks. Conversion between graphical models. Graphical Models 2-1

Weighted Treewidth: Algorithmic Techniques and Results

Graphs and Discrete Structures

Probabilistic Graphical Models

Junction Trees and Chordal Graphs

Colouring graphs with no odd holes

Exact Algorithms Lecture 7: FPT Hardness and the ETH

Exact Inference: Elimination and Sum Product (and hidden Markov models)

Minimal Dominating Sets in Graphs: Enumeration, Combinatorial Bounds and Graph Classes

Honour Thy Neighbour Clique Maintenance in Dynamic Graphs

On 2-Subcolourings of Chordal Graphs

Exact algorithm for the Maximum Induced Planar Subgraph Problem

AND/OR Cutset Conditioning

Variational Methods for Graphical Models

5 Minimal I-Maps, Chordal Graphs, Trees, and Markov Chains

Graph Theory. Probabilistic Graphical Models. L. Enrique Sucar, INAOE. Definitions. Types of Graphs. Trajectories and Circuits.

CS 343: Artificial Intelligence

Lecture 3: Conditional Independence - Undirected

Expectation Propagation

Graphical Models as Block-Tree Graphs

Faster parameterized algorithms for Minimum Fill-In

Exponential time algorithms for the minimum dominating set problem on some graph classes

Conflict Graphs for Combinatorial Optimization Problems

Faster parameterized algorithms for Minimum Fill-In

Information Processing Letters

Probabilistic Graphical Models

Applied Mathematics Letters. Graph triangulations and the compatibility of unrooted phylogenetic trees

arxiv: v1 [cs.dm] 21 Dec 2015

FEDOR V. FOMIN. Lectures on treewidth. The Parameterized Complexity Summer School 1-3 September 2017 Vienna, Austria

Learning the Structure of Sum-Product Networks. Robert Gens Pedro Domingos

Bayesian Networks, Winter Yoav Haimovitch & Ariel Raviv

Homework 1: Belief Propagation & Factor Graphs

OSU CS 536 Probabilistic Graphical Models. Loopy Belief Propagation and Clique Trees / Join Trees

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014

Graph Isomorphism. Algorithms and networks

Iterative Algorithms for Graphical Models 1

Graphs Representable by Caterpillars. Nancy Eaton

EE512A Advanced Inference in Graphical Models

Computing Largest Correcting Codes and Their Estimates Using Optimization on Specially Constructed Graphs p.1/30

Models for grids. Computer vision: models, learning and inference. Multi label Denoising. Binary Denoising. Denoising Goal.

Modeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA

EE512 Graphical Models Fall 2009

Chordal graphs MPRI

Chordal deletion is fixed-parameter tractable

Lecture 11: May 1, 2000

4 Factor graphs and Comparing Graphical Model Types

Transcription:

3750 Machine earning ecture 4 Markov Random ields Milos auskrecht milos@cs.pitt.edu 5329 ennott quare 3750 dvanced Machine earning Markov random fields Probabilistic models with symmetric dependences. Typically models spatially varying quantities - potential function (defined over factors) - f is strictly positive we can rewrite the definition as: - nergy function - ibbs (oltzman) distribution - partition function 3750 dvanced Machine earning 1

raphical representation of MRs n undirected network (also called independence graph) = (, ) =1, 2,.. N correspond to random variables xample: or x i and x j appear within the same factor c variables,.. ssume the full joint of MR 3750 dvanced Machine earning Markov random fields regular lattice (sing model) rbitrary graph 3750 dvanced Machine earning 2

Markov random fields regular lattice (sing model) rbitrary graph 3750 dvanced Machine earning Markov random fields Pairwise Markov property Two nodes in the network that are not directly connected can be made independent given all other nodes 3750 dvanced Machine earning 3

Markov random fields Pairwise Markov property Two nodes in the network that are not directly connected can be made independent given all other nodes ocal Markov property set of nodes (variables) can be made independent from the rest of nodes variables given its immediate neighbors lobal Markov property vertex set is independent of the vertex set ( and are disjoint) given set if all chains in between elements in and intersect 3750 dvanced Machine earning Types of Markov random fields MRs with discrete random variables lique potentials can be defined by mapping all cliquevariable instances to R xample: ssume two binary variables, with values {a1,a2,a3} and {b1,b2} are in the same clique c. Then: a1 b1 0.5 a1 b2 0.2 a2 b1 0.1 a2 b2 0.3 a3 b1 0.2 a3 b2 0.4 3750 dvanced Machine earning 4

Types of Markov random fields aussian Markov Random ield Precision matrix Variables in x are connected in the network only if they have a nonzero entry in the precision matrix ll zero entries are not directly connected Why? 3750 dvanced Machine earning MR variable elimination inference xample: liminate 3750 dvanced Machine earning 5

actors actor: is a function that maps value assignments for a subset of random variables to R (reals) The scope of the factor: a set of variables defining the factor xample: ssume discrete random variables x (with values a1,a2, a3) and y (with values b1 and b2) actor: a1 b1 0.5 a1 b2 0.2 cope of the factor: a2 b1 0.1 a2 b2 0.3 a3 b1 0.2 a3 b2 0.4 3750 dvanced Machine earning Variables:,, actor Product a1 b1 c1 0.5*0.1 a1 b1 c2 0.5*0.6 b1 c1 0.1 b1 c2 0.6 b2 c1 0.3 b2 c2 0.4 a1 b1 0.5 a1 b2 0.2 a2 b1 0.1 a2 b2 0.3 a3 b1 0.2 a3 b2 0.4 a1 b2 c1 0.2*0.3 a1 b2 c2 0.2*0.4 a2 b1 c1 0.1*0.1 a2 b1 c2 0.1*0.6 a2 b2 c1 0.3*0.3 a2 b2 c2 0.3*0.4 a3 b1 c1 0.2*0.1 a3 b1 c2 0.2*0.6 a3 b2 c1 0.4*0.3 a3 b2 c2 0.4*0.4 3750 dvanced Machine earning 6

Variables:,, actor Marginalization a1 b1 c1 0.2 a1 b1 c2 0.35 a1 b2 c1 0.4 a1 b2 c2 0.15 a2 b1 c1 0.5 a2 b1 c2 0.1 a2 b2 c1 0.3 a2 b2 c2 0.2 a3 b1 c1 0.25 a1 c1 0.2+0.4=0.6 a1 c2 0.35+0.15=0.5 a2 c1 0.8 a2 c2 0.3 a3 c1 0.4 a3 c2 0.7 a3 b1 c2 0.45 a3 b2 c1 0.15 a3 b2 c2 0.25 3750 dvanced Machine earning MR variable elimination inference xample (cont): liminate 3750 dvanced Machine earning 7

8 3750 dvanced Machine earning MR variable elimination inference xample (cont): liminate 3750 dvanced Machine earning MR variable elimination inference xample (cont): liminate

9 3750 dvanced Machine earning MR variable elimination inference xample (cont): liminate 3750 dvanced Machine earning MR variable elimination inference xample (cont): liminate

MR variable elimination inference xample (cont): liminate 3750 dvanced Machine earning nduced graph graph induced by a specific variable elimination order: a graph extended by links that represent intermediate factors. 3750 dvanced Machine earning 10

Tree decomposition of the graph tree decomposition of a graph : tree T with a vertex set associated to every node. or all edges {v,w} : there is a set containing both v and w in T. or every v : the nodes in T that contain v form a connected subtree. 3750 dvanced Machine earning Tree decomposition of the graph tree decomposition of a graph : tree T with a vertex set associated to every node. or all edges {v,w} : there is a set containing both v and w in T. or every v : the nodes in T that contain v form a connected subtree. liques in the graph 3750 dvanced Machine earning 11

Tree decomposition of the graph tree decomposition of a graph : tree T with a vertex set associated to every node. or all edges {v,w} : there is a set containing both v and w in T. or every v : the nodes in T that contain v form a connected subtree. 3750 dvanced Machine earning Tree decomposition of the graph nother tree decomposition of a graph : tree T with a vertex set associated to every node. or all edges {v,w} : there is a set containing both v and w in T. or every v : the nodes in T that contain v form a connected subtree. 3750 dvanced Machine earning 12

Tree decomposition of the graph nother tree decomposition of a graph : tree T with a vertex set associated to every node. or all edges {v,w} : there is a set containing both v and w in T. or every v : the nodes in T that contain v form a connected subtree. 3750 dvanced Machine earning Treewidth of the graph Width of the tree decomposition: Treewidth of a graph : tw()= minimum width over all tree decompositions of. 3750 dvanced Machine earning 13

Treewidth of the graph Treewidth of a graph : tw()= minimum width over all tree decompositions of Why is it important? The calculations can take advantage of the structure and be performed more efficiently treewidth gives the best case complexity vs 3750 dvanced Machine earning Trees Why do we like trees? nference in trees structures can be done in time linear in the number of nodes in the tree 3750 dvanced Machine earning 14

onverting Ns to MRs Moral-graph []: of a bayesian network over X is an undirected graph over X that contains an edge between x and y if: There exists a directed edge between them in. They are both parents of the same node in. 3750 dvanced Machine earning Moral raphs Why moralization? 3750 dvanced Machine earning 15

hordal graphs hordal raph: an undirected graph whose minimum cycle contains 3 verticies. hordal. Not hordal. 3750 dvanced Machine earning hordal raphs Properties: There exists an elimination ordering that adds no edges. The minimal induced treewidth of the graph is equal to the size of the largest clique - 1. 3750 dvanced Machine earning 16

Triangulation The process of converting a graph into a chordal graph is called Triangulation. new graph obtained via triangulation is: 1) uaranteed to be chordal. 2) Not guaranteed to be (treewidth) optimal. There exist exact algorithms for minimal chordal graphs, and heuristic methods with a guaranteed upper bound. 3750 dvanced Machine earning hordal raphs iven a minimum triangulation for a graph, we can carry out the variable-elimination algorithm in the minimum possible time. omplexity of the optimal triangulation: inding the minimal triangulation is NP-ard. The inference limit: nference time is exponential in terms of the largest clique (factor) in. 3750 dvanced Machine earning 17

nference: conclusions We cannot escape exponential costs in the treewidth. ut in many graphs the treewidth is much smaller than the total number of variables till a problem: inding the optimal decomposition is hard ut, paying the cost up front may be worth it. Triangulate once, query many times. Real cost savings if not a bounded one. 3750 dvanced Machine earning 18