Probabilistic Graphical Models Part III: Example Applications

Similar documents
Non-Bayesian Classifiers Part I: k-nearest Neighbor Classifier and Distance Functions

Artistic ideation based on computer vision methods

Improving Recognition through Object Sub-categorization

Texture Analysis. Selim Aksoy Department of Computer Engineering Bilkent University

Undirected Graphical Models. Raul Queiroz Feitosa

Comparing Local Feature Descriptors in plsa-based Image Models

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering

Object recognition (part 2)

A Graph Theoretic Approach to Image Database Retrieval

Semantic-based image analysis with the goal of assisting artistic creation

Unsupervised Learning. CS 3793/5233 Artificial Intelligence Unsupervised Learning 1

Context Based Object Categorization: A Critical Survey

Structural and Syntactic Pattern Recognition

Using Multiple Segmentations to Discover Objects and their Extent in Image Collections

Computer Vision. Exercise Session 10 Image Categorization

Improving the Efficiency of Fast Using Semantic Similarity Algorithm

Modeling semantic aspects for cross-media image indexing

Integrating co-occurrence and spatial contexts on patch-based scene segmentation

Clustering Lecture 5: Mixture Model

Semantic Segmentation of Street-Side Images

Continuous Visual Vocabulary Models for plsa-based Scene Recognition

Multi-resolution Segmentation and Shape Analysis for Remote Sensing Image Classification

Figure 1: Workflow of object-based classification

Video Google: A Text Retrieval Approach to Object Matching in Videos

INF 4300 Classification III Anne Solberg The agenda today:

Unsupervised discovery of category and object models. The task

Object of interest discovery in video sequences

Segmentation. Bottom up Segmentation Semantic Segmentation

Computer vision: models, learning and inference. Chapter 10 Graphical Models

Expectation Maximization (EM) and Gaussian Mixture Models

Image Segmentation. Selim Aksoy. Bilkent University

Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines

Analysis: TextonBoost and Semantic Texton Forests. Daniel Munoz Februrary 9, 2009

Textural Features for Image Database Retrieval

D-Separation. b) the arrows meet head-to-head at the node, and neither the node, nor any of its descendants, are in the set C.

Image Segmentation. Selim Aksoy. Bilkent University

10-701/15-781, Fall 2006, Final

AS STORAGE and bandwidth capacities increase, digital

Norbert Schuff VA Medical Center and UCSF

Context Based Object Categorization: A Critical Survey

Computer Vision Group Prof. Daniel Cremers. 4. Probabilistic Graphical Models Directed Models

Computer Vision Group Prof. Daniel Cremers. 4. Probabilistic Graphical Models Directed Models

Automatic Linguistic Indexing of Pictures by a Statistical Modeling Approach

Introduction to Pattern Recognition Part II. Selim Aksoy Bilkent University Department of Computer Engineering

Automatic Segmentation of Semantic Classes in Raster Map Images

Markov Networks in Computer Vision. Sargur Srihari

Scene Modeling Using Co-Clustering

Unsupervised Identification of Multiple Objects of Interest from Multiple Images: discover

IJREAT International Journal of Research in Engineering & Advanced Technology, Volume 1, Issue 5, Oct-Nov, 2013 ISSN:

Context. CS 554 Computer Vision Pinar Duygulu Bilkent University. (Source:Antonio Torralba, James Hays)

INF4820, Algorithms for AI and NLP: Hierarchical Clustering

Part-based and local feature models for generic object recognition

Semantic Segmentation. Zhongang Qi

arxiv: v3 [cs.cv] 3 Oct 2012

Unsupervised Learning of Categories from Sets of Partially Matching Image Features

Bus Detection and recognition for visually impaired people

Local Feature Detectors

Unsupervised Learning

Spatial Latent Dirichlet Allocation

Latent Variable Models for Structured Prediction and Content-Based Retrieval

570 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 2, FEBRUARY 2011

Multiple-Choice Questionnaire Group C

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Markov Random Fields for Recognizing Textures modeled by Feature Vectors

Previously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011

Lab #4 Introduction to Image Processing II and Map Accuracy Assessment

Using the Kolmogorov-Smirnov Test for Image Segmentation

Markov Networks in Computer Vision

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Bioimage Informatics

An Introduction to Content Based Image Retrieval

Contrast adjustment via Bayesian sequential partitioning

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Region-based Segmentation

COMPUTATIONAL STATISTICS UNSUPERVISED LEARNING

Automatic Image Annotation and Retrieval Using Hybrid Approach

Part III: Affinity Functions for Image Segmentation

Unsupervised Segmentation of Synthetic Aperture Radar Sea Ice Imagery Using A Novel Markov Random Field Model

Computer Vision Group Prof. Daniel Cremers. 4a. Inference in Graphical Models

Generative and discriminative classification techniques

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

COMS 4771 Clustering. Nakul Verma

Simultaneous Appearance Modeling and Segmentation for Matching People under Occlusion

Modeling Scenes with Local Descriptors and Latent Aspects

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

Variational Methods for Discrete-Data Latent Gaussian Models

Pattern recognition (4)

Computer Vision I - Filtering and Feature detection

Dietrich Paulus Joachim Hornegger. Pattern Recognition of Images and Speech in C++

MATRIX BASED SEQUENTIAL INDEXING TECHNIQUE FOR VIDEO DATA MINING

Tri-modal Human Body Segmentation

Probabilistic Graphical Models

Det De e t cting abnormal event n s Jaechul Kim

Using Structural Features to Detect Buildings in Panchromatic Satellite Images

Lecture 3: Conditional Independence - Undirected

Probabilistic Generative Models for Machine Vision

Introduction to Machine Learning CMU-10701

Machine Learning. Unsupervised Learning. Manfred Huber

Jing Gao 1, Feng Liang 1, Wei Fan 2, Chi Wang 1, Yizhou Sun 1, Jiawei i Han 1 University of Illinois, IBM TJ Watson.

Transcription:

Probabilistic Graphical Models Part III: Example Applications Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2014 CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 1 / 38

Introduction We will look at example uses of Bayesian networks and Markov networks for the following applications: Alarm network for monitoring intensive care patients Bayesian networks Recommendation system Bayesian networks Diagnostic systems Bayesian networks Statistical text analysis probabilistic latent semantic analysis Scene classification probabilistic latent semantic analysis Object detection probabilistic latent semantic analysis Image segmentation Markov random fields Contextual classification conditional random fields CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 2 / 38

Intensive Care Monitoring Figure 1: The alarm network for monitoring intensive care patients. The network has 37 variables and 509 parameters (full joint has 2 37 ). CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 3 / 38

Recommendation Systems Given user preferences, the system can suggest recommendations. Input: movie preferences of many users. Output: model correlations between movie features. Users that like comedy, often like drama. Users that like action, often do not like cartoons. Users that like Robert De Niro films, often like Al Pacino films. Given user preferences, the system can predict the probability that new movies match preferences. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 4 / 38

Diagnostic Systems Figure 2: Diagnostic indexing for home health site at Microsoft. Users can enter symptoms and can get recommendations. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 5 / 38

Statistical Text Analysis T. Hofmann, Unsupervised learning by probabilistic latent semantic analysis, Machine Learning, vol. 42, no. 1 2, pp. 177 196, January February 2001. The probabilistic latent semantic analysis (PLSA) algorithm has been originally developed for statistical text analysis to discover topics in a collection of documents that are represented using the frequencies of words from a vocabulary. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 6 / 38

Statistical Text Analysis PLSA uses a graphical model for the joint probability of the documents and their words in terms of the probability of observing a word given a topic (aspect) and the probability of a topic given a document. Suppose there are N documents having content coming from a vocabulary with M words. The collection of documents is summarized in an N-by-M co-occurrence table n where n(d i, w j ) stores the number of occurrences of word w j in document d i. In addition, there is a latent topic variable z k associated with each observation, an observation being the occurrence of a word in a particular document. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 7 / 38

Statistical Text Analysis d z w P (d) P (z d) P (w z) Figure 3: The graphical model used by PLSA for modeling the joint probability P (w j, d i, z k ). CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 8 / 38

Statistical Text Analysis The generative model P (d i, w j ) = P (d i )P (w j d i ) for word content of documents can be computed using the conditional probability P (w j d i ) = K P (w j z k )P (z k d i ). k=1 P (w j z k ) denotes the topic-conditional probability of word w j occurring in topic z k. P (z k d i ) denotes the probability of topic z k observed in document d i. K is the number of topics. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 9 / 38

Statistical Text Analysis Then, the topic specific word distribution P (w j z k ) and the document specific word distribution P (w j d i ) can be used to determine similarities between topics and documents. In PLSA, the goal is to identify the probabilities P (w j z k ) and P (z k d i ). These probabilities are learned using the EM algorithm. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 10 / 38

Statistical Text Analysis In the E-step, the posterior probability of the latent variables are computed based on the current estimates of the parameters as P (z k d i, w j ) = P (w j z k )P (z k d i ) K l=1 P (w j z l )P (z l d i ). In the M-step, the parameters are updated to maximize the expected complete data log-likelihood as N i=1 P (w j z k ) = n(d i, w j )P (z k d i, w j ) M N m=1 i=1 n(d i, w m )P (z k d i, w m ), P (z k d i ) = M j=1 n(d i, w j )P (z k d i, w j ) M j=1 n(d. i, w j ) CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 11 / 38

Statistical Text Analysis Figure 4: Four aspects (topics) to most likely generate the word segment, derived from a K = 128 aspects model of a document collection consisting of abstracts of 1568 documents on clustering. The displayed word stems are the most probable words in the class-conditional distribution P (w j z k ), from top to bottom in descending order. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 12 / 38

Statistical Text Analysis Figure 5: Abstracts of four examplary documents from the collection along with latent class posterior probabilities P (z k d i, w = segment ) and word probabilities P (w = segment d i ). CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 13 / 38

Scene Classification P. Quelhas, F. Monay, J.-M. Odobez, D. Gatica-Perez, T. Tuytelaars, A Thousand Words in a Scene, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 29, no. 9, pp. 1575 1589, September 2007. The PLSA model is used for scene classification by modeling images using visual words (visterms). The topic (aspect) probabilities are used as features as an alternative representation to the word histograms. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 14 / 38

Scene Classification Figure 6: Image representation as a collection of visual words (visterms). CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 15 / 38

Scene Classification Figure 7: 10 most probable images from a data set consisting of city and landscape images for seven topics (aspects) out of 20. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 16 / 38

Object Detection H. G. Akcay, S. Aksoy, Automatic Detection of Geospatial Objects Using Multiple Hierarchical Segmentations, IEEE Transactions on Geoscience and Remote Sensing, vol. 46, no. 7, pp. 2097 2111, July 2008. We used the PLSA technique for object detection to model the joint probability of the segments and their features in terms of the probability of observing a feature given an object and the probability of an object given the segment. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 17 / 38

Object Detection k means quantization histogram of pixels Figure 8: After image segmentation, each segment is modeled using the statistical summary of its pixel content (e.g., quantized spectral values). CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 18 / 38

Object Detection P (s) s P (t s) t P (x t) x building s t s x x t = (a) P (x s) P (x t) P (t s) (b) Figure 9: (a) PLSA graphical model. The filled nodes indicate observed random variables whereas the unfilled node is unobserved. The red arrows show examples for the measurements represented at each node. (b) In PLSA, the object specific feature probability, P (x j t k ), and the segment specific object probability, P (t k s i ), are used to compute the segment specific feature probability, P (x j s i ). CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 19 / 38

Object Detection After learning the parameters of the model, we want to find good segments belonging to each object type. This is done by comparing the object specific feature distribution P (x t) and the segment specific feature distribution P (x s). The similarity between two distributions can be measured using the Kullback-Leibler (KL) divergence D(p(x s) p(x t)). Then, for each object type, the segments can be sorted according to their KL divergence scores, and the most representative ones for that object type can be selected. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 20 / 38

Object Detection (a) Image (b) Buildings (c) Roads (d) Vegetation (e) Water Figure 10: Examples of object detection. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 21 / 38

Object Detection (a) Image (b) Buildings (c) Roads (d) Vegetation Figure 11: Examples of object detection. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 22 / 38

Image Segmentation Z. Kato, T.-C. Pong, A Markov random field image segmentation model for color textured images, Image and Vision Computing, vol. 24, no. 10, pp. 1103 1114, October 2006. Markov random fields are used as a neighborhood model for image segmentation by classifying pixels into different pixel classes. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 23 / 38

Image Segmentation The goal is to assign each pixel into a set of labels w Ω. Pixels are modeled using color and texture features. Pixel features are modeled using multivariate Gaussians, p(x w). A first-order neighborhood system is used as the prior for the labeling process. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 24 / 38

Image Segmentation Figure 12: The Markov random field used as the first-order neighborhood model for the labeling process. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 25 / 38

Image Segmentation The prior is modeled as ( p(w) = 1 Z exp ) V c (w c ) c C where V c denotes the clique potential of clique c C having the label configuration w c. Each clique corresponds to a pair of neighboring pixels. The potentials favor similar classes in neighboring pixels as +1 if w s w r, V c = δ(w s, w r ) = 1 otherwise. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 26 / 38

Image Segmentation The prior is proportional to the length of the region boundaries. Thus, homogeneous segmentations will get a higher probability. The final labeling for each pixel is done by maximizing the posterior probability p(w x) p(x w)p(w). CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 27 / 38

Image Segmentation Figure 13: Example segmentation results. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 28 / 38

Image Segmentation Figure 14: Example Markov random field models used in the literature. (a) First-order neighborhood system. (b) Non-regular planar graph associated to an image partition. (c) Quad-tree. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 29 / 38

Contextual Classification A. Rabinovich, A. Vedaldi, C. Galleguillos, E. Wiewiora, S. Belongie, Objects in Context, IEEE International Conference on Computer Vision, 2007. Semantic context among objects is used for improving object categorization. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 30 / 38

Contextual Classification Figure 15: Idealized context based object categorization system: an original image is perfectly segmented into objects; each object is categorized; and object s labels are refined with respect to semantic context in the image. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 31 / 38

Contextual Classification Figure 16: Object categorization framework: S 1,..., S k is the set of k segments for an image; L 1,..., L n is a ranked list of n labels for each segment; O 1,..., O m is a set of m object categories in the image. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 32 / 38

Contextual Classification A conditional random field (CRF) framework is used to incorporate semantic context into the object categorization. Given an image I and its segmentation S 1,..., S k, the goal is to find segment labels c 1,..., c k such that they agree with the segment contents and are in contextual agreement with each other. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 33 / 38

Contextual Classification This interaction is modeled as a probability distribution p(c 1,..., c k S 1,..., S k ) = B(c 1,..., c k ) k i=1 A(i) Z(φ, S 1,..., S k ) with ( k ) A(i) = p(c i S i ) and B(c 1,..., c k ) = exp i,j=1 φ(c i, c j ), where Z( ) is the partition function. The semantic context information is modeled using context matrices that are symmetric, nonnegative matrices that contain the co-occurrence frequency among object labels in the training set. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 34 / 38

Contextual Classification Figure 17: An example conditional random field. Squares indicate feature functions and circles indicate variable nodes. Arrows represent single node potentials due to feature functions, and undirected edges represent pairwise potentials. Global context is represented by h. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 35 / 38

Contextual Classification Figure 18: An example context matrix. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 36 / 38

Contextual Classification Figure 19: Example results where context improved the categorization accuracy. Left to right: original segmentation, categorization w/o contextual constraints, categorization w/ contextual constraints, ground truth. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 37 / 38

Contextual Classification Figure 20: Example results where context reduced the categorization accuracy. Left to right: original segmentation, categorization w/o contextual constraints, categorization w/ contextual constraints, ground truth. CS 551, Fall 2014 c 2014, Selim Aksoy (Bilkent University) 38 / 38