VC 17/18 TP14 Pattern Recognition

Similar documents
VC 16/17 TP5 Single Pixel Manipulation

VC 11/12 T14 Visual Feature Extraction

VC 10/11 T9 Region-Based Segmentation

VC 10/11 T4 Colour and Noise

VC 12/13 T16 Video Compression

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Clustering Lecture 5: Mixture Model

VC 11/12 T11 Optical Flow

Recognition: Face Recognition. Linda Shapiro EE/CSE 576

Data: a collection of numbers or facts that require further processing before they are meaningful

Introduction to Pattern Recognition Part II. Selim Aksoy Bilkent University Department of Computer Engineering

ECG782: Multidimensional Digital Signal Processing

Image processing & Computer vision Xử lí ảnh và thị giác máy tính

Waleed Pervaiz CSE 352

Unsupervised Learning

Network Traffic Measurements and Analysis

Non-Bayesian Classifiers Part I: k-nearest Neighbor Classifier and Distance Functions

Some questions of consensus building using co-association

Fuzzy Segmentation. Chapter Introduction. 4.2 Unsupervised Clustering.

Classifying Images with Visual/Textual Cues. By Steven Kappes and Yan Cao

DIGITAL IMAGE ANALYSIS. Image Classification: Object-based Classification

I211: Information infrastructure II

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Instance-based Learning

Wavelet Applications. Texture analysis&synthesis. Gloria Menegaz 1

Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions

Textural Features for Image Database Retrieval

Human motion analysis: methodologies and applications

Introduction to Mobile Robotics

Practical Image and Video Processing Using MATLAB

Object Recognition. Lecture 11, April 21 st, Lexing Xie. EE4830 Digital Image Processing

Perceptron as a graph

22 October, 2012 MVA ENS Cachan. Lecture 5: Introduction to generative models Iasonas Kokkinos

IBL and clustering. Relationship of IBL with CBR

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER

Machine learning Pattern recognition. Classification/Clustering GW Chapter 12 (some concepts) Textures

Machine Learning Lecture 3

Sensor Tasking and Control

Generative and discriminative classification techniques

Machine Learning Lecture 3

COMP 551 Applied Machine Learning Lecture 13: Unsupervised learning

K-Means and Gaussian Mixture Models

Analysis of Image and Video Using Color, Texture and Shape Features for Object Identification

COMPUTER AND ROBOT VISION

Content-based image and video analysis. Machine learning

MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A

Last week. Multi-Frame Structure from Motion: Multi-View Stereo. Unknown camera viewpoints

Introduction to Machine Learning CMU-10701

Image Processing, Analysis and Machine Vision

Unsupervised Learning

Computer Vision 6 Segmentation by Fitting

Clustering. Bruno Martins. 1 st Semester 2012/2013

Short Survey on Static Hand Gesture Recognition

EE795: Computer Vision and Intelligent Systems

Robot Learning. There are generally three types of robot learning: Learning from data. Learning by demonstration. Reinforcement learning

Lecture 11: Classification

Sonar Image Compression

Remote Sensing & Photogrammetry W4. Beata Hejmanowska Building C4, room 212, phone:

Mixture Models and the EM Algorithm

Note Set 4: Finite Mixture Models and the EM Algorithm

Clustering and Visualisation of Data

IMAGE SEGMENTATION. Václav Hlaváč

Tutorial 3. Jun Xu, Teaching Asistant csjunxu/ February 16, COMP4134 Biometrics Authentication

CSE 158. Web Mining and Recommender Systems. Midterm recap

Computer Vision 5 Segmentation by Clustering

Histograms. h(r k ) = n k. p(r k )= n k /NM. Histogram: number of times intensity level rk appears in the image

Pattern recognition. Classification/Clustering GW Chapter 12 (some concepts) Textures

INF 4300 Classification III Anne Solberg The agenda today:

Data Mining Classification: Bayesian Decision Theory

Homework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures:

Verification: is that a lamp? What do we mean by recognition? Recognition. Recognition

Clustering K-means. Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, Carlos Guestrin

DATA MINING AND MACHINE LEARNING. Lecture 6: Data preprocessing and model selection Lecturer: Simone Scardapane

Chapter 4 Fuzzy Logic

Machine Learning Lecture 3

9.1. K-means Clustering

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

Unsupervised Learning. Clustering and the EM Algorithm. Unsupervised Learning is Model Learning

Radial Basis Function Neural Network Classifier

Application of partial differential equations in image processing. Xiaoke Cui 1, a *

Olmo S. Zavala Romero. Clustering Hierarchical Distance Group Dist. K-means. Center of Atmospheric Sciences, UNAM.

Contextual priming for artificial visual perception

Forestry Applied Multivariate Statistics. Cluster Analysis

Motion Patterns: Signal Interpretation towards the Laban Movement Analysis Semantics

A Neural Classifier for Anomaly Detection in Magnetic Motion Capture

Machine Learning. Unsupervised Learning. Manfred Huber

An Introduction to Pattern Recognition

CS325 Artificial Intelligence Ch. 20 Unsupervised Machine Learning

Unsupervised Learning: Clustering

Real time game field limits recognition for robot self-localization using collinearity in Middle-Size RoboCup Soccer

Case-Based Reasoning. CS 188: Artificial Intelligence Fall Nearest-Neighbor Classification. Parametric / Non-parametric.

CS 188: Artificial Intelligence Fall 2008

Image segmentation. Václav Hlaváč. Czech Technical University in Prague

INTRODUCTION TO DATA MINING. Daniel Rodríguez, University of Alcalá

arxiv: v1 [physics.data-an] 27 Sep 2007

Supervised vs unsupervised clustering

Clustering web search results

Automatic Image Alignment (feature-based)

A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models

Announcements. Recognition I. Gradient Space (p,q) What is the reflectance map?

Transcription:

VC 17/18 TP14 Pattern Recognition Mestrado em Ciência de Computadores Mestrado Integrado em Engenharia de Redes e Sistemas Informáticos Miguel Tavares Coimbra

Outline Introduction to Pattern Recognition Statistical Pattern Recognition Classifiers

Topic: Introduction to Pattern Recognition Introduction to Pattern Recognition Statistical Pattern Recognition Classifiers

This is a horse http://www.flickr.com/photos/kimbar/2027234083/

This is a horse http://www.flickr.com/photos/genewolf/2031802050/

This is a... Horse? http://www.flickr.com/photos/masheeebanshee/413465808/

Decisions I can manipulate images. I want to make decisions! Classify / Identify features. Recognize patterns.

One definition Pattern recognition "the act of taking in raw data and taking an action based on the category of the data". Wikipedia How do I do this so well? How can I make machines do this?

The problem Do you see a horse? What a computer sees

Mathematics We only deal with numbers. How do we represent knowledge? How do we represent visual features? How do we classify them? Very complex problem!! Let s break it into smaller ones...

Typical PR system Sensor Gathers the observations to be classified or described Feature Extraction Computes numeric or symbolic information from the observations; Classifier Does the actual job of classifying or describing observations, relying on the extracted features.

Sensor In our specific case: Image acquiring mechanism. Let s assume we don t control it. One observation = One Image Video = Multiple Observations

Feature Extraction What exactly are features? Colour, texture, shape, etc. Animal with 4 legs. Horse. Horse jumping. These vary a lot! Some imply some sort of recognition e.g. How do I know the horse is jumping?

Broad classification of features Low-level Colour, texture Middle-level Object with head and four legs. Object moving up. Horse High-level Horse jumping. Horse competition.

Low-level features Objective Directly reflect specific image and video features. Colour Texture Shape Motion Etc.

Middle-level features Some degree of subjectivity They are typically one solution of a problem with multiple solutions. Examples: Segmentation Optical Flow Identification Etc.

High-level features Semantic Interpretation Knowledge Context Examples: This person suffers from epilepsy. The virus attacks the cell with some degree of intelligence. This person is running from that one. How do humans do this so well?

The semantic gap Fundamental problem of current research! Low-level: -Colour -Texture -Shape - High-level: -Interpretation -Decision -Understanding - Now what?? How do i cross this bridge?

Features & Decisions Low-Level Features Various Possible Solutions Decision One Solution Middle-Level Features How do I decide? Decision High-Level Features

Classification Middle-Level Features Horse Rider Upward Motion Classifier High-Level Features Competition Horse Jumping M inputs, N outputs

Layers of classification Brown Head Four legs......... First layer of classificaiton Horse Rider Upward Motion Second layer of classificaiton Competition Horse Jumping

Classifiers How do I map my M inputs to my N outputs? Mathematical tools: Distance-based classifiers. Rule-based classifiers. Neural Networks. Support Vector Machines...

Types of PR methods Statistical pattern recognition based on statistical characterizations of patterns, assuming that the patterns are generated by a probabilistic system. Syntactical (or structural) pattern recognition based on the structural interrelationships of features.

Topic: Statistical Pattern Recognition Introduction to Pattern Recognition Statistical Pattern Recognition Classifiers

Is Porto in Portugal?

Porto is in Portugal I want to make decisions. Is Porto in Portugal? I know certain things. A world map including cities and countries. I can make this decision! Porto is in Portugal. I had enough a priori knowledge to make this decision.

What if I don t have a map? I still want to make this decision. I observe: Amarante has coordinates x 1,y 1 and is in Portugal. Viseu has coordinates x 2, y 2 and is in Portugal. Vigo has coordinates x 3, y 3 and is in Spain. I classify: Porto is close to Amarante and Viseu so Porto is in Portugal. What if I try to classify Valença?

Statistical PR I used statistics to make a decision. I can make decisions even when I don t have full a priori knowledge of the whole process. I can make mistakes. How did I recognize this pattern? I learned from previous observations where I knew the classification result. I classified a new observation. What pattern?

Back to the Features Feature F i F i f i Feature F i with N values. i F f f,...,, i1 i2 Feature vector F with M features. F F F... 1 2 f in F M Naming conventions: Elements of a feature vector are called coefficients. Features may have one or more coefficients. Feature vectors may have one or more features.

Back to our Porto example I ve classified that Porto is in Portugal. What feature did I use? Spatial location Let s get more formal I ve defined a feature vector F with one feature F 1, which has two coefficients f 1x, f 1y. F [ F ] [ f, f 1x 1 1 y ]

Y Coordinate Feature Space Feature Vector Two total coefficients. Can be seen as a feature space with two orthogonal axis. Feature Space Hyper-space with N dimensions where N is the total number of coefficients of my feature vector. Feature Space 35 30 Vigo 25 20 15 10 Braga 5 0 Porto Amarante -2 0 2 4 6 8 10 12-5 -10 X Coordinate

Y Coordinate A Priori Knowledge I have a precise model of my feature space based on a priori knowledge. City is in Spain if F 1Y >23 35 30 25 Vigo Feature Space I know the border is here Great models = Great classifications. 20 15 F 1Y (London) = 100 10 Braga London is in Spain (??) 5 Porto is in Portugal! Amarante 0 Porto -2 0 2 4 6 8 10 12-5 -10 X Coordinate

What if I don t have a model? I need to learn from observations. Derive a model. Direct classification. Training stage. Learn model parameters. Classification Y Coordinate 35 30 25 20 15 10 5 Vigo Braga Feature Space Amarante 0 Porto -10 0 10 20 30-5 Learned Model -10 X Coordinate

Y Coordinate Classes In our example, cities can belong to: Portugal Spain I have two classes of cities. A class represents a sub-space of my feature space. 35 30 25 20 15 10 5 Vigo Braga Feature Space SPAIN PORTUGAL Amarante 0 Porto -2 0 2 4 6 8 10 12-5 -10 X Coordinate

Classifiers A Classifier C maps a class into the feature space. C Spain ( x, y) Various types of classifiers. Nearest-Neighbours. Bayesian. Soft-computing machines. Etc... true false, y K, otherwise

Topic: Classifiers Introduction to Pattern Recognition Statistical Pattern Recognition Classifiers

Distance to Mean I can represent a class by its mean feature vector. C F To classify a new object, I choose the class with the closest mean feature vector. Different distance measures! Y Coordinate 35 30 25 20 15 10 5 Spain Feature Space Portugal 0 Porto -10 0 10 20 30-5 -10 Euclidean Distance X Coordinate

Possible Distance Measures L1 Distance L1 1 S( x) v( x) N 1 x Euclidean Distance (L2 Distance) N L1 or Taxicab Distance L2 1 N S( x) v( x) 2 N x 1

Gaussian Distribution Defined by two parameters: Mean: μ f ( x) 1 exp 2 ( x u) 2 2 2 Variance: σ 2 Great approximation to the distribution of many phenomena. Central Limit Theorem

Multivariate Distribution For N dimensions: Mean feature vector: Covariance Matrix: F

Mahalanobis Distance Based on the covariance of coefficients. Superior to the Euclidean distance.

K-Nearest Neighbours Algorithm Choose the closest K neighbours to a new observation. Classify the new object based on the class of these K objects. Characteristics Assumes no model. Does not scale very well...

Resources Gonzalez & Woods, 3rd Ed, Chapter 12. Andrew Moore, Statistic Data Mining Tutorial, http://www.autonlab.org/tutorials/