Patch-based Object Recognition. Basic Idea

Similar documents
Scale Invariant Feature Transform

Introduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.

Scale Invariant Feature Transform

Local Features Tutorial: Nov. 8, 04

The SIFT (Scale Invariant Feature

Fitting: The Hough transform

Scale Invariant Feature Transform by David Lowe

Fitting: The Hough transform

CAP 5415 Computer Vision Fall 2012

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

SIFT: Scale Invariant Feature Transform

Outline 7/2/201011/6/

Previously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011

Object Recognition with Invariant Features

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Fitting: The Hough transform

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

Object Detection. Sanja Fidler CSC420: Intro to Image Understanding 1/ 1

Local features: detection and description. Local invariant features

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Advanced Video Content Analysis and Video Compression (5LSH0), Module 4

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

Object Category Detection. Slides mostly from Derek Hoiem

Motion Estimation and Optical Flow Tracking

Visuelle Perzeption für Mensch- Maschine Schnittstellen

Part-based and local feature models for generic object recognition

Local features: detection and description May 12 th, 2015

CS 558: Computer Vision 4 th Set of Notes

Beyond bags of features: Adding spatial information. Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Implementing the Scale Invariant Feature Transform(SIFT) Method

Model Fitting: The Hough transform II

Local invariant features

Computer Vision for HCI. Topics of This Lecture

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

Local Features: Detection, Description & Matching

Image Features: Detection, Description, and Matching and their Applications

Beyond Bags of features Spatial information & Shape models

Distinctive Image Features from Scale-Invariant Keypoints

Distinctive Image Features from Scale-Invariant Keypoints

Verslag Project beeldverwerking A study of the 2D SIFT algorithm

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability

Local Image Features

Distinctive Image Features from Scale-Invariant Keypoints

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

Image Segmentation and Registration

SIFT - scale-invariant feature transform Konrad Schindler

Feature Based Registration - Image Alignment

Part based models for recognition. Kristen Grauman

Lecture 8: Fitting. Tuesday, Sept 25

Local Image Features

School of Computing University of Utah

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Local features and image matching. Prof. Xin Yang HUST

Key properties of local features

Instance-level recognition part 2

Feature Matching and Robust Fitting

Instance-level recognition II.

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

Local Feature Detectors

Eppur si muove ( And yet it moves )

Problems with template matching

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

Motion illusion, rotating snakes

Feature Detection and Matching

Wikipedia - Mysid

Computer Vision. Recap: Smoothing with a Gaussian. Recap: Effect of σ on derivatives. Computer Science Tripos Part II. Dr Christopher Town

Image Processing. Image Features

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

Designing Applications that See Lecture 7: Object Recognition

Obtaining Feature Correspondences

Prof. Feng Liu. Spring /26/2017

High-Level Computer Vision

Introduction to SLAM Part II. Paul Robertson

Automatic Image Alignment (feature-based)

Discovering Visual Hierarchy through Unsupervised Learning Haider Razvi

Comparison of Feature Detection and Matching Approaches: SIFT and SURF

Image features. Image Features

Deformable Part Models

2D Image Processing Feature Descriptors

Performance Evaluation of Scale-Interpolated Hessian-Laplace and Haar Descriptors for Feature Matching

Recognition of Degraded Handwritten Characters Using Local Features. Markus Diem and Robert Sablatnig

Determinant of homography-matrix-based multiple-object recognition

Object Detection by Point Feature Matching using Matlab

Lecture: RANSAC and feature detectors

CS231A Section 6: Problem Set 3

Hough Transform and RANSAC

Corner Detection. GV12/3072 Image Processing.

3D Object Recognition using Multiclass SVM-KNN

Feature descriptors and matching

Scale Invariant Feature Transform (SIFT) CS 763 Ajit Rajwade

CS 223B Computer Vision Problem Set 3

Bias-Variance Trade-off (cont d) + Image Representations

3D Reconstruction From Multiple Views Based on Scale-Invariant Feature Transform. Wenqi Zhu

CS 231A Computer Vision (Fall 2012) Problem Set 3

10/03/11. Model Fitting. Computer Vision CS 143, Brown. James Hays. Slides from Silvio Savarese, Svetlana Lazebnik, and Derek Hoiem

Robotics Programming Laboratory

Finding 2D Shapes and the Hough Transform

3D OBJECT RECONTRUCTION USING MULTIPLE-VIEW GEOMETRY: SIFT DETECTION LEOW RUEY SHYAN

Transcription:

Patch-based Object Recognition 1! Basic Idea Determine interest points in image Determine local image properties around interest points Use local image properties for object classification Example: Interest points determined by Haralick Operator Example: Gabor-Filterbank for local image description 2! 1

Interest Operators (1) Moravec interest operator: M(i,j) = 1 8 i+ 1 " j+ 1 " m= i! 1n= j! 1 g(m,n)! g(i,j) ij Zuniga-Haralick operator: fit a cubic polynomial f(i,j) = c 1 + c 2 x + c 3 y + c 4 x 2 + c 5 xy + c 6 y 2 + c 7 x 3 + c 8 x 2 y + c 9 xy 2 + c 10 y 3 For a 5x5 neighbourhood the coefficients of the best-fitting polynomial can be directly determined from the 25 greyvalues compute interest value from polynomial coefficients ZH(i,j) =!2(c 2 2c 6! c 2 c 3 c 5! c 32 c 4 ) c 2 2 ( 2 + c 3 ) 3 2 measure of "cornerness" of the polynomial 3! Interest Operators (2) Difference-of-Gaussians (DoG) Locates edges at zero crossings of second derivative of smoothed image, approximated by difference of Gaussians "mexican-hat operator" Harris interest operator: R = det(h) - k tr(h) = αβ - k (α+β)! Determine points with two strong principle curvatures α and β are eigenvalues of Hessian matrix H and proportional to main curvatures 4! 2

SIFT Features SIFT = Scale Invariant Image Features David G. Lowe: Distinctive Image Features from Scale-Invariant Keypoints International Journal of Computer Vision, 2004 5! Computation Steps for SIFT Features 1. Scale-space extrema detection: The first stage of computation searches over all scales and image locations. It is implemented efficiently by using a difference-of- Gaussian function to identify potential interest points that are invariant to scale and orientation. 2. Keypoint localization: At each candidate location, a detailed model is fit to determine location and scale. Keypoints are selected based on measures of their stability. 3. Orientation assignment: One or more orientations are assigned to each keypoint location based on local image gradient directions. All future operations are performed on image data that has been transformed relative to the assigned orientation, scale, and location for each feature, thereby providing invariance to these transformations. 4. Keypoint descriptor: The local image gradients are measured at the selected scale in the region around each keypoint. These are transformed into a representation that allows for significant levels of local shape distortion and change in illumination. 6! 3

Scale Space! 7! Example Image in Scale Space! 8! 4

Maxima Detection! 9! Keypoint Selection! 10! 5

Eliminating Edge Responses! Compute Hessian at keypoint:! Eigenvalues α and β are proportional to principal curvatures.! Both principle curvatures must be significant for a keypoint to be stable.! Note that! Hence one can check the ratio r = α/β of the principle curvatures by evaluating! 11! Local Image Descriptor! 12! 6

SIFT Feature Matching Find nearest neighbour in a database of SIFT features from training images. For robustness, use ratio of nearest neighbor to ratio of second nearest neighbour. Neighbour with minimum Euclidean distance => expensive search. Use an approximate, fast method to find nearest neighbour with high probability 13! Recognition Using SIFT Features Compute SIFT features on the input image Match these features to the SIFT feature database Each keypoint specifies 4 parameters: 2D location, scale, and orientation. To increase recognition robustness: Hough transform to identify clusters of matches that vote for the same object pose. Each keypoint votes for the set of object poses that are consistent with the keypoint's location, scale, and orientation. Locations in the Hough accumulator that accumulate at least 3 votes are selected as candidate object/pose matches. A verfication step matches the training image for the hypothesized object/ pose to the image using a least-squares fit to the hypothesized location, scale, and orientation of the object. 14! 7

Experiments (1)! 15! Experiments (2)! 16! 8

SIFT Features Summary SIFT features are reasonably invariant to rotation, scaling, and illumination changes. They can be used for matching and object recognition (among other things). Robust to occlusion: as long as we can see at least 3 features from the object we can compute the location and pose. Efficient on-line matching: recognition can be performed in close-toreal time (at least for small object databases). 17! Patch-based Object Categorization and Segmentation Bastian Leibe, Ales Leonardis, and Bernt Schiele: Combined Object Categorization and Segmentation with an Implicit Shape Model In ECCV 04 Workshop on Statistical Learning in Computer Vision, Prague, May 2004. Define a shape model by for an object class (or category) by a class-specific collection of local appearances (a codebook), a spatial probability distribution specifying where each codebook, entry may be found on the object To recognize an object, extract image patches around interest points and and compare them with the codebook. Matching patches cast probabilistic votes leading to object hypotheses. Each pixel of an object hypothesis is classified as object or background based on the contributing patches. 18! 9

22.12.10 Implicit Shape Model - Representation! 105 training images (+ motion segmentation) y Learn appearance codebook Ø Ø Ø y Extract patches at DoG interest points Agglomerative clustering codebook s s x y Learn spatial distributions Ø Appearance codebook x y Match codebook to training images Record matching positions on object s s x x Spatial occurrence distributions 19! Implicit Shape Model - Recognition (1)! Interest Points Image Patch e Matched Codebook Entries Interpretation (Codebook match) I Probabilistic Voting Object Position o,x 20! 10

22.12.10 Implicit Shape Model - Recognition (2)! Matched Codebook Entries Interest Points Probabilistic Voting!Spatial feature configurations!!interleaved object recognition and segmentation! Voting Space (continuous) Segmentation Refined Hypotheses (uniform sampling) Backprojected Hypotheses Backprojection of Maxima 21! Car Detection Ø Ø Recognizes different kinds of cars Robust to clutter, occlusion, noise, low contrast 22! 11

Cow Detection frame-by-frame detection no temporal continuity exploited 23! 12