Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

Similar documents
Local features: detection and description. Local invariant features

Local features: detection and description May 12 th, 2015

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability

Local features and image matching. Prof. Xin Yang HUST

CS4670: Computer Vision

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford

Local Features: Detection, Description & Matching

Image matching. Announcements. Harder case. Even harder case. Project 1 Out today Help session at the end of class. by Diva Sian.

Feature Detectors and Descriptors: Corners, Lines, etc.

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford

Feature descriptors and matching

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Patch Descriptors. CSE 455 Linda Shapiro

Feature Based Registration - Image Alignment

Local Patch Descriptors

Local Image Features

Local invariant features

The SIFT (Scale Invariant Feature

Image Features. Work on project 1. All is Vanity, by C. Allan Gilbert,

Homography estimation

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

Patch Descriptors. EE/CSE 576 Linda Shapiro

Local Image Features

Automatic Image Alignment (feature-based)

Local Image Features

Feature Matching and RANSAC

CS 231A Computer Vision (Winter 2018) Problem Set 3

Prof. Feng Liu. Spring /26/2017

SURF. Lecture6: SURF and HOG. Integral Image. Feature Evaluation with Integral Image

Object Recognition with Invariant Features

Computer Vision for HCI. Topics of This Lecture

Instance-level recognition

Local Features Tutorial: Nov. 8, 04

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Instance-level recognition

Motion Estimation and Optical Flow Tracking

Eppur si muove ( And yet it moves )

2D Image Processing Feature Descriptors

CAP 5415 Computer Vision Fall 2012

Object Detection. Sanja Fidler CSC420: Intro to Image Understanding 1/ 1

CS 558: Computer Vision 4 th Set of Notes

Feature Matching and Robust Fitting

Instance-level recognition part 2

Automatic Image Alignment

Recognition. Topics that we will try to cover:

Lecture: RANSAC and feature detectors

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

CSE 527: Introduction to Computer Vision

CS5670: Computer Vision

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

Scale Invariant Feature Transform

A Comparison of SIFT, PCA-SIFT and SURF

Motion illusion, rotating snakes

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

A System of Image Matching and 3D Reconstruction

TA Section 7 Problem Set 3. SIFT (Lowe 2004) Shape Context (Belongie et al. 2002) Voxel Coloring (Seitz and Dyer 1999)

HISTOGRAMS OF ORIENTATIO N GRADIENTS

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

Outline 7/2/201011/6/

Local Feature Detectors

Automatic Image Alignment

Wikipedia - Mysid

Scale Invariant Feature Transform

Introduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.

Coarse-to-fine image registration

Instance-level recognition II.

CSE 252B: Computer Vision II

A Comparison and Matching Point Extraction of SIFT and ISIFT

Evaluation and comparison of interest points/regions

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

Feature Detection and Matching

Image warping and stitching

Image warping and stitching

CS 231A Computer Vision (Winter 2014) Problem Set 3

Image Features: Detection, Description, and Matching and their Applications

Lecture 12 Recognition

SIFT - scale-invariant feature transform Konrad Schindler

Lecture 4.1 Feature descriptors. Trym Vegard Haavardsholm

Lecture 10 Detectors and descriptors

A Novel Algorithm for Color Image matching using Wavelet-SIFT

Image warping and stitching

3D Environment Reconstruction

Keypoint-based Recognition and Object Search

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Bias-Variance Trade-off (cont d) + Image Representations

Prof. Kristen Grauman

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Mosaics. Today s Readings

Homographies and RANSAC

Key properties of local features

Prof. Noah Snavely CS Administrivia. A4 due on Friday (please sign up for demo slots)

Stereo Epipolar Geometry for General Cameras. Sanja Fidler CSC420: Intro to Image Understanding 1 / 33

Miniature faking. In close-up photo, the depth of field is limited.

AK Computer Vision Feature Point Detectors and Descriptors

Recognition: Face Recognition. Linda Shapiro EE/CSE 576

Photo by Carl Warner

Image correspondences and structure from motion

Why is computer vision difficult?

Transcription:

Image Features: Local Descriptors Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

[Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 2/ 58 Local Features Detection: Identify the interest points. Description: Extract a feature descriptor around each interest point. Matching: Determine correspondence between descriptors in two views.

The Ideal Feature Descriptor Repeatable: Invariant to rotation, scale, photometric variations Distinctive: We will need to match it to lots of images/objects! Compact: Should capture rich information yet not be too high-dimensional (otherwise matching will be slow) E cient: We would like to compute it (close-to) real-time Sanja Fidler CSC420: Intro to Image Understanding 3/ 58

Invariances [Source: T. Tuytelaars] Sanja Fidler CSC420: Intro to Image Understanding 4/ 58

Invariances [Source: T. Tuytelaars] Sanja Fidler CSC420: Intro to Image Understanding 4/ 58

What If We Just Took Pixels? The simplest way is to write down the list of intensities to form a feature vector, and normalize them (i.e., mean 0, variance 1). Why normalization? But this is very sensitive to even small shifts, rotations and any a transformation. ne Sanja Fidler CSC420: Intro to Image Understanding 5/ 58

Tones Of Better Options SIFT PCA-SIFT GLOH HOG SURF DAISY LBP Shape Contexts Color Histograms Sanja Fidler CSC420: Intro to Image Understanding 6/ 58

Tones Of Better Options SIFT TODAY PCA-SIFT GLOH HOG SURF DAISY LBP Shape Contexts Color Histograms Sanja Fidler CSC420: Intro to Image Understanding 7/ 58

SIFT Descriptor [Lowe 2004] SIFT stands for Scale Invariant Feature Transform Invented by David Lowe, who also did DoG scale invariant interest points Actually in the same paper, which you should read: David G. Lowe Distinctive image features from scale-invariant keypoints International Journal of Computer Vision, 2004 Paper: http://www.cs.ubc.ca/~lowe/papers/ijcv04.pdf Sanja Fidler CSC420: Intro to Image Understanding 8/ 58

SIFT Descriptor 1 Our scale invariant interest point detector gives scale for each keypoint [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 9/ 58

SIFT Descriptor 2 For each keypoint, we take the Gaussian-blurred image at corresponding scale [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 9/ 58

SIFT Descriptor 3 Compute the gradient magnitude and orientation in neighborhood of each keypoint [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 9/ 58

SIFT Descriptor 3 Compute the gradient magnitude and orientation in neighborhood of each keypoint Sanja Fidler CSC420: Intro to Image Understanding 9/ 58

SIFT Descriptor 4 Compute dominant orientation of each keypoint. How? [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 9/ 58

SIFT Descriptor: Computing Dominant Orientation Compute a histogram of gradient orientations, each bin covers 10 [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 10 / 58

SIFT Descriptor: Computing Dominant Orientation Compute a histogram of gradient orientations, each bin covers 10 Orientations closer to the keypoint center should contribute more [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 10 / 58

SIFT Descriptor: Computing Dominant Orientation Compute a histogram of gradient orientations, each bin covers 10 Orientations closer to the keypoint center should contribute more Orientation giving the peak in the histogram is the keypoint s orientation Sanja Fidler CSC420: Intro to Image Understanding 10 / 58

SIFT Descriptor 4 Compute dominant orientation [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 11 / 58

SIFT Descriptor 5 Compute a 128 dimensional descriptor: 4 4 grid, each cell is a histogram of 8 orientation bins relative to dominant orientation [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 11 / 58

SIFT Descriptor: Computing the Feature Vector Compute the orientations relative to the dominant orientation [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 12 / 58

SIFT Descriptor: Computing the Feature Vector Compute the orientations relative to the dominant orientation [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 12 / 58

SIFT Descriptor: Computing the Feature Vector Compute the orientations relative to the dominant orientation Form a 4 4 grid. For each grid cell compute a histogram of orientations for 8 orientation bins spaced apart by 45 [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 12 / 58

SIFT Descriptor: Computing the Feature Vector Compute the orientations relative to the dominant orientation Form a 4 4 grid. For each grid cell compute a histogram of orientations for 8 orientation bins spaced apart by 45 [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 12 / 58

SIFT Descriptor: Computing the Feature Vector Compute the orientations relative to the dominant orientation Form a 4 4 grid. For each grid cell compute a histogram of orientations for 8 orientation bins spaced apart by 45 Form the 128 dimensional feature vector [Adopted from: F. Flores-Mangas] Sanja Fidler CSC420: Intro to Image Understanding 12 / 58

SIFT Descriptor: Post-processing The resulting 128 non-negative values form a raw version of the SIFT descriptor vector. To reduce the e ects of contrast or gain (additive variations are already removed by the gradient), the 128-D vector is normalized to unit length: f i = f i / f i Sanja Fidler CSC420: Intro to Image Understanding 13 / 58

SIFT Descriptor: Post-processing The resulting 128 non-negative values form a raw version of the SIFT descriptor vector. To reduce the e ects of contrast or gain (additive variations are already removed by the gradient), the 128-D vector is normalized to unit length: f i = f i / f i To further make the descriptor robust to other photometric variations, values are clipped to 0.2 and the resulting vector is once again renormalized to unit length. Sanja Fidler CSC420: Intro to Image Understanding 13 / 58

SIFT Descriptor: Post-processing The resulting 128 non-negative values form a raw version of the SIFT descriptor vector. To reduce the e ects of contrast or gain (additive variations are already removed by the gradient), the 128-D vector is normalized to unit length: f i = f i / f i To further make the descriptor robust to other photometric variations, values are clipped to 0.2 and the resulting vector is once again renormalized to unit length. Great engineering e ort! Sanja Fidler CSC420: Intro to Image Understanding 13 / 58

SIFT Descriptor: Post-processing The resulting 128 non-negative values form a raw version of the SIFT descriptor vector. To reduce the e ects of contrast or gain (additive variations are already removed by the gradient), the 128-D vector is normalized to unit length: f i = f i / f i To further make the descriptor robust to other photometric variations, values are clipped to 0.2 and the resulting vector is once again renormalized to unit length. Great engineering e ort! What is SIFT invariant to? Sanja Fidler CSC420: Intro to Image Understanding 13 / 58

SIFT Descriptor: Post-processing The resulting 128 non-negative values form a raw version of the SIFT descriptor vector. To reduce the e ects of contrast or gain (additive variations are already removed by the gradient), the 128-D vector is normalized to unit length: f i = f i / f i To further make the descriptor robust to other photometric variations, values are clipped to 0.2 and the resulting vector is once again renormalized to unit length. Great engineering e ort! What is SIFT invariant to? Sanja Fidler CSC420: Intro to Image Understanding 13 / 58

Properties of SIFT Invariant to: Scale Rotation Partially invariant to: Illumination changes (sometimes even day vs. night) Camera viewpoint (up to about 60 degrees of out-of-plane rotation) Occlusion, clutter (why?) Also important: Fast and e cient can run in real time Lots of code available Sanja Fidler CSC420: Intro to Image Understanding 14 / 58

Examples Figure: Matching in day / night under viewpoint change [Source: S. Seitz] Sanja Fidler CSC420: Intro to Image Understanding 15 / 58

Example [Source: N. Snavely] Figure: NASA Mars Rover images with SIFT feature matches Sanja Fidler CSC420: Intro to Image Understanding 16 / 58

PCA-SIFT The dimensionality of SIFT is pretty high, i.e., 128D for each keypoint Reduce the dimensionality using linear dimensionality reduction In this case, principal component analysis (PCA) Use 10D or so descriptor [Source: R. Urtasun] Sanja Fidler CSC420: Intro to Image Understanding 17 / 58

Gradient location-orientation histogram (GLOH) Developed by Mikolajczyk and Schmid (2005): variant of SIFT that uses a log-polar binning structure instead of the four quadrants. The spatial bins are 11, and 15, with eight angular bins (except for the central region), for a total of 17 spatial bins and 16 orientation bins. The 272D histogram is then projected onto a 128D descriptor using PCA trained on a large database. [Source: R. Szeliski] Sanja Fidler CSC420: Intro to Image Understanding 18 / 58

Other Descriptors SURF DAISY LBP HOG Shape Contexts Color Histograms Sanja Fidler CSC420: Intro to Image Understanding 19 / 58

[Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 20 / 58 Local Features Detection: Identify the interest points. Description: Extract feature descriptor around each interest point. Matching: Determine correspondence between descriptors in two views.

Image Features: Matching the Local Descriptors Sanja Fidler CSC420: Intro to Image Understanding 21 / 58

Matching the Local Descriptors Once we have extracted keypoints and their descriptors, we want to match the features between pairs of images. Ideally a match is a correspondence between a local part of the object on one image to the same local part of the object in another image How should we compute a match? Figure: Images from K. Grauman Sanja Fidler CSC420: Intro to Image Understanding 22 / 58

Matching the Local Descriptors Once we have extracted keypoints and their descriptors, we want to match the features between pairs of images. Ideally a match is a correspondence between a local part of the object on one image to the same local part of the object in another image How should we compute a match? Figure: Images from K. Grauman Sanja Fidler CSC420: Intro to Image Understanding 22 / 58

Matching the Local Descriptors Simple: Compare them all, compute Euclidean distance Sanja Fidler CSC420: Intro to Image Understanding 23 / 58

Matching the Local Descriptors Simple: Compare them all, compute Euclidean distance Sanja Fidler CSC420: Intro to Image Understanding 23 / 58

Matching the Local Descriptors Find closest match (min distance). How do we know if match is reliable? Sanja Fidler CSC420: Intro to Image Understanding 23 / 58

Matching the Local Descriptors Find also the second closest match. Match reliable if first distance much smaller than second distance Sanja Fidler CSC420: Intro to Image Understanding 23 / 58

Matching the Local Descriptors Compute the ratio: where f 0 i f 0 i = f i i f i f 0 i is the closest and f 0 i second closest match to f i. Sanja Fidler CSC420: Intro to Image Understanding 23 / 58

Figure: Images from R. Szeliski Sanja Fidler CSC420: Intro to Image Understanding 24 / 58 Which Threshold to Use? Setting the threshold too high results in too many false positives, i.e., incorrect matches being returned. Setting the threshold too low results in too many false negatives, i.e., too many correct matches being missed

Which Threshold to Use? Threshold ratio of nearest to 2nd nearest descriptor Typically: i < 0.8 [Source: K. Grauman] Figure: Images from D. Lowe Sanja Fidler CSC420: Intro to Image Understanding 25 / 58

Applications of Local Invariant Features Wide baseline stereo Motion tracking Panorama stitching Mobile robot navigation 3D reconstruction Recognition Retrieval [Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 26 / 58

Wide Baseline Stereo [Source: T. Tuytelaars] Sanja Fidler CSC420: Intro to Image Understanding 27 / 58

Recognizing the Same Object [Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 28 / 58

Motion Tracking Figure: Images from J. Pilet Sanja Fidler CSC420: Intro to Image Understanding 29 / 58

Now What Now we know how to extract scale and rotation invariant features We even know how to match features across images Can we use this to find Waldo in an even more sneaky scenario? Sanja Fidler CSC420: Intro to Image Understanding 30 / 58

Now What Now we know how to extract scale and rotation invariant features We even know how to match features across images Can we use this to find Waldo in an even more sneaky scenario? template Waldo on the road Sanja Fidler CSC420: Intro to Image Understanding 30 / 58

Now What Now we know how to extract scale and rotation invariant features We even know how to match features across images Can we use this to find Waldo in an even more sneaky scenario? template He comes closer... We know how to solve this Sanja Fidler CSC420: Intro to Image Understanding 30 / 58

Now What Now we know how to extract scale and rotation invariant features We even know how to match features across images Can we use this to find Waldo in an even more sneaky scenario? template Someone takes a (weird) picture of him! Sanja Fidler CSC420: Intro to Image Understanding 30 / 58

Find My DVD! More interesting: If we have DVD covers (e.g., from Amazon), can we match them to DVDs in real scenes? Sanja Fidler CSC420: Intro to Image Understanding 31 / 58

Matching Planar Objects In New Viewpoints Sanja Fidler CSC420: Intro to Image Understanding 32 / 58

What Kind of Transformation Happened To My DVD? Sanja Fidler CSC420: Intro to Image Understanding 33 / 58

What Kind of Transformation Happened To My DVD? Rectangle goes to a parallelogram (almost but not really, but let s believe that for now) Sanja Fidler CSC420: Intro to Image Understanding 34 / 58

All 2D Linear Transformations Linear transformations are combinations of Scale, Rotation Shear Mirror apple x 0 y 0 = apple a c b d apple x y [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 35 / 58

All 2D Linear Transformations Properties of linear transformations: Origin maps to origin Lines map to lines Sanja Fidler CSC420: Intro to Image Understanding 36 / 58

All 2D Linear Transformations Properties of linear transformations: Origin maps to origin Lines map to lines Parallel lines remain parallel Sanja Fidler CSC420: Intro to Image Understanding 36 / 58

All 2D Linear Transformations Properties of linear transformations: Origin maps to origin Lines map to lines Parallel lines remain parallel Ratios are preserved Sanja Fidler CSC420: Intro to Image Understanding 36 / 58

All 2D Linear Transformations Properties of linear transformations: Origin maps to origin Lines map to lines Parallel lines remain parallel Ratios are preserved Closed under composition Sanja Fidler CSC420: Intro to Image Understanding 36 / 58

All 2D Linear Transformations Properties of linear transformations: Origin maps to origin Lines map to lines Parallel lines remain parallel Ratios are preserved Closed under composition apple x 0 y 0 = apple a c b d apple e g f h apple i k j l apple x y Sanja Fidler CSC420: Intro to Image Understanding 36 / 58

All 2D Linear Transformations Properties of linear transformations: Origin maps to origin Lines map to lines Parallel lines remain parallel Ratios are preserved Closed under composition apple x 0 y 0 = apple a c b d apple e g f h apple i k j l apple x y What about the translation? [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 36 / 58

All 2D Linear Transformations Properties of linear transformations: Origin maps to origin Lines map to lines Parallel lines remain parallel Ratios are preserved Closed under composition apple x 0 y 0 = apple a c b d apple e g f h apple i k j l apple x y What about the translation? [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 36 / 58

A A ne Transformations ne transformations are combinations of Linear transformations, and Translations apple x 0 y 0 = same as: apple a c b d apple apple x 0 a b e y 0 = c d f apple x apple e y + f 2 3 x 4y5 1 Sanja Fidler CSC420: Intro to Image Understanding 37 / 58

A A ne Transformations ne transformations are combinations of Linear transformations, and Translations apple x 0 y 0 = same as: apple a c b d apple apple x 0 a b e y 0 = c d f apple x apple e y + f 2 3 x 4y5 1 Sanja Fidler CSC420: Intro to Image Understanding 37 / 58

A A ne Transformations ne transformations are combinations of Linear transformations, and Translations apple apple x 0 a b e y 0 = c d f 2 3 x 4y5 1 Properties of a ne transformations: Origin does not necessarily map to origin Lines map to lines Parallel lines remain parallel Ratios are preserved Closed under composition Rectangles go to parallelograms [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 38 / 58

2D Image Tranformations These transformations are a nested set of groups Closed under composition and inverse is a member Sanja Fidler CSC420: Intro to Image Understanding 39 / 58

What Transformation Happened to My DVD? A ne transformation approximates viewpoint changes for roughly planar objects and roughly orthographic cameras (more about these later in class) DVD went a ne! Sanja Fidler CSC420: Intro to Image Understanding 40 / 58

Computing the (A ne) Transformation Given a set of matches between images I and J How can we compute the a ne transformation A from I to J? Find transform A that best agrees with the matches [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 41 / 58

Computing the (A ne) Transformation Given a set of matches between images I and J How can we compute the a ne transformation A from I to J? Find transform A that best agrees with the matches [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 41 / 58

Computing the A ne Transformation Let (x i, y i ) be a point on the reference (model) image, and (x 0 i, y 0 i )itsmatch in the test image An a ne transformation A maps (x i, y i )to(xi 0, y i 0 2 ): apple apple x 0 i a b e yi 0 = 4 x 3 i y c d f i 5 1 Sanja Fidler CSC420: Intro to Image Understanding 42 / 58

Computing the A ne Transformation Let (x i, y i ) be a point on the reference (model) image, and (x 0 i, y 0 i )itsmatch in the test image An a ne transformation A maps (x i, y i )to(xi 0, y i 0 2 ): apple apple x 0 i a b e yi 0 = 4 x 3 i y c d f i 5 1 We can rewrite this into a simple linear system: 2 3 a apple b xi y i 0 0 1 0 c 0 0 x i y i 0 1 6 d = 7 4 e 5 f apple x 0 i y 0 i Sanja Fidler CSC420: Intro to Image Understanding 42 / 58

Computing the A ne Transformation But we have many matches: 2 6 6 6 6 4.. x i y i 0 0 1 0 0 0 x i y i 0 1. 3 7 7 7 7 5 {z } P 2 6 6 6 6 6 6 4 a b c d e f 3 7 7 7 7 7 7 5 {z } a = 2 6 6 6 6 4.. x 0 i y 0 i. 3 7 7 7 7 5 {z } P 0 For each match we have two more equations Sanja Fidler CSC420: Intro to Image Understanding 43 / 58

Computing the A ne Transformation But we have many matches: 2 6 6 6 6 4.. x i y i 0 0 1 0 0 0 x i y i 0 1. 3 7 7 7 7 5 {z } P 2 6 6 6 6 6 6 4 a b c d e f 3 7 7 7 7 7 7 5 {z } a = 2 6 6 6 6 4.. x 0 i y 0 i. 3 7 7 7 7 5 {z } P 0 For each match we have two more equations How many matches do we need to compute A? Sanja Fidler CSC420: Intro to Image Understanding 43 / 58

Computing the A ne Transformation But we have many matches: 2 6 6 6 6 4.. x i y i 0 0 1 0 0 0 x i y i 0 1. 3 7 7 7 7 5 {z } P 2 6 6 6 6 6 6 4 a b c d e f 3 7 7 7 7 7 7 5 {z } a = 2 6 6 6 6 4.. x 0 i y 0 i. 3 7 7 7 7 5 {z } P 0 For each match we have two more equations How many matches do we need to compute A? Sanja Fidler CSC420: Intro to Image Understanding 43 / 58

Computing the A ne Transformation But we have many matches: 2 6 4.. x i y i 0 0 1 0 0 0 x i y i 0 1 2 3 7 5 6 4. {z } P For each match we have two more equations a b c d e f {z } a How many matches do we need to compute A? 6parameters! 3matches But the more, the better (more reliable) How do we compute A? 3 2 = 6 7 4 5.. x 0 i y 0 i 3 7 5. {z } P 0 Sanja Fidler CSC420: Intro to Image Understanding 43 / 58

Computing the A ne Transformation 2 6 6 6 6 4. x i y i 0 0 1 0 0 0 x i y i 0 1. 3 7 7 7 7 5 {z } P 2 6 6 6 6 6 6 4 a b c d e f 3 7 7 7 7 7 7 5 {z } a = 2 6 6 6 6 4. x 0 i y 0 i. 3 7 7 7 7 5 {z } P 0 If we have 3 matches, then computing A is really easy: a = P 1 P 0 Sanja Fidler CSC420: Intro to Image Understanding 44 / 58

Computing the A ne Transformation 2 6 4. x i y i 0 0 1 0 0 0 x i y i 0 1 2 3 7 5 6 4. {z } P a b c d e f {z } a 3 2 = 6 7 4 5. xi 0 yi 0 3 7 5. {z } P 0 If we have 3 matches, then computing A is really easy: a = P 1 P 0 If we have more than 3, then we do least-squares estimation: min a,b,,f Pa P0 2 2 Sanja Fidler CSC420: Intro to Image Understanding 44 / 58

Computing the A ne Transformation 2 6 4. x i y i 0 0 1 0 0 0 x i y i 0 1 2 3 7 5 6 4. {z } P a b c d e f {z } a 3 2 = 6 7 4 5. xi 0 yi 0 3 7 5. {z } P 0 If we have 3 matches, then computing A is really easy: a = P 1 P 0 If we have more than 3, then we do least-squares estimation: Which has a closed form solution: min a,b,,f Pa P0 2 2 a =(P T P) 1 P T P 0 Sanja Fidler CSC420: Intro to Image Understanding 44 / 58

Image Alignment Algorithm: A ne Case Given images I and J 1 Compute image features for I and J 2 Match features between I and J 3 Compute a ne transformation A between I and J using least squares on the set of matches Sanja Fidler CSC420: Intro to Image Understanding 45 / 58

Image Alignment Algorithm: A ne Case Given images I and J 1 Compute image features for I and J 2 Match features between I and J 3 Compute a ne transformation A between I and J using least squares on the set of matches Is there a problem with this? [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 45 / 58

Robustness!"#$%&'() %*$%&'() [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 46 / 58

Simple Case (This example is unrelated to the object matching example, but it nicely shows how to robustify estimation) Lets consider a simpler example... Fit a line to the points below!!"#$%&'()*+,)-)%+.&),#),/&0&)1-,-2#+.,0) 3&-0,)045-"&0)6,) Sanja Fidler CSC420: Intro to Image Understanding 47 / 58

Simple Case (This example is unrelated to the object matching example, but it nicely shows how to robustify estimation) Lets consider a simpler example... Fit a line to the points below!!"#$%&'()*+,)-)%+.&),#),/&0&)1-,-2#+.,0) 3&-0,)045-"&0)6,) How can we fix this? [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 47 / 58

Simple Idea: RANSAC Take the minimal number of points to compute what we want. In the line example, two points (in our a ne example, three matches) Sanja Fidler CSC420: Intro to Image Understanding 48 / 58

Simple Idea: RANSAC Take the minimal number of points to compute what we want. In the line example, two points (in our a ne example, three matches) By take we mean choose at random from all points Sanja Fidler CSC420: Intro to Image Understanding 48 / 58

Simple Idea: RANSAC Take the minimal number of points to compute what we want. In the line example, two points (in our a ne example, three matches) By take we mean choose at random from all points Fit a line to the selected pair of points Sanja Fidler CSC420: Intro to Image Understanding 48 / 58

Simple Idea: RANSAC Take the minimal number of points to compute what we want. In the line example, two points (in our a ne example, three matches) By take we mean choose at random from all points Fit a line to the selected pair of points Count the number of all points that agree with the line: We call the agreeing points inliers Sanja Fidler CSC420: Intro to Image Understanding 48 / 58

Simple Idea: RANSAC Take the minimal number of points to compute what we want. In the line example, two points (in our a ne example, three matches) By take we mean choose at random from all points Fit a line to the selected pair of points Count the number of all points that agree with the line: We call the agreeing points inliers Agree = within a small distance of the line Sanja Fidler CSC420: Intro to Image Understanding 48 / 58

Simple Idea: RANSAC Take the minimal number of points to compute what we want. In the line example, two points (in our a ne example, three matches) By take we mean choose at random from all points Fit a line to the selected pair of points Count the number of all points that agree with the line: We call the agreeing points inliers Agree = within a small distance of the line Repeat this many times, remember the number of inliers for each trial Among several trials, select the one with the largest number of inliers This procedure is called RAndom SAmple Consensus Sanja Fidler CSC420: Intro to Image Understanding 48 / 58

RANSAC for Line Fitting Example 1 Randomly select minimal subset of points [Source: R. Raguram] Sanja Fidler CSC420: Intro to Image Understanding 49 / 58

RANSAC for Line Fitting Example 1 Randomly select minimal subset of points 2 Hypothesize a model [Source: R. Raguram] Sanja Fidler CSC420: Intro to Image Understanding 49 / 58

RANSAC for Line Fitting Example 1 Randomly select minimal subset of points 2 Hypothesize a model 3 Compute error function [Source: R. Raguram] Sanja Fidler CSC420: Intro to Image Understanding 49 / 58

RANSAC for Line Fitting Example 1 Randomly select minimal subset of points 2 Hypothesize a model 3 Compute error function 4 Select points consistent with model [Source: R. Raguram] Sanja Fidler CSC420: Intro to Image Understanding 49 / 58

RANSAC for Line Fitting Example 1 Randomly select minimal subset of points 2 Hypothesize a model 3 Compute error function 4 Select points consistent with model 5 Repeat hypothesize and verify loop [Source: R. Raguram] Sanja Fidler CSC420: Intro to Image Understanding 49 / 58

RANSAC for Line Fitting Example 1 Randomly select minimal subset of points 2 Hypothesize a model 3 Compute error function 4 Select points consistent with model 5 Repeat hypothesize and verify loop [Source: R. Raguram] Sanja Fidler CSC420: Intro to Image Understanding 49 / 58

RANSAC for Line Fitting Example 1 Randomly select minimal subset of points 2 Hypothesize a model 3 Compute error function 4 Select points consistent with model 5 Repeat hypothesize and verify loop 6 Choose model with largest set of inliers [Source: R. Raguram] Sanja Fidler CSC420: Intro to Image Understanding 49 / 58

Translations [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 50 / 58

RAndom SAmple Consensus!"#"$%&!"#&'(%$)&(%&*(+,-'.&$-/+%&$"%$#&'& [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 51 / 58

RAndom SAmple Consensus!"#"$%&'()%*"+&,'%$*&'%&+'(-),.&$)/(%&!"#!$%&& [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 51 / 58

RAndom SAmple Consensus!"#$"#%#&'%#()*+,)-.*%/0#&%#&'%&01&'+#%*"23'(%.4%0*,0'(+% [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 51 / 58

RANSAC All the inliers will agree with each other on the translation vector; the (hopefully small) number of outliers will (hopefully) disagree with each other Sanja Fidler CSC420: Intro to Image Understanding 52 / 58

RANSAC All the inliers will agree with each other on the translation vector; the (hopefully small) number of outliers will (hopefully) disagree with each other RANSAC only has guarantees if there are < 50% outliers Sanja Fidler CSC420: Intro to Image Understanding 52 / 58

RANSAC All the inliers will agree with each other on the translation vector; the (hopefully small) number of outliers will (hopefully) disagree with each other RANSAC only has guarantees if there are < 50% outliers All good matches are alike; every bad match is bad in its own way. [Tolstoy via Alyosha Efros] [Source: N. Snavely] Sanja Fidler CSC420: Intro to Image Understanding 52 / 58

A ne Transformation How? Sanja Fidler CSC420: Intro to Image Understanding 53 / 58

A ne Transformation How? Find matches across images I and J. This gives us points X I in image I and X J in J, where we know that the point XI k is a match with XJ k Iterate: Choose 3 pairs of matches randomly Compute the a ne transformation Project all matched points X I from I to J via the computed transformation. This gives us ˆX I Find how many matches are inliers, i.e., ˆX I k XJ k < thresh. Choose the transformation with the most inliers Sanja Fidler CSC420: Intro to Image Understanding 53 / 58

RANSAC Inlier threshold related to the amount of noise we expect in inliers Sanja Fidler CSC420: Intro to Image Understanding 54 / 58

RANSAC Inlier threshold related to the amount of noise we expect in inliers Often model noise as Gaussian with some standard deviation (e.g., 3 pixels) Sanja Fidler CSC420: Intro to Image Understanding 54 / 58

RANSAC Inlier threshold related to the amount of noise we expect in inliers Often model noise as Gaussian with some standard deviation (e.g., 3 pixels) Number of rounds related to the percentage of outliers we expect, and the probability of success we d like to guarantee Sanja Fidler CSC420: Intro to Image Understanding 54 / 58

RANSAC Inlier threshold related to the amount of noise we expect in inliers Often model noise as Gaussian with some standard deviation (e.g., 3 pixels) Number of rounds related to the percentage of outliers we expect, and the probability of success we d like to guarantee Suppose there are 20% outliers, and we want to find the correct answer with 99% probability Sanja Fidler CSC420: Intro to Image Understanding 54 / 58

RANSAC Inlier threshold related to the amount of noise we expect in inliers Often model noise as Gaussian with some standard deviation (e.g., 3 pixels) Number of rounds related to the percentage of outliers we expect, and the probability of success we d like to guarantee Suppose there are 20% outliers, and we want to find the correct answer with 99% probability How many rounds do we need? [Source: R. Urtasun] Sanja Fidler CSC420: Intro to Image Understanding 54 / 58

How many rounds? Su cient number of trials S must be tried. Sanja Fidler CSC420: Intro to Image Understanding 55 / 58

How many rounds? Su cient number of trials S must be tried. Let p be the probability that any given correspondence is valid and P be the total probability of success after S trials. Sanja Fidler CSC420: Intro to Image Understanding 55 / 58

How many rounds? Su cient number of trials S must be tried. Let p be the probability that any given correspondence is valid and P be the total probability of success after S trials. The likelihood in one trial that all k random samples are inliers is p k Sanja Fidler CSC420: Intro to Image Understanding 55 / 58

How many rounds? Su cient number of trials S must be tried. Let p be the probability that any given correspondence is valid and P be the total probability of success after S trials. The likelihood in one trial that all k random samples are inliers is p k The likelihood that S such trials will all fail is 1 P =(1 p k ) S Sanja Fidler CSC420: Intro to Image Understanding 55 / 58

How many rounds? Su cient number of trials S must be tried. Let p be the probability that any given correspondence is valid and P be the total probability of success after S trials. The likelihood in one trial that all k random samples are inliers is p k The likelihood that S such trials will all fail is 1 P =(1 p k ) S The required minimum number of trials is S = log(1 P) log(1 p k ) Sanja Fidler CSC420: Intro to Image Understanding 55 / 58

How many rounds? Su cient number of trials S must be tried. Let p be the probability that any given correspondence is valid and P be the total probability of success after S trials. The likelihood in one trial that all k random samples are inliers is p k The likelihood that S such trials will all fail is 1 P =(1 p k ) S The required minimum number of trials is S = log(1 P) log(1 p k ) The number of trials grows quickly with the number of sample points used. Sanja Fidler CSC420: Intro to Image Understanding 55 / 58

How many rounds? Su cient number of trials S must be tried. Let p be the probability that any given correspondence is valid and P be the total probability of success after S trials. The likelihood in one trial that all k random samples are inliers is p k The likelihood that S such trials will all fail is 1 P =(1 p k ) S The required minimum number of trials is S = log(1 P) log(1 p k ) The number of trials grows quickly with the number of sample points used. [Source: R. Urtasun] Sanja Fidler CSC420: Intro to Image Understanding 55 / 58

RANSAC pros and cons Pros Simple and general Sanja Fidler CSC420: Intro to Image Understanding 56 / 58

RANSAC pros and cons Pros Simple and general Applicable to many di erent problems Sanja Fidler CSC420: Intro to Image Understanding 56 / 58

RANSAC pros and cons Pros Simple and general Applicable to many di erent problems Often works well in practice Sanja Fidler CSC420: Intro to Image Understanding 56 / 58

RANSAC pros and cons Pros Simple and general Applicable to many di erent problems Often works well in practice Cons Parameters to tune Sanja Fidler CSC420: Intro to Image Understanding 56 / 58

RANSAC pros and cons Pros Simple and general Applicable to many di erent problems Often works well in practice Cons Parameters to tune Sometimes too many iterations are required Sanja Fidler CSC420: Intro to Image Understanding 56 / 58

RANSAC pros and cons Pros Simple and general Applicable to many di erent problems Often works well in practice Cons Parameters to tune Sometimes too many iterations are required Can fail for extremely low inlier ratios Sanja Fidler CSC420: Intro to Image Understanding 56 / 58

RANSAC pros and cons Pros Simple and general Applicable to many di erent problems Often works well in practice Cons Parameters to tune Sometimes too many iterations are required Can fail for extremely low inlier ratios We can often do better than brute-force sampling [Source: N. Snavely, slide credit: R. Urtasun] Sanja Fidler CSC420: Intro to Image Understanding 56 / 58

Ransac Verification [Source: K. Grauman, slide credit: R. Urtasun] Sanja Fidler CSC420: Intro to Image Understanding 57 / 58

Summary Stu You Need To Know To match image I and J under a ne transformation: Compute scale and rotation invariant keypoints in both images Compute a (rotation invariant) feature vector in each keypoint (e.g., SIFT) Match all features in I to all features in J For each feature in reference image I find closest match in J If ratio between closest and second closest match is < 0.8, keep match Do RANSAC to compute a ne transformation A: Select 3 matches at random Compute A Compute the number of inliers Repeat Find A that gave the most inliers Sanja Fidler CSC420: Intro to Image Understanding 58 / 58