FAST-MATCH: FAST AFFINE TEMPLATE MATCHING

Similar documents
FAsT-Match: Fast Affine Template Matching

FAsT-Match: Fast Affine Template Matching

Chapter 3 Image Registration. Chapter 3 Image Registration

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

Comparison of Local Feature Descriptors

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

Requirements for region detection

Feature Detectors and Descriptors: Corners, Lines, etc.

Feature Based Registration - Image Alignment

The Lucas & Kanade Algorithm

Dense Image-based Motion Estimation Algorithms & Optical Flow

Coarse-to-fine image registration

Lucas-Kanade Image Registration Using Camera Parameters

Lecture 16: Computer Vision

Lecture 16: Computer Vision

Homographies and RANSAC

Visual Tracking (1) Pixel-intensity-based methods

Midterm Examination CS 534: Computational Photography

Agenda. Rotations. Camera calibration. Homography. Ransac

CS231A Section 6: Problem Set 3

Object Recognition with Invariant Features

Optical Flow-Based Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides.

arxiv: v1 [cs.cv] 28 Sep 2018

Combining Appearance and Topology for Wide

Stereo and Epipolar geometry

Representing Moving Images with Layers. J. Y. Wang and E. H. Adelson MIT Media Lab

The SIFT (Scale Invariant Feature

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

Prof. Feng Liu. Spring /26/2017

Lecture 8 Fitting and Matching

CS201: Computer Vision Introduction to Tracking

EE795: Computer Vision and Intelligent Systems

Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features

Motion Estimation and Optical Flow Tracking

Computer Vision for HCI. Topics of This Lecture

Comparison of Viewpoint-Invariant Template Matchings

Fast Affine Template Matching over Galois Field

Motion Estimation. There are three main types (or applications) of motion estimation:

Discovering Visual Hierarchy through Unsupervised Learning Haider Razvi

Fast Natural Feature Tracking for Mobile Augmented Reality Applications

CS443: Digital Imaging and Multimedia Binary Image Analysis. Spring 2008 Ahmed Elgammal Dept. of Computer Science Rutgers University

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Image stitching. Announcements. Outline. Image stitching

Motion Tracking and Event Understanding in Video Sequences

Observations. Basic iteration Line estimated from 2 inliers

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Local Feature Detectors

Automatic Image Alignment (feature-based)

Tight Approximation of Image Matching

Final Exam Study Guide

COMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE

CSE 252B: Computer Vision II

Visual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania.

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

Image Stitching. Slides from Rick Szeliski, Steve Seitz, Derek Hoiem, Ira Kemelmacher, Ali Farhadi

Targil 10 : Why Mosaic? Why is this a challenge? Exposure differences Scene illumination Miss-registration Moving objects

Computer Vision I. Announcements. Random Dot Stereograms. Stereo III. CSE252A Lecture 16

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac

arxiv: v1 [cs.cv] 2 May 2016

Computer Vision. Recap: Smoothing with a Gaussian. Recap: Effect of σ on derivatives. Computer Science Tripos Part II. Dr Christopher Town

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac

Lecture 28 Intro to Tracking

Recall: Blob Merge/Split Lecture 28

Visual Tracking (1) Feature Point Tracking and Block Matching

Local invariant features

Eppur si muove ( And yet it moves )

Instance-level recognition

Image matching. Announcements. Harder case. Even harder case. Project 1 Out today Help session at the end of class. by Diva Sian.

AK Computer Vision Feature Point Detectors and Descriptors

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

A novel point matching method for stereovision measurement using RANSAC affine transformation

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Lecture 9 Fitting and Matching

Image processing and features

Comparison between Motion Analysis and Stereo

Review for the Final

Introduction to Computer Vision

CSE 527: Introduction to Computer Vision

INFO0948 Fitting and Shape Matching

Filtering Images. Contents

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

CS-465 Computer Vision

Fitting. Instructor: Jason Corso (jjcorso)! web.eecs.umich.edu/~jjcorso/t/598f14!! EECS Fall 2014! Foundations of Computer Vision!

Global Flow Estimation. Lecture 9

Lecture: RANSAC and feature detectors

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --

Model Fitting. Introduction to Computer Vision CSE 152 Lecture 11

Occluded Facial Expression Tracking

Ruch (Motion) Rozpoznawanie Obrazów Krzysztof Krawiec Instytut Informatyki, Politechnika Poznańska. Krzysztof Krawiec IDSS

Notes 9: Optical Flow

Visual Tracking. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

CS 231A Computer Vision (Winter 2014) Problem Set 3

CS231A Midterm Review. Friday 5/6/2016

Edge Detection. CS664 Computer Vision. 3. Edges. Several Causes of Edges. Detecting Edges. Finite Differences. The Gradient

Fitting. Fitting. Slides S. Lazebnik Harris Corners Pkwy, Charlotte, NC

2D Image Processing Feature Descriptors

Agenda. Rotations. Camera models. Camera calibration. Homographies

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Transcription:

Seminar on Sublinear Time Algorithms FAST-MATCH: FAST AFFINE TEMPLATE MATCHING KORMAN, S., REICHMAN, D., TSUR, G., & AVIDAN, S., 2013 Given by: Shira Faigenbaum-Golovin Tel-Aviv University 27.12.2015

Problem Definition Image processing in a nutshell Prior Art of image alignment Suggested algorithm

Image matching: Given two grayscale images, I1 and I2 Find affine transformation T that maps pixels from I1 to pixels to I2. So that the difference over pixels p between I1(T(p)) and I2(p) is minimized

Find the best transformation between two given images:

Some results I

Some results III

Align two images before comparison Align for image enhancement Panoramic mosaics. Match images in a video sequence

Gray scale image I is an nxm matrix with values between [0,1]. where 0 is black, and 1 is white. The intermediate values are the gray levels

A pixel p in an nxn image I is a pair (x,y)in {1,,n}². A value of a pixel p=(x,y) in an image I is I(x,y). Two different pixels p =(x,y) and q =(x,y ) are adjacent if x-x 1 and y-y 1. A pixel p is boundary in an image I if there is an adjacent pixel q s.t I(p) I(q).

An affine transformation matrix T can be decomposed into T( I ) Tr R S R I 1 2 1 1 where Tr, R, S are translation, rotation and non-uniform scaling matrices. There exist 6 degrees of freedom: a rotation angle, x and y scales, another rotation angle and x and y translations.

Given two grayscale images, I1 and I2 and affine transformation T :I1 I2. We define a sum of absolute differences (SAD) and 1 dt ( I1, I2) 2 I1( T ( p)) I2( p) n p I1 T ( p) I2 The optimal transformation satisfies: d( I, I ) min d ( I, I ) 1 2 T 1 2 T

The algorithm: 1. Take a sample of the Affine transformations 2. Evaluate the SAD for each transformation in the sample 3. Return the best Questions: Which transformations to use? How does is guarantee a bound?

Direct methods parametric OF Indirect methods (feature based)

Lucas, Kanade An iterative image registration technique with an application to stereo vision [ICAI 1981] Baker, Matthews Lucas-Kanade 20 years on: A unifying framework [IJCV 04]

Enumerate ~ n¹⁸ affine transformation (for nxn images)

e.g. SIFT Computational complexity 2 ( n )

Select (at random) a subset of k pairs Compute a motion estimate T By using least squares, to minimize the sum of squared residuals. Counts the number of inliers that are within ε of their predicted location The random selection process is repeated m times, The sample set with largest number of inliers is kept as the final solution Computational complexity ( k( t mn)) k=#samples N=#data points t= time of single model m=avg # of models per sample

Lowe Distinctive image features from scale-invariant key-points [IJCV 04] Morel, Yu Asift: A new framework for fully affine invariant image comparison [SIAM 09] M.A. Fichler, R.C. Bolles Random sample consensus [Comm. of ACM 81]

template image Transformation space (e.g. affine) Observation: Due to image smoothness assumption, the SAD measure will not change significantly, when small variations in the parameters of the transformation

I1 ( n1 n1) I 2 2 I1 T I 1 ( I, I ) I ( T( p)) I ( p) T 1 2 2 1 2 n1 p I 1 T ( I1, I2) min T ( I1, I2) T T * I, I ) ( I, ) ( 1 2 T* 1 I2

Given two transformations T and T, we define l ( T, T ') max T( p) T '( p) p I 1 2 l ( T, T ') as: I1 2 T I For a positive α, a net of (affine) transformations τ={ti} is an α-cover if In our case α=δn1. T T l ( T, T) O( ) Claim: We can construct a net of transformation T Aδ And prove that it is δn1-cover i Ti i T '

An affine transformation matrix T can be decomposed into T( I ) Tr R S R I 1 2 1 1 There exist 6 degrees of freedom: a rotation angle, x and y scales, another rotation angle and x and y translations. The basic idea is to discretize the space of Affine transformations, by dividing each of the dimensions into Θ(δ) equal segments, such that for any two consecutive transformations T and T on any of the dimensions it will hold that l ( T,T ') ( n ) 1

1 n2 2 ( 6 ( n ) ) 1

Input: Grayscale images I1 and I2, a precision parameter δ and a transformation T Output: An estimate of the distance Sample m = Θ(1/ δ²) values of pixels p1 pm in I1 Return Claim: Given images I1 and I2 and an affine transformation T, the algorithm returns a value dt such that dt - < δ with probability 2/3. It performs (1/ δ²) samples. Estimate the SAD to within O(1/ δ²)

1 n2 2 ( 6 ( n ) ) 1 O(1/ δ²) Total runtime is: 2 A (1/ ) 1 n2 2 ( ( ) ) 8 n1

Achieving a satisfactory error rate would require using a net Nδ where δ is small. Thus causing the execution time to grow. Therefore branch-and-bound scheme is used, by running the algorithm on subset of the net Nδ. and refining the δ parameter.

Pascal VOC 2010 data-set 200 random image/templates Template dimensions of 10%, 30%, 50%, 70%, 90% Comparison to a feature-based method - ASIFT Image degradations (template left in-tact): Gaussian Blur with STD of {0,1,2,4,7,11} pixels Gaussian Noise with STD of {0,5,10,18,28,41} JPEG compression of quality {75,40,20,10,5,2}

Pascal VOC 2010 data-set 200 random image/templates Template dimensions of 10%, 30%, 50%, 70%, 90% Comparison to a feature-based method - ASIFT Image degradations (template left in-tact): Gaussian Blur with STD of {0,1,2,4,7,11} pixels Gaussian Noise with STD of {0,5,10,18,28,41} JPEG compression of quality {75,40,20,10,5,2}

Fast-Match vs. ASIFT template dimension 50%

Fast-Match vs. ASIFT template dimension 20%

Runtimes

Template Dim: 45%

Template Dim: 35%

Template Dim: 25%

Template Dim: 15%

Template Dim: 10%

Experiment 2: Varying conditions Mikolajczyk data-set (for features and descriptors) 8 sequences of 6 images, with increasingly harsh conditions Including: Zoom+Rotation (bark) Blur (bikes) Zoom+rotation (boat) Viewpoint change (graffiti) Brightness change (light) Blur (trees) Jpeg compression (UBC) Viewpoint change (wall)

Experiment 3: Matching in real-world scenes The Zurich Building Data-set 200 buildings, 5 different views each 200 random instances Random choice of building, 2 views, template in one view We seek the best possible affine transformation In most cases homography or non-rigid is needed Results: 129 cases - good matches 40 cases template doesn t appear in second image 12 cases bad occlusion of template in second image 19 cases failure (none of the above)

Experiment 3: Good cases

Experiment 3: Good cases

Experiment 3: failures, occlusions, out of img.

Handles template matching under arbitrary Affine (6 dof) transformations with Guaranteed error bounds Fast execution Main ingredients Sampling of transformation space (based on variation) Quick transformation evaluation ( property testing ) Branch-and-Bound scheme

Limitations Smoothness assumption Global transformation Partial matching Extensions Higher dimensions - Matching 3D shapes Other registration problems Symmetry detection

Thank you for your attention