Introduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.

Similar documents
SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

CAP 5415 Computer Vision Fall 2012

The SIFT (Scale Invariant Feature

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

Local Features: Detection, Description & Matching

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Scale Invariant Feature Transform

Scale Invariant Feature Transform

Key properties of local features

School of Computing University of Utah

Local Features Tutorial: Nov. 8, 04

Outline 7/2/201011/6/

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Scale Invariant Feature Transform by David Lowe

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Patch-based Object Recognition. Basic Idea

Implementing the Scale Invariant Feature Transform(SIFT) Method

Computer Vision for HCI. Topics of This Lecture

Distinctive Image Features from Scale-Invariant Keypoints

Obtaining Feature Correspondences

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

Distinctive Image Features from Scale-Invariant Keypoints

Object Recognition with Invariant Features

SIFT - scale-invariant feature transform Konrad Schindler

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

Verslag Project beeldverwerking A study of the 2D SIFT algorithm

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

Distinctive Image Features from Scale-Invariant Keypoints

SIFT: Scale Invariant Feature Transform

Motion Estimation and Optical Flow Tracking

Local features: detection and description. Local invariant features

Local Feature Detectors

Local features: detection and description May 12 th, 2015

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

Image Matching. AKA: Image registration, the correspondence problem, Tracking,

CS 558: Computer Vision 4 th Set of Notes

Feature Based Registration - Image Alignment

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability

Ulas Bagci

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Advanced Video Content Analysis and Video Compression (5LSH0), Module 4

Image Features: Detection, Description, and Matching and their Applications

Scale Invariant Feature Transform (SIFT) CS 763 Ajit Rajwade

Eppur si muove ( And yet it moves )

2D Image Processing Feature Descriptors

3D Reconstruction From Multiple Views Based on Scale-Invariant Feature Transform. Wenqi Zhu

Local invariant features

TA Section 7 Problem Set 3. SIFT (Lowe 2004) Shape Context (Belongie et al. 2002) Voxel Coloring (Seitz and Dyer 1999)

Local features and image matching. Prof. Xin Yang HUST

A Novel Algorithm for Color Image matching using Wavelet-SIFT

Implementation and Comparison of Feature Detection Methods in Image Mosaicing

Feature Detection and Matching

CS231A Section 6: Problem Set 3

Corner Detection. GV12/3072 Image Processing.

Other Linear Filters CS 211A

SURF. Lecture6: SURF and HOG. Integral Image. Feature Evaluation with Integral Image

Computer Vision. Recap: Smoothing with a Gaussian. Recap: Effect of σ on derivatives. Computer Science Tripos Part II. Dr Christopher Town

Feature Matching and Robust Fitting

Lecture 10 Detectors and descriptors

Object Detection by Point Feature Matching using Matlab

A SIFT Descriptor with Global Context

Motion illusion, rotating snakes

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford

Problems with template matching

AK Computer Vision Feature Point Detectors and Descriptors

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford

Comparison of Feature Detection and Matching Approaches: SIFT and SURF

Edge and corner detection

HISTOGRAMS OF ORIENTATIO N GRADIENTS

Performance Evaluation of Scale-Interpolated Hessian-Laplace and Haar Descriptors for Feature Matching

Feature Descriptors. CS 510 Lecture #21 April 29 th, 2013

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

Feature Detectors and Descriptors: Corners, Lines, etc.

Feature descriptors and matching

Image matching. Announcements. Harder case. Even harder case. Project 1 Out today Help session at the end of class. by Diva Sian.

Prof. Feng Liu. Spring /26/2017

Image Features. Work on project 1. All is Vanity, by C. Allan Gilbert,

Faster Image Feature Extraction Hardware

Anno accademico 2006/2007. Davide Migliore

Edge and local feature detection - 2. Importance of edge detection in computer vision

EE795: Computer Vision and Intelligent Systems

Image features. Image Features

Segmentation and Grouping

Automatic Image Alignment (feature-based)

Local Image Features

3D from Photographs: Automatic Matching of Images. Dr Francesco Banterle

Using Geometric Blur for Point Correspondence

Shape recognition with edge-based features

CS664 Lecture #21: SIFT, object recognition, dynamic programming

A Comparison and Matching Point Extraction of SIFT and ISIFT

Computer vision: models, learning and inference. Chapter 13 Image preprocessing and feature extraction

Coarse-to-fine image registration

Chapter 3 Image Registration. Chapter 3 Image Registration

Local Image Features

Multi-modal Registration of Visual Data. Massimiliano Corsini Visual Computing Lab, ISTI - CNR - Italy

SIFT Descriptor Extraction on the GPU for Large-Scale Video Analysis. Hannes Fassold, Jakub Rosner

IMAGE MATCHING USING SCALE INVARIANT FEATURE TRANSFORM (SIFT) Naotoshi Seo and David A. Schug

Transcription:

Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe presented by, Sudheendra Invariance Intensity Scale Rotation Affine View point Introduction Introduction SIFT (Scale Invariant Feature Transformation) approach Local features that are invariant to translation, rotation, scale, and other imaging parameters Features are highly distinctive, each feature finds its correct match in the database with high probability Robust against occlusion and clutter Related Research Local interest points Rover visual obstacle avoidance, Moravec 1981 corner detectors A combined corner and edge detector, Harris and Stephens 1988 Harris corner detector Rotationally invariant points Local greyvalue invariants for image retrieval, Schmid and Mohr 1997 Scale invariance Scale space theory, Lindeberg 1993 identifying appropriate scale for features Invariance to affine transformation 1. Feature Extraction SIFT method Keypoint Detection Search over all scales and image locations for stable interest points Keypoint Localization Fit a quadratic func and find extrema for more accuracy. Orientation Assignment Assign an orientaion to the keypoint using local gradient information Local image Descriptor Form a historgram of local gradients around the keypoint. Invariance to scale Invariance to rotation Partial invariance to viewpoint. Object recognition SIFT method Match key descriptors of test image to the database of features Euclidean distance ratio of nearest to second nearest neighbor Best-Bin Bin-First approximate algorithm Hough transform Cluster matched features with a consistent interpretation Vote for all object poses consistent with the feature Select clusters with more than 3 features Affine transformation Solve for affine parameters and perform geometric verification of o the object s pose in the test image 1

Keypoint detection Goal detect keypoints that are invariant to scale If the scale of the data set is not available then the only way to work with it is to represent it in all possible scales scale space theory Scale-space representation one parameter family of images where fine-scale image is information is successively suppressed (smoothened) Using some constrainsts on scale space functions Lindeberg defines the scale space represention of an image as Scale space representation (Example) And that the maxima and minima of the laplacian of L produce stable image features Difference of gaussian approximates laplacian of gaussian which is required for true scale invariance Extrema of the difference of gaussian gives interest points Different levels in the scale-space space representation of a two-dimensional image at scale levels t = 0,, 8, 3, 18 and 51 together with grey-level blobs indicating local minima at each scale Keypoint detection Keypoint Detection (Diagram) Incrementally convolve with gaussians seperated by a constant factor k Octave doubling of sigma Choose k such there are s images in a particular octave (k= 1/s ) Compute DoG from adjacent scales for entire octave Choose every second row and column from the image convolved with gaussian with twice the width and repeat above steps for the next octave Extrema detection - select points in each octave that are greater than or less than all their neighbors (8 in image space, 9+9 with different scales) Issues Keypoint detection 1. Frequency of sampling in scale Choosing s (defined previously) Affects the accuracy of the extrema Keypoint Localization Local extrema of DoG gives approximate location of keypoints Accuracy pixel, scaling factor level Quadratic function is fit using the 1 st and nd derivatives at the obtained keypoints Extrema of quadratic gives more accurate and stable keypoint location in scale space ( y, s) Accuracy - sub-pixel and sub-scale scale level Highest repeatability for 3 scales per octave. Frequency of sampling in the spatial domain Choosing the initial width of the gaussian Similar graph for repeatability against initialwidth provides an optimal value of 1.6 T D 1 T D D( x) = D+ x + x x D x = 1 D

Keypoint Elimination Function value at extrema used to reject unstable keypoints T 1 D D( x) = D + x discard if D(x D(x) < 0.03 Eliminate edge responses Edges have strong responses even if the location along the edge is poorly determined Principle curvatures Large along the edge, small perpendicular to it Eigenvalues of the Hessian proportional to principle curvatures Ratio of principle curvatures found from the sum and product of eigenvalues All points with a ratio greater than 10 are rejected Orientation assignment Goal to provide invariance to rotation Assign orientation based on local gradients Match features with respect to the orientation Orientation assignment To make the feature invariant to rotation If L is the Gaussian smoothed image in some scale Points in region around keypoint are selected and magnitude and orientations of gradient are calculated m ( y) = ( L( x + 1, y) L( x 1, y)) + ( L( y + 1) L( y 1 θ ( y) = tan (( L( y + 1) L( y 1)) /( L( x + 1, y) L( x 1, y))) Orientation histogram formed with 36 bins of points within a region around the keypoint Sample is added to appropriate bin and weighted by gradient magnitude and Gaussian-weighted circular window with a of σ 1.5 times scale of keypoint Highest peak in the histogram is selected along with any peak that is 80% of the highest peak to form multiple keypoints with different orientations Quadratic fitting performed to get peaks with higher accuracy 1)) Keypoint Descriptor Goal distinctive feature with invariance to viewpoint and intensity Motivation complex cells in the visual cortex Simple cells respond to a bar of light at the center Complex cells respond to a bar of light anywhere in the receptive field 16x16 region around keypoint divided into 16 4x4 regions Histogram of 8 orienations with gaussian weighted magitudes are formed (8 x 4 x 4 = 18 dimensional vector) Gradients rotated relative to keypoint orientation Similar to receptive field idea, allows upto four shifts of a point while still being in the same histogram Entries weighted using the distance from central point to avoid boundary effects Invariance to intensity vector is normalized Keypoint Descriptor (Diagram) Keypoint selection (Diagram) Rotated with respect to keypoint orientation 3

Matching Keypoints Feature location, scale, orientation of keypoint and 18 dimensional keydescriptor Independently match all keypoints of test image over all octaves with all keypoints of training image over all octaves Matching function - Euclidean distance Ratio of distance of closest neighbor to second-closest closest Eliminates matches with background noise and clutter Exhaustive search not feasible Best bin first approximate algorithm Similar to KD tree algorithm Matching Keypoints (Diagram) Threshold at 0.8 Eliminates 90% of false matches Eliminates less than 5% of correct matches KD tree and Best bin first KD tree Data structure for searches involving multi-dimensional keys Split data into two sets based on the median value for a particular dimension Choose another dimension and repeat on both sets Cycle thru the dimensions till all points have been covered Search performed by starting at the root and finding the closest leaf and then backtracking along the parent and pruning impossible branches Best-bin bin-first Instead of backtracking according to tree structure choose branches/bins that are close to the query point Search 00 nearest candidates and stop Provides a speedup of about orders of magnitude KD tree and Best Bin First Hough transform Keypoint parameters location, scale and orientation available in test image Using this and the keypoint parameters relative to the training image object position in the test image can be found Bins created for object parameters in test image 30 degrees for orientation A factor of for scale.5 times the size of the training image (according to scale) for location If more than 3 features fall into a bin the bin is subject to geometric verification for affine transformations Affine transformation Accounts for the 3d rotation of a planar surface The transformation of a model point [x y] to an image point [u v] is given by where m i are the affine rotation, stretch and scale parameters and t x, t y are the translational parameters The equation is solved for the least squares solution of these parameters Using this solution the error between each projected image feature and model feature can be calculated The features and parameters are accepted if e < 0.05 4

Results Results Results Conclusion Advantages Invariance to Scale Rotation Intensity Viewpoint (partially) Disadvantages Large feature size (18 dimensional floating opint vector) Large number of features (000 for typical images) 5