Using Geometric Blur for Point Correspondence

Similar documents
Scale Invariant Feature Transform

Scale Invariant Feature Transform

The SIFT (Scale Invariant Feature

Introduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Prof. Feng Liu. Spring /26/2017

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Motion Estimation and Optical Flow Tracking

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

Image Matching. AKA: Image registration, the correspondence problem, Tracking,

Outline 7/2/201011/6/

Invariant Features from Interest Point Groups

Local Features: Detection, Description & Matching

Local Feature Detectors

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Scale Invariant Feature Transform by David Lowe

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Object Recognition with Invariant Features

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

Wide Baseline Matching using Triplet Vector Descriptor

Corner Detection. GV12/3072 Image Processing.

Local features: detection and description. Local invariant features

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

Visual Tracking (1) Pixel-intensity-based methods

SIFT - scale-invariant feature transform Konrad Schindler

Advanced Video Content Analysis and Video Compression (5LSH0), Module 4

A SIFT Descriptor with Global Context

Lecture 10 Detectors and descriptors

Local invariant features

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

Feature Based Registration - Image Alignment

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability

Problems with template matching

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

HISTOGRAMS OF ORIENTATIO N GRADIENTS

Local Image Features

Visual Tracking (1) Feature Point Tracking and Block Matching

Computer Vision for HCI. Topics of This Lecture

Omni-directional Multi-baseline Stereo without Similarity Measures

Local features: detection and description May 12 th, 2015

Patch-based Object Recognition. Basic Idea

Announcements. Recognition (Part 3) Model-Based Vision. A Rough Recognition Spectrum. Pose consistency. Recognition by Hypothesize and Test

Implementing the Scale Invariant Feature Transform(SIFT) Method

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Implementation and Comparison of Feature Detection Methods in Image Mosaicing

Department of Electrical and Computer Systems Engineering

Previously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011

Recognition (Part 4) Introduction to Computer Vision CSE 152 Lecture 17

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

A Rapid Automatic Image Registration Method Based on Improved SIFT

Local features and image matching. Prof. Xin Yang HUST

Chapter 3 Image Registration. Chapter 3 Image Registration

IMPROVING SPATIO-TEMPORAL FEATURE EXTRACTION TECHNIQUES AND THEIR APPLICATIONS IN ACTION CLASSIFICATION. Maral Mesmakhosroshahi, Joohee Kim

Image Features: Detection, Description, and Matching and their Applications

CS 378: Autonomous Intelligent Robotics. Instructor: Jivko Sinapov

CS229: Action Recognition in Tennis

Performance Evaluation of Scale-Interpolated Hessian-Laplace and Haar Descriptors for Feature Matching

Matching Local Invariant Features with Contextual Information: An Experimental Evaluation.

Local Image Features

2D Image Processing Feature Descriptors

AK Computer Vision Feature Point Detectors and Descriptors

DETECTORS AND DESCRIPTORS FOR PHOTOGRAMMETRIC APPLICATIONS

Fast Image Matching Using Multi-level Texture Descriptor

AN EFFICIENT BINARY CORNER DETECTOR. P. Saeedi, P. Lawrence and D. Lowe

Automatic Image Alignment (feature-based)

Application questions. Theoretical questions

Edge and corner detection

Comparing Local Feature Descriptors in plsa-based Image Models

Object Recognition Algorithms for Computer Vision System: A Survey

Subpixel Corner Detection Using Spatial Moment 1)

Shape Matching and Object Recognition using Low Distortion Correspondences

Evaluation and comparison of interest points/regions

Computer Vision 558 Corner Detection Overview and Comparison

Shape Matching. Brandon Smith and Shengnan Wang Computer Vision CS766 Fall 2007

Filters (cont.) CS 554 Computer Vision Pinar Duygulu Bilkent University

Local Image Features

Research on an Adaptive Terrain Reconstruction of Sequence Images in Deep Space Exploration

EE795: Computer Vision and Intelligent Systems

A Novel Algorithm for Color Image matching using Wavelet-SIFT

Scale Invariant Segment Detection and Tracking

Computer Vision Lecture 20

Evaluation of the Influence of Feature Detectors and Photometric Descriptors in Object Recognition

Image Processing. Image Features

Computer Vision. Recap: Smoothing with a Gaussian. Recap: Effect of σ on derivatives. Computer Science Tripos Part II. Dr Christopher Town

Part-based and local feature models for generic object recognition

Computer Vision Lecture 20

Aggregating Descriptors with Local Gaussian Metrics

Digital Image Processing (CS/ECE 545) Lecture 5: Edge Detection (Part 2) & Corner Detection

Local Features Tutorial: Nov. 8, 04

Detecting Multiple Symmetries with Extended SIFT

A Comparison of SIFT and SURF

Automatic Image Alignment

Verslag Project beeldverwerking A study of the 2D SIFT algorithm

Pictures at an Exhibition: EE368 Project

Reliable Feature Matching Across Widely Separated Views

Deformation Invariant Image Matching

Final Exam Study Guide

Sketchable Histograms of Oriented Gradients for Object Detection

Transcription:

1 Using Geometric Blur for Point Correspondence Nisarg Vyas Electrical and Computer Engineering Department, Carnegie Mellon University, Pittsburgh, PA Abstract In computer vision applications, point correspondence plays an important role. Here, we present a technique called geometric blur to find point correspondences between two different images, even in the presence of affine distortions. We compare the results of this technique with other prevalent techniques for finding point correspondences, such as SSD (Sum of Squared Differences) with uniform Gaussian blur. Experimental results are shown on various face images and other object images. Index Terms point correspondences, geometric blur, interest points, object detection and recognition, SIFT P I. INTRODUCTION oint correspondence between two or more images is crucial component for many computer vision and image analysis tasks. Most methods for 3D reconstruction, object detection and recognition, image alignment and matching and camera calibration techniques assume that feature points were extracted and put to reliable correspondence. Hence, this problem is a key step in most computer vision applications, and so far there does not exist any completely automated method to solve this problem reliably. There have been considerable efforts to define correspondence between images based on local interest points, firstly used for stereo image matching by Moravec [1] before two decades. The corner detector of [1] was substantially improved by Harris et. al [2], which has been widely used for many computer vision and image matching tasks. Zhang et. al [3] matched Harris corners between two images by using correlation window of the neighborhood surrounding the Harris Manuscript received Dec 19th, 2005. * - The author is affiliated with the Electrical and Computer Engineering Department, Carnegie Mellon University, Pittsburgh, PA 15213, contact e-mail address for the author is : nisarg@ cmu.edu corners. Schmid and Mohr [4] used rotationally invariant descriptor of the local image region based on Harris interest points to match set of images. Traditionally, Harris corner detector is very sensitive to changes in image scale. Lowe [5][6] extended local descriptorbased approach to achieve scale invariance, which is known as Scale Invariant Feature Transform (SIFT). The SIFT operator finds characteristic scale of the interest point by finding an extrema of the interest point in a Difference of Gaussian (DoG) pyramid. The filter used for constructing DoG pyramid is a Gaussian blur filter of size 3x3. Berg et. al [7] argue that uniform blur filter might not be correct way to apply blurring, when one is interested in finding point correspondences. Moreover, they provide a new blurring operator called Geometric Blur. This paper compares results of experiments based on simple Sum of Squared Differences (SSD) based image patch matching technique, used on uniform Gaussian blurred neighborhood and geometric blurred neighborhood over Harris interest points. The paper is organized as follows. Section 2 discusses Geometric Blur and the underlying intuition behind using it for finding point correspondences and Section 3 discusses results of using different blurring techniques for solving the problem. In last section, we conclude with some future directions of this work. II. GEOMETRIC BLUR Geometric Blur is a notion of blurring, developed especially to compute measure of similarity between image patches (templates). It targets making the correspondence procedure more by making the templates discriminative and the matching robust. The standard strategy employed in the vision community to find similarity measure between two image patches is the coarse-to-fine way approach by

2 building Gaussian pyramids. At each level of the pyramid, the image is subsampled and convolved with the Gaussian filter in the spatial domain. Gaussian blur introduces a positional blur uniformly at each level. Using a uniform blur at each level of the pyramid creates a positional uncertainty about the central region of the image patch, which may be incorrect thing to do when one is interested in finding correspondence. The situation even worsens when there are affine distortions present in the images. of object catergories for our point correspondence experiments. Results We conducted Harris corner detection from set of pairs from the dataset we used. Figure 3 displays the outputs of applying Harris corner detector for example pairs of images. We assume that under the presence of an affine distortion, that fixes a single point, the distance the piece of signal changes is linearly proportional to the distance that the piece of signal is away from the feature point. This assumption can be encoded in the filter by making use of a spatially varying blurring kernel. Hence, instead of blurring by Gaussian filter of constant standard deviation (σ), we propose using a variable standard deviation, in linear proportion with the distance. (e.g. α x ). The following figure shows the difference between using uniform Gaussian blur filter and the geometric blur with spatially varying Gaussian Kernel. (a) (b) Figure 1: Comparison between geometric blur and uniform Gaussian blur. Geometric blur blurs the signal more farther from the origin (Figure Courtesy [7]) ( c ) ( d ) Geometric blur is more effective when applied to sparse signals[7], hence we have compute geometric blur on 4 distinct gradient channels: i) positive gradient in x direction, ii) negative gradient in x direction, iii) positive gradient in y direction and iv) negative gradient in y direction. III. EXPERIMENTS Database We have used face subset of the Caltech 101 dataset [8] (e) Figure 3 (a-f) Best 50 Harris interest points for, the interest points are marked with + (f)

3 We found out 50 most prominent interest points from each pair of images by using Harris corner detector. Given one interest point in the first image, we attempt to find a matching interest point in the second image The images are split into four sparse gradient channels as described in the previous section, and the geometric blur descriptors are extracted from the selected interest points. The matching of the descriptors is carried by Sum of the Squared Differences (SSD) operator over the descriptors. The geometric blur descriptor is taken by subsampled points of concentric circles around feature points. Geometric blur descriptor of each feature point consists of total 10 concentric circles, with each circle having 8 points subsampled. Calculation of geometric blur is carried out over four gradient channels. Thus, we get the final geometric descriptor having total of 320 dimensions per each feature point selected by Harris corner detector. Figure 4 displays the results of the point correspondences for the. The images are displayed on larger scales so that the same color coded correspondence points can be visualized in both the. (b) (a) ( c )

4 (f) ( d ) Figure 4 point correspondences derived from geometric blur extractor technique, the correspondence is established between images (a)-(b), (c)-(d) and (e)-(f), the correct correspondence is marked with same color and same shape at the similar locations in the pair of images. First pair has 17 correct matches, second pair has 18 correct correspondences and the third pair has 16 correct correspondences. For comparison, we established point correspondences using SSD for uniform Gaussian blur for the same images, and sample results are shown in figure 5. ( e )

5 ( a ) ( c ) ( b ) (d)

6 on uniform Gaussian blur, the correspondence is established between images (a)-(b), (c)-(d) and (e)- (f), the correct correspondence is marked with same color and same shape at the similar locations in the pair of images. First pair has 11 correct matches, second pair has 15 correct correspondences and the third pair has 12 correct correspondences. We tested on 30 pairs of facial images, once with same person, and once with different persons. The performance for both geometric blur and uniform blur techniques can be summarized from the following table. As it can be seen from the table, geometric blur descriptor performs much better than the uniform Gaussian blur. ( e ) Description Geometric blur, Same person Uniform Gaussian blur, Same person Geometric blur, different persons Uniform Gaussian blur, different persons Average number of successful correspondences out of first 25 correspondences Accuracy Percentage 17 68 % 12.2 48. 8 % 10.4 41.6% 7.8 31.2 % ( f ) Figure 5 point correspondences derived from SSD

7 IV. CONCLUSION In this paper, we propose to use geometric blur for finding point correspondences, and we claim that it is more effective than the normal uniform descriptors used for the task on hands. We apply both types of descriptors for finding correspondences on the facial images. Upon comparison of both descriptors, our results convey that geometric blur descriptors perform much better than the uniform Gaussian blur descriptors for the task of point correspondence. These early experiments provide motivation of using geometric blur application which require template matching, such as stereo vision and object detection. It would be interesting to combine geometric blur technique with successful point correspondence strategies such as SIFT, which use uniform Gaussian blurring method till this date. ACKNOWLEDGEMENT The author would like to thank Prof. Alexei A. Efros for providing valuable insights and for ftuiful discussions regarding project experiments. REFERENCES [1] H. Moravec, Rover visual.object avoidance, In International Joint Conference on Artificial Intelligence, Vancouver, Canada, pages 785-790, 1981 [2] C. Harris and M. Stephens, A combined corner and edge detector, In fourth Alvey Vision Conference, Manchester, UK, pages 147-151, 1988 [3] Z. Zhang, R. Deriche, O. Faugeras, and Q. T. Luong, A robust technique for matching two uncalibrated images through the recovery of the unknown epipolar geometry, Artificial Intelligence, issue 78, pages 87-119, 1995 [4] C. Schmid and R. Mohr, Local grayvalue invariants for image retrieval, in IEEE transactions on Pattern Analysis and Machine Intelligence, issue 9, volume 5, pages 530-534, 1997 [5] D. G. Lowe, Object recognition from local scale-invariant features, In International Conference on Computer Vision, Corfu, Greece, pages 1150-1157 [6] D. G. Lowe, Distinctive Image Features from Scale- Invariant Keypoints, International Journal of Computer Vision, 2004 [7] A. C. Berg and J. Malik, Geometric Blur for Template Matching, In International Conference on Computer Vision and Pattern Recognition, 2001 [8] L. Fei-Fei, R. Fergus and P. Perona, Learning generative visual models from few training examples: an incremental Bayesian approach tested on 101 object categories. In International conference on Computer Vision and Pattern Recognition, 2004