The Log Polar Transformation for Rotation Invariant Image Registration of Aerial Images

Similar documents
The SIFT (Scale Invariant Feature

Local Features: Detection, Description & Matching

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Introduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

School of Computing University of Utah

Implementing the Scale Invariant Feature Transform(SIFT) Method

Comparison of Feature Detection and Matching Approaches: SIFT and SURF

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

Motion Estimation and Optical Flow Tracking

Scale Invariant Feature Transform

Scale Invariant Feature Transform

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Obtaining Feature Correspondences

Feature Based Registration - Image Alignment

Object Recognition with Invariant Features

CAP 5415 Computer Vision Fall 2012

Local Features Tutorial: Nov. 8, 04

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

Outline 7/2/201011/6/

A Novel Algorithm for Color Image matching using Wavelet-SIFT

Local Feature Detectors

Local invariant features

Computer Vision for HCI. Topics of This Lecture

Implementation and Comparison of Feature Detection Methods in Image Mosaicing

SIFT: Scale Invariant Feature Transform

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

CS 558: Computer Vision 4 th Set of Notes

SIFT - scale-invariant feature transform Konrad Schindler

3D Reconstruction From Multiple Views Based on Scale-Invariant Feature Transform. Wenqi Zhu

Faster Image Feature Extraction Hardware

Key properties of local features

Anno accademico 2006/2007. Davide Migliore

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

Feature Detection and Matching

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Motion illusion, rotating snakes

Feature Detectors and Descriptors: Corners, Lines, etc.

Advanced Video Content Analysis and Video Compression (5LSH0), Module 4

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

Automatic Image Alignment (feature-based)

Panoramic Image Stitching

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

A Rapid Automatic Image Registration Method Based on Improved SIFT

Local features: detection and description. Local invariant features

Feature Matching and Robust Fitting

Coarse-to-fine image registration

Image features. Image Features

Edge and local feature detection - 2. Importance of edge detection in computer vision

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

Computer Vision I. Announcements. Fourier Tansform. Efficient Implementation. Edge and Corner Detection. CSE252A Lecture 13.

Feature descriptors and matching

Chapter 3 Image Registration. Chapter 3 Image Registration

Subpixel Corner Detection Using Spatial Moment 1)

Local features: detection and description May 12 th, 2015

A Comparison of SIFT and SURF

Distinctive Image Features from Scale-Invariant Keypoints

Instance-level recognition

2D Image Processing Feature Descriptors

Detection of deformable objects in a nonstationary

Lecture 6: Edge Detection

Instance-level recognition

CS 223B Computer Vision Problem Set 3

A Comparison and Matching Point Extraction of SIFT and ISIFT

Local Image Features

Image Matching. AKA: Image registration, the correspondence problem, Tracking,

CS 556: Computer Vision. Lecture 3

IJMTES International Journal of Modern Trends in Engineering and Science ISSN:

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

Automatic Image Alignment

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

A Comparison of SIFT, PCA-SIFT and SURF

Car Detecting Method using high Resolution images

Distinctive Image Features from Scale-Invariant Keypoints

HISTOGRAMS OF ORIENTATIO N GRADIENTS

Invariant Features from Interest Point Groups

Pictures at an Exhibition: EE368 Project

Robotics Programming Laboratory

CS 231A Computer Vision (Fall 2012) Problem Set 3

Other Linear Filters CS 211A

A Summary of Projective Geometry

Local features and image matching. Prof. Xin Yang HUST

THE ANNALS OF DUNAREA DE JOS UNIVERSITY OF GALATI FASCICLE III, 2007 ISSN X ELECTROTECHNICS, ELECTRONICS, AUTOMATIC CONTROL, INFORMATICS

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Visual Tracking (1) Pixel-intensity-based methods

Verslag Project beeldverwerking A study of the 2D SIFT algorithm

Acquisition of high resolution geo objects Using image mosaicking techniques

Object Recognition Algorithms for Computer Vision System: A Survey

Distinctive Image Features from Scale-Invariant Keypoints

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability

Copy-Move Forgery Detection using DCT and SIFT

Feature Extraction and Image Processing, 2 nd Edition. Contents. Preface

Scale Invariant Feature Transform by David Lowe

Combining Appearance and Topology for Wide

NOWADAYS, the computer vision is one of the top

Pictures at an Exhibition

Local Image Features

CSE 527: Introduction to Computer Vision

Transcription:

The Log Polar Transformation for Rotation Invariant Image Registration of Aerial Images B. Janardhana Rao 1 and O.Venkata Krishna 2 1 CVR College of Engineering/ECE, Hyderabad, India Email: janardhan.bitra@gmail.com 2 CVR College of Engineering/EIE, Hyderabad, India Email: venkatakrishna.odugu@gmail.com Abstract Aerial image registration requires a high degree of precision. To improve the feature-based registration accuracy, this paper proposes a novel Log-Polar Transform (LPT) based image registration. Instead of using the whole image in the conventional method, feature points are used here, which reduces the computational time. It is not important how the image patch is rotated for rotation invariance. The key is focusing on the feature points. So a circular image patch is used, instead of using square image patches as used in conventional methods. In the existing techniques of registration always do FFT first and then Log-Polar Transformation (LPT). In this proposed technique does LPT first and then the FFT. The proposed technique contains four steps. First, feature points are selected in both the reference image and the sensed image with corner detector. Secondly, using feature point positions as centers image patches are created. Circular image patches are cropped choosing a feature point as center. Then the circular images are transformed from Cartesian co-ordinates to Log-Polar coordinates. Next, the LPT images are dealt using phase correlation. Experimental results demonstrate the reliability and rotation invariance of the proposed method. IndexTerms Image registration, Log-Polar Transformation (LPT), Fast Fourier Transform (FFT), Phase correlation. I. INTRODUCTION Image Registration is an essential step in most image processing tasks. It is widely used in remote sensing, weather forecasting, medical imaging and so on. Image Registration is the process of overlaying two or more images that have the same scene taken at different image sensors, different times, and different geometric viewpoints, or share the same visual information such as the same objects. The main goal of image registration is to find the transformation between the reference image and the sensed images. However, it could not design a universal registration system to solve all kinds of tasks depending on the different type of image and various types of degradations. A. Image Registration Process Image registration process involves the following four steps. First step is Feature Extraction which consists feature construction and feature selection of image. The feature construction and feature selection gives feature points (key points) in the image using one of the corner detector methods. Each feature point is unique. Second step is Feature Matching which establish correspondence between the reference and the sensed images. It gives the best matching point pairs between the images using similarity measures. Third step is Homography Estimation, based on the matching points obtained from feature matching estimate the parameter of the mapping function. In order to get the parameters, homography estimation uses only the true matching points. If the wrong matching points are used then the images will be incorrectly aligned. Fourth step is Transformation, after estimating the homography parameters, the sensed image is transformed using any one of transformation technique. The transformation consists translation, rotation and scaling. The existing registration methods are firstly, taking feature points over the images to correct the geometric errors such as translation, rotation and scaling, and secondly, establishing the correspondence model between images. Phase correlation, based on the translation property of the Fourier transform, computes the FFT ratio instead of any feature points. The sharp peak of the inverse of this ratio represents the highest correlation. Before phase correlation, the reference and sensed images are transformed from the Cartesian coordinates to the logpolar image and then phase correlation computes the similar ratio, rotation and scaling errors which are represented as shift. The matching is establish by using these parameters. The existing registration methods use the whole image to do log-polar transformation and phase correlation. However, in the common log-polar transformation for image registration is to crop a square patch in both the reference image and the sensed image. After the image has been transformed from Cartesian coordinates to Log- Polar coordinates, black semicircles appear on the border of the image, overlapping to form sharp edges / corners in between. Phase correlation focuses on these corners at the black semicircles. There is no information at these black semicircles, which reduces the accuracy of the image registration. Fig 1 shows a sample of previous registration methods. In order to improve the accuracy of feature-based registration for aerial image registration requires a high degree of precision. This paper proposes a novel Log- Polar Transform (LPT) based image registration. Here instead of using the whole image in the existing method, 833

feature points are used, which reduces the computational time. neighborhood shows edges in more than one direction. The Harris Corner Detector algorithm has the following steps: 1. Define a matrix M for each pixel (x,y) in the image domain, as the positive semi-definite, symmetric matrix Fig 1: The previous Image registration method Furthermore, the black semicircles also need to be avoided. The proposed process contains four steps. First, feature points are selected in both the reference image and the sensed image with corner detector (Harris or SIFT). Secondly, image patches are created using feature point positions as centers. Each point is a center point of LPT, so circular image patches are cropped choosing feature point as center. The radius of the circle can be changed, then the circular images are transformed to Log- Polar coordinates. Next, the LPT images are dealt with using phase correlation. Fig 2 shows the proposed solution, which removes the black semicircles. Fig 2: The Proposed Image Registration Technique II. FEATURE EXTRACTION The feature extraction is the process of finding the feature points since we are dealing with aerial images there is some problem with these, for example old aerial photos may appear yellow due to decay or general wear and tear, a better stitching result will not be achieved if the two aerial images have different illuminations, even after registration and it takes a long time to deal with the high resolution images. In order to avoid these they should be preprocessed before going to used for feature extraction. The preprocessing includes modifying the intensity and removal of noise. The feature points are extracted from the Harris corner detection method then go to Scale Invariant Feature Transform (SIFT). The Harris Corner Detector is one of the simplest corner detector algorithm, proposed by Chris Harris and Mike Stephens. It is based on the second moment matrix, called the auto- correlation matrix. The main idea is to find the interest points called corners where the surrounding 2. Let w be a weight matrix(a Gaussian filter), the convolution M = w*m The width of w determines the interesting region around x. The result M is auto-correlation matrix, sometimes called Harris matrix. It gives a averaging of M over the neighborhood pixels. 3. For each pixel (x,y), construct an indicator function, calculate the cornerness(x,y). cornerness(x,y)= det(m) k(trace(m)) 2 det(m) = λ 1 λ 2 (trace(m)) 2 = λ 1 + λ 2 k is a constant. A typical value for k is 0.04. 4. Set all cornerness(x,y) below a threshold T to zero. Threshold T controls the number of corners. The scale invariant feature transform (SIFT) algorithm was developed by David G. Lowe [5] [6] [7]. The SIFT features are invariant to image scale and rotation. Because of the highly distinctive property of the SIFT feature, a single feature can be correctly matched with high probability against a large number of features[7]. SIFT feature detection works in four consecutive steps. The first step is scale-space extrema detection, used to get the potential feature points which are invariant to scale and orientation. This step is implemented by using the extrema of Difference of Gaussians(DOG) to search over all scales and image locations. First, it establishes the scale-space with Gaussian function. The scale space can be written as L(x, y, σ) = G(x, y, σ) * I(x, y) Where * is the convolution operation, I(x,y) is an input image and is the Gaussian function in 2D situation. While σ needs to be assigned a different value to construct n scaled image in its first scale space. The DOG image is obtained by the Difference of Gaussian function convolved with the image I(x,y). It can be described as D(x, y, σ) = ((L(x, y, kσ) L(x, y, σ)) * I(x, y) which is equal to D(x, y, σ) = L(x, y, kσ) L(x, y, σ) where k = 2. The size of each octave scale space is half the previous one. Fig 3 shows a set of scale space images are produced by the initial image convolved with Gaussian repeatedly 834

on the left and adjacent Gaussian images are subtracted to produce the difference of Gaussian(DOG) on the right. Fig 3: Construction of DOG image[7] Secondly, it detects the local extrema(maxima or minima) of the DOG image. In Fig 4, each pixel is compared against 26 neighbour points (8+2 9=26, 8 neighbours in the current scale space image, 9 neighbours in the scale above and 9 neighbours in the scale below). The pixel is selected only if its value larger or smaller than all of these neighbors. The second step is keypoint localization, The feature points are obtained from local extrema (maxima and minima) of DOG. For stability, it should reject the points whose locations are along the edge. Same as the Harris Corner Detector, it computes Hessian matrix and checks that the ratio of principal curvatures is under some threshold, r. We need to check Fig 4: Maxima and minima of the difference-of-gaussian images The fourth step is keypoint descriptor, an image location, scale and orientation of each keypoint are defined from the previous steps. This step is to compute a descriptor which is highly distinctive and invariant to 3D viewpoint or changes in illumination. Fig 5 illustrates how to create a keypoint descriptor. Here a 16 16 Gaussian window is selected around the keypoint location and broken it into 4 4 subwindows. It calculates the magnitude m(x,y) and orientation θ(x,y) of the gradient at each point in each 4 4 subwindow, then these orientations are put into an 8 bin histogram which contains 8 features in each subwindow. In this case, the 16 16 window is used to create a keypoint, so that each keypoint has 128 feature vectors(16 8=128). The third step is orientation assignment, The keypoint descriptor can tolerate image rotation by being assigned an orientation. The following method can provide the most stable results. For each image sample, L(x,y) (Gaussian smoothed image), the gradient of each key point should be calculated. The magnitude m(x,y)and orientation θ(x,y) of the gradient is computed respectively as Fig 5: Creation of keypoint descriptor[11] III. FEATURE MATCHING An orientation histogram is formed from the gradient orientations of sample points within a region around the keypoint according to Lowe s suggestion [19], which has 36 bins covering the 360 degree range of orientations. The highest peak in the histogram is detected later any other local peak with 80% of the peak is used to also create a keypoint with that orientation. Because of locations with multiple peaks of similar magnitude, there will be multiple keypoints created at the same location and scale but with a different orientation. Based on the feature points extracted from the previous chapter, we will start feature matching. Original feature matching process establishes correspondence between the reference image and the sensed image. The proposed method is matching the correspondence between the circular image patches in both images. The circular image patches with feature point as center from square image patches are obtained by using Log- Polar Transform (LPT), a nonlinear sampling method which converts an image from the Cartesian coordinates I(x,y) to the log-polar coordinates LPT(ρ, θ). The mapping function is given by 835

Where, (xc, yc) is the center pixel of the transformation in the Cartesian coordinates, (x, y) is the sampling pixel in the Cartesian coordinates, (ρ, θ) is the log-radius and the angular position in the log-polar coordinates. The distance between two successive sampling points increases exponentially in the radius direction as the feature point gets far away from the center which is shown in fig 6. In the angular direction, the circumference is sampled with the same number of samples. So image pixels close to the center are oversampled while image pixels far away from the center are undersampled or missed. In the log-polar coordinates, the row represents the log-radius while the column represents the angular direction. The benefit of using log-polar coordinates instead of Cartesian coordinates is that any rotation and scale in the Cartesian coordinates are represented as translation in the log-polar coordinates. So LPT based image registration can deal with changes in rotation and scale. The common method transforms a square image to log-polar coordinates. The LPT transformed image appears with black semicircle shape regions due to undersampling or exempt information. A second benefit of log-polar coordinates is the space variant property. The sampling resolution in the Cartesian coordinates is higher around the center point and decreases exponentially as it moves away from the center. So the center of the LPT plays a major part than the surrounding areas which are far away from the feature point center. M x N in the spatial domain. It is the sampled Fourier Transform, so it does not contain all the frequencies which are forming an image, but the set of samples is large enough to fully describe the spatial domain image. The Discrete Fourier Transform (DFT) is computed as follows: The concept behind the Fourier transform is using a sum of sine and cosine waves of different frequencies to construct any waveform. The frequencies are determined by expanding the exponential in the above equation sine and cosine with variables u and v. The inverse discrete Fourier transform of the above equation is So, if F(u,v) is known, using above equation corresponding image f(x,y) will obtain. The registration of the translated images are done by Phase Correlation method, which is based on the Fourier Shift property[1]. The shift obtained transforms as phase difference in the Fourier domain. The Phase Correlation algorithm can be described as follows: (a) Given the two input images ga and gb, then apply a window function (e.g. a Hanning window) on both the images to reduce edge effects. Next, calculate the discrete Fourier transform of both images. Ga = F{ga}, Gb = F{gb} The experimental results show that the images without a window function get better results in this case, so we can ignore this step. (b) Normalize the cross-power spectrum, given by Fig 6: LPT mapping: (a) LPT sampling in the Cartesian Coordinates (b) the sample result in the angular and logradius directions [8] Using any one of the feature point as the center of the circle, our proposed method extracts a circle image patch from Cartesian coordinates to the log-polar coordinate, then it will avoid the black semicircle shape regions. The phase correlation can become more accurate if there is no noise regions. The image processing operators are more efficient when they are applied in the frequency domain. The Fourier Transform is a technique which is used to convert images from the spatial domain into the frequency domain and decompose an image into its cosine and sine components. In the Fourier domain, each point represents a particular frequency contained in the spatial domain image. The main idea of Discrete Fourier Transform is that F(u,v) in the frequency domain represents the image f(x,y) of size where * is the complex conjugate. (c) Apply the inverse Fourier transform to get the normalized crosscorrelation. r = F 1 {R} (d) Look for the location of the peak in its inverse Fourier transform r. ( x, y) = argmax (x,y) {r} The false matches are removed using the symmetry test in which the false match (i.e. the match pair could not pass the symmetry test in both directions ) are removed by computing the matches using the symmetrical (two-sided) matching conditions and keep the accurate match. IV. HOMOGRAPHY ESTIMATION The Homography is a method of mapping points and lines in one plane to another plane. The estimation of homographies between images using point feature correspondences by common algorithm. According to the definition of homography from Hartley and Zisserman [4], homography matrix H of size 3 3 is suffcient to calculate the homography. The homography matrix H has 8 independent degrees. 836

The algorithms for Homography Estimation includes the basic Direct Linear Transform (DLT) algorithm which is an algorithm for computing the homography matrix H if it gives four or more point correspondences. The relationship between corresponding points x and x If the 2D points are mapping in homogeneous coordinates, described as In homogeneous coordinates the points are scaled, so the 2D point can be expressed as x=[x,y,w]=[αx,αy,αw] = [x/w,y/w,1 ]. To get a unique identification of the image coordinates x,y, the points are often normalized with w=1. In order to obtain the correct result, Hartley and Zisserman [4] propose a normalization step on DLT. when there are more than 4 points of correspondence available minimization of a suitable cost function should be considered to solve for a homography. The homography estimation algorithm requires a number of correspondences at least 4. However, in some situations the two features in the images do not correspond to the same real feature. Therefore we need to distinguish inlier, the true matching correspondence, is used for homography estimation and outlier, the wrong matching correspondence, is not considered for homography estimation. Only inlier matches can be used for homography estimation. We will discuss two homography estimation methods, one is RANSAC (Random Sample Consensus) is the most popular robust homography estimation method [2]. The method can be described as first, it selects 4 correspondences at random and computes a homography matrix H. Secondly, it classifies the other correspondences as inliers or outliers depending on its concurrence with H. Next it repeats this process in all iterations and then chooses the iteration which contains the largest number of inliers. The homography matrix H in that iteration is considered as the best transformation model. The distance threshold t between x and Hx needs to be assigned when RANSAC classifies correspondences as inliers or outliers. If the distance > t, the point is considered an outlier. We also need to decide the number of iterations for running RANSAC. Hartley and Zisserman [4] suggest the formula that where ε is the probability of outliers and s is the number of correspondences used in each iteration (s = 4 in this case ). The distance threshold is chosen to find out outliers and discard them in RANSAC and it deals with the sum of squared difference algorithms. However, the Algebraic distance version of DLT is not very robust with outliers. The sum of squared residuals obtained in RANSAC replaced by the Least Median of Squares (LMS) estimation method. Peter J.Rousseeuw [9] states that LMS works very well if there are less than 50% outliers and it does not need to set any parameters such as the threshold or the probability of outliers. The major disadvantage of the LMS is that it could not deal with the outliers, which are more than half the data. V. SYSTEM DESCRIPTION The paper analyze the transformation relationship between the reference image and the sensed image. For obtaining a satisfying final result, preprocessing work is usually done to remove noise and modify intensity after selecting interesting aerial images. The corner detector method is used to obtain feature points in the reference image and the sensed image. Circle image patches are created using these feature points as centers. In order to deal with rotation situation, circle shaped image patches are used. It can avoid the black semicircle shaped noise after log-polar transformation. Next, we transform all the image patches from the Cartesian coordinates I(x,y) to the logpolar coordinates LPT(ρ, θ). Any rotation and scale in the Cartesian coordinates are represented as translation in the log-polar coordinates. Then Fast Fourier Transform is used to convert log-polar based images from the spatial domain into the frequency domain. Based on the Fourier Shift property, the Phase Correlation method is proposed for the registration of translated images. Symmetry test is used for removing false matches. A good matching result can give accurate transformation parameters. At last, RANSAC is used for homography estimation. Some parameters will influence the accuracy of this proposed registration system, the radius of circular image patches and the size of input images of FFT are the important parameters considered in this paper. As shown in Fig 7, the radius of circular image patches decides the size of input image of LPT. It also shows that the size of input images of FFT affects the FFT. The output image of the LPT is the same one for input image of FFT. The Area of the input image of LPT should be bigger than the area of input image of FFT to get high accuracy. Fig 7: The parameters that affect registration are radius and size 837

The pixel of the center point is removed for avoiding oversampling in LPT. Fig 8 shows in detail when the size of square subimage is m m, the radius of the largest circle image patch inside is m = 2 radius + 1. The input size of FFT 64x64 needs to be tested, so the size of the interesting square should be bigger than 64x64 then 71x71 is chosen for the input size of FFT 64x64 shown in Fig 9(b). The radius of the circle image patch in LPT are 10, 12, 14, 16, 18, 20 and the input sizes of FFT are 8x8, 16x16, 32x32, 64x64 are taken for the test. 10 sets of images have been tested. Fig 10 shows one sample of test results. Fig 8: The radius of circular image patch The best value for the radius obtained by testing the radius of the circle image patches to get more accuracy of registration. The size of input image of LPT also will be tested. Then two methods are compared, first Log-polar then FFT or first FFT then Log-polar. VI. EXPERIMENTAL RESULTS The algorithm is developed using MATLAB. All of the images that have been tested are screenshots from Google Earth. The resolution of these images are 418*320. Irrespective of registration method and the particular images used, the most important thing for the user is an estimate how accurate the registration is. The accuracy evaluation is not a normal problem, because the errors can be dragged into each of stages in the registration process. Inorder to evaluate the accuracy of these experimental results, the homography with RANSAC was estimated ten times and the average number of inliers was calculated. If the same results were obtained ten times in each actual experiment, it proved that the proposed method is stable and reliable. The proposed registration method is LPT PC (Log-Polar Transform and Phase correlation). At first, the experiment will test different parameters and select the optimal parameters for the proposed algorithm. Later shows the experimental results of feature matching applying first log-polar then FFT and first FFT then log-polar. Then it shows the comparison of number of inliers in the proposed algorithm LPT PC using SIFT selecting feature points with the original SIFT. A. Results of Optimal Parameter Selection The optimal parameters selected from the parameters considering the radius of the circle image patch in LPT, as well as the input size of FFT. Then these optimal parameters will be used to registration. Here 400 feature points are extracted using SIFT in the reference images and the sensed images. Because the interesting square is extracted from the original image, using feature points as centers, when the square is 41x41, the biggest radius should be 20 shown Fig 9(a). In order to get high precision, the input size of FFT should be smaller than the size of the interesting square. Furthermore, the input size of FFT should be 2n(n = 1, 2, 3, 4...). When the square is 41x41, the biggest input size of FFT is 32x32. (a) (b) Fig 9: The principle of selecting parameters. The chart presented in Fig 10, shows the optimal parameters, the radius of circle image patch for LPT is 20, the input size of FFT is 32x32. In this case, the highest number of the inliers is 238 for this set of images. Fig 10: Chart showing the number inliers for different parameters B. Results of first LPT then FFT and first FFT then LPT Here the two methods are compared, first LPT then FFT and first FFT then LPT. The proposed log-polar method is used to implement the first LPT then FFT method. In both the methods SIFT is chosen to extract the feature points then the optimal parameters are used to complete this experiment. The lines are used to connect sets of feature points which have top high phrase correlation. Looking at the matching connection in Fig 11(a) and (c), there are some false matching connections between the reference image and the sensed image in first FFT then LPT method. Fig 11 (b) and (d) shows all the match connections are correct in first LPT then FFT. C. Registration Results Results of the proposed registration method LPT PC with SIFT feature detector and original SIFT are compared here. The SIFT is selected to extract the feature points. 838

ISSN:2229-6093 image, the inliers in the original SIFT is 165 and in the LPT PC is 234, it means that the LPT PC works better. For scale image, the original SIFT performs better than the LPT PC. (a) (b) Fig 12: Chart showing the number of inliers for different rotation degree( one image is rotated in different angle degrees). (c) Fig 13: Chart showing the number of inliers for random angle rotation(four different images are rotated in random angle degree). (d) Fig 11: Two sequence method: (a)(c) first FFT then LPT, (b)(d) first LPT then FFT. Next, the first LPT then FFT method are chosen, which is compared with Lowe s SIFT. Both methods select 400 feature points from the original images. After homography estimation, the number of inliers are compared. At the beginning, in both registration methods the different angle rotated images are tested. In this experiment there are 450, 900, 1350, 1800, 2250, 2700 and 3150 rotation images. The results of inliers have been presented in the chart shown in Fig 12. As shown in Fig 12, the original SIFT get the number of inliers above 180, the highest number of inliers is 205. The number of inliers in the LPT PC are range in 143 to 181. Fig 13 shows 4 sets of images with random angle degrees. It also proves the proposed method can handle rotation images. At last, the number of inliers have been compared in the different type of images which are Brightened image, Rotated image (45 degree), Scaled image and Noisy image. The chart in Fig 14 shows that the LPT PC has high ability for dealing with the Noisy image which got 196 inliers, while the original SIFT gives low number of inliers 17. For rotated image, the inliers in the original SIFT is 198 and in the LPT PC is 181. For the Brighten Fig 14: Chart showing the number of inliers for different images. CONCLUSION This paper contains an overview of image registration methods and proposes a novel registration method which can be used for 2D aerial images. In order to improve the accuracy of registration, it selected the feature points with corner detector and created image patches using these feature points as centers, instead of using the whole image in the conventional method, which reduces the computational time. For rotation invariant, it is not important how the image patch is rotated, the key is focusing on the feature points. Each point was a center point of LPT, so the circular image patches were cropped 839

over the center point. Then it transformed the circular images to Log-Polar coordinates, which can avoid black semicircles from appearing on the border of the LPT image. Next, it coped with the LPT images with phase correlation. Due to the radius of the circle image patches can be changed. After testing, the best parameters have been found (the radius for LPT as 20, the input size of FFT as 32x32) for this novel method. Image registration with FFT always do the FFT first. But in order to avoid the black semicircles, applied LPT first to get the circular region and then the FFT. Taken the comparison first LPT then FFT and first FFT then LPT methods. The number of inliers for different rotation degree, the number of inliers for four different images are rotated in random angle degree and the number of inliers for different types of images with noise, rotation, high intensity and scaling are compared for proposed LPT PC and original SIFT. According to the comparison, the LPT PC works better. ACKONOWLEDGEMENTS I would like to express my gratitude to all those who gave me the possibility to complete this paper. I wish to express my deepest gratitude to Mr. O. Venkata Krishna, Sr. Assistant Professor for his genuine support, valuable advice and sincere comments which helped me a lot in finishing this paper. His excellent scientific expertise, in combination with his friendly and helpful character has made him an outstanding academic teacher. REFERENCES [1] Ronald N. Bracewell. The Fourier Transform and Its Applications, 2nd ed. McGraw-Hill, USA, 1986. [2] Elan Dubrofsky. Homography estimation. Master s thesis, University of British Columbia, 2009. [3] J.B. Zerubia Foroosh, H. and M. Berthod Extension of phase correlation to subpixel registration. IEEE Transactions of Image Processing, Volume 11, Mars 2002. [4] R. I. Hartley and A. Zisserman. Multiple View Geometry in Computer Vision. Cambridge University Press, ISBN: 0521540518, second edition, 2004. [5] David G. Lowe. Object recognition from local scaleinvariant features. In The Proceedings of the Seventh IEEE International Conference on Computer Vision, Corfu, Greece, 1999. [6] David G. Lowe. Local feature view clustering for 3d object recognition. In Proceedings of the 2001 IEEE Computer Society Conference on omputer Vision and Pattern Recognition, Kauai, Hawaii, 2001. [7] David G. Lowe. Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision, 60:91 110, 2004. [8] R. Matungka, Y.F. Zheng, and R.L. Ewing. Image registration using adaptive polar transform. Image Processing, IEEE Transactions on, 18(10):2340 2354, 2009. [9] P.J.Rousseeuw. Least median of square regression. Journal of the American Statistical Association, 79(388):871 880, 1984. [10] Utkarsh Sinha. Sift: Scale invariant feature transform. http://www.aishack.in/2010/05/sift-scaleinvariant-featuretransform, May 2010. Date visited: 2013-01-20. BIOGRAPHIES Mr. Bitra Janardhana Rao was born in 1984.He received his B.Tech in ECE from Nagarjuna University,Guntur in 2006. He has completed his Master of Technology in Communication and Radar Systems from KL University in 2010.He is currently working as an Assistant professor in CVR College Of Engineering, Hyderabad. His research interest includes Signal Processing and Image processing.. Mr. O. Venaka Krishna was born in 1983. He received his B.Tech in ECE from Nagarjuna University, Guntur in 2004.He has completed his Master of Technology in VLSI System Design from JNTU, Hyderabad in 2009.He is currently working as an Sr. Assistant professor in CVR College Of Engineering, Hyderabad. His research interest includes VLSI Signal Processing. 840