Lucas-Kanade Scale Invariant Feature Transform for Uncontrolled Viewpoint Face Recognition

Similar documents
A Comparison of SIFT, PCA-SIFT and SURF

III. VERVIEW OF THE METHODS

Specular 3D Object Tracking by View Generative Learning

Object Recognition with Invariant Features

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Scale Invariant Feature Transform

Scale Invariant Feature Transform

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

Patch-based Object Recognition. Basic Idea

Learning Patch Correspondences for Improved Viewpoint Invariant Face Recognition

Stable Keypoint Recognition using Viewpoint Generative Learning

Motion Estimation and Optical Flow Tracking

A Comparison and Matching Point Extraction of SIFT and ISIFT

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

A Novel Real-Time Feature Matching Scheme

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Viewpoint Invariant Features from Single Images Using 3D Geometry

Feature Based Registration - Image Alignment

Evaluation and comparison of interest points/regions

The SIFT (Scale Invariant Feature

VEHICLE MAKE AND MODEL RECOGNITION BY KEYPOINT MATCHING OF PSEUDO FRONTAL VIEW

Keypoint Recognition with Two-Stage Randomized Trees

Computer Vision for HCI. Topics of This Lecture

SURF applied in Panorama Image Stitching

A Rapid Automatic Image Registration Method Based on Improved SIFT

Introduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.

arxiv: v1 [cs.cv] 28 Sep 2018

Local features: detection and description. Local invariant features

Local Features Tutorial: Nov. 8, 04

Local features and image matching. Prof. Xin Yang HUST

Tensor Decomposition of Dense SIFT Descriptors in Object Recognition

Local Feature Detectors

IMPROVED FACE RECOGNITION USING ICP TECHNIQUES INCAMERA SURVEILLANCE SYSTEMS. Kirthiga, M.E-Communication system, PREC, Thanjavur

Performance Evaluation of Scale-Interpolated Hessian-Laplace and Haar Descriptors for Feature Matching

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

Feature Matching and Robust Fitting

Pose Normalization for Robust Face Recognition Based on Statistical Affine Transformation

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

arxiv: v1 [cs.cv] 28 Sep 2018

Implementation of Face Recognition Using STASM and Matching Algorithm for Differing Pose and Brightness

Implementing the Scale Invariant Feature Transform(SIFT) Method

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

A Novel Extreme Point Selection Algorithm in SIFT

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

Local features: detection and description May 12 th, 2015

Image Features: Detection, Description, and Matching and their Applications

School of Computing University of Utah

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Eppur si muove ( And yet it moves )

Local Image Features

Implementation and Comparison of Feature Detection Methods in Image Mosaicing

CAP 5415 Computer Vision Fall 2012

SIFT: Scale Invariant Feature Transform

Three-Dimensional Face Recognition: A Fishersurface Approach

Prof. Feng Liu. Spring /26/2017

Local Image Features

Scale Invariant Feature Transform by David Lowe

Visual Tracking (1) Pixel-intensity-based methods

A Novel Algorithm for Pose and illumination Invariant Image Matching

Stereoscopic Images Generation By Monocular Camera

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

An Evaluation of Volumetric Interest Points

Decorrelated Local Binary Pattern for Robust Face Recognition

CS229: Action Recognition in Tennis

Distinctive Image Features from Scale-Invariant Keypoints

Outline 7/2/201011/6/

Obtaining Feature Correspondences

Image Feature Evaluation for Contents-based Image Retrieval

2D Image Processing Feature Descriptors

Image Matching Using SIFT Algorithm And Pose - Illimination

A Novel Algorithm for Color Image matching using Wavelet-SIFT

Simultaneous Recognition and Homography Extraction of Local Patches with a Simple Linear Classifier

Local invariant features

Lecture 10 Detectors and descriptors

Corner Detection. GV12/3072 Image Processing.

Kanade Lucas Tomasi Tracking (KLT tracker)

Structure Guided Salient Region Detector

FAST AFFINE-INVARIANT IMAGE MATCHING BASED ON GLOBAL BHATTACHARYYA MEASURE WITH ADAPTIVE TREE. Jongin Son, Seungryong Kim, and Kwanghoon Sohn

TIED FACTOR ANALYSIS FOR FACE RECOGNITION ACROSS LARGE POSE DIFFERENCES

Lucas-Kanade Image Registration Using Camera Parameters

ASIFT: An Algorithm for Fully Affine Invariant Comparison

Video Google faces. Josef Sivic, Mark Everingham, Andrew Zisserman. Visual Geometry Group University of Oxford

Chapter 3 Image Registration. Chapter 3 Image Registration

Local Features: Detection, Description & Matching

An MKD-SRC Approach for Face Recognition from Partial Image

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model

CS 231A Computer Vision (Fall 2012) Problem Set 3

Tracking in image sequences

Extracting Spatio-temporal Local Features Considering Consecutiveness of Motions

3D model search and pose estimation from single images using VIP features

A Comparison of SIFT and SURF

Distinctive Image Features from Scale-Invariant Keypoints

Key properties of local features

Object Detection by Point Feature Matching using Matlab

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H.

Transcription:

Lucas-Kanade Scale Invariant Feature Transform for Uncontrolled Viewpoint Face Recognition Yongbin Gao 1, Hyo Jong Lee 1, 2 1 Division of Computer Science and Engineering, 2 Center for Advanced Image and Information Technology Chonbuk National University, Jeonju 561-756, Korea Abstract - Face recognition has been widely investigated in the last decade. However, real world application for face recognition is still a challenge. Most of these face recognition algorithms are under controlled settings, such as limited viewpoint and illumination changes. In this paper, we focus on face recognition which tolerates large viewpoint change. A novel framework named Lucas-Kanade Scale Invariant Feature Transform (LK-SIFT) is proposed. LK-SIFT is an extension of SIFT algorithm. SIFT is a scale and rotation invariant algorithm, which is powerful for small viewpoint changes in face recognition, but it fails when large viewpoint change exists. To handle this problem, we propose to use Lucas-Kanade algorithm to generate different viewpoint face from a single frontal face. After that, SIFT is used to detect local features from these viewpoints, these SIFT features contain information of different viewpoint face, which can deal with the problem of face viewpoint change. Finally, our framework is compared with the SIFT algorithm and other similar solutions. Experiment results show our framework achieves better recognition accuracy than SIFT algorithm at the cost of acceptable computational time gains compared with other similar algorithms. Keywords: Face recognition, Lucas-Kanade, Scale Invariant Feature Transform. 1 Introduction Real world face recognition has many useful applications, such as identifying subjects from surveillance camera for public security and annotating people from digital photos automatically for individuals. There are some successful commercial face recognition systems available like Google Picasa and Apple iphoto. However, face recognition research is still far from mature [1]. Earlier face recognition algorithms are only effective under controlled settings, such as the probe and gallery images are frontal. This algorithm fails when it is applied to cases as pose and illumination changes. This paper focuses on the viewpoint invariant face recognition, which identify face when probe faces are from different viewpoints while gallery faces are frontal. The problem of face recognition under different viewpoint is the distance between different poses is bigger than distance between different subjects. One solution is to eliminate the distance between different poses. Among which, face normalization is an effective method to remove the pose difference. Face normalization can be used as 2D or 3D model. As for 2D model, Markov Random Fields (MRF) is widely used to find correspondences between frontal face and the profile probe faces [2, 3]. MRF is to find 2D displacement by minimizing the energy, which consists of two parts, one is distance of corresponding node, another one represents the smoothness between neighbour nodes. Lucas-Kanade method is also used for face alignment [4, 5]. As for 3D model, Blanz et al. proposes an effective 3D morphable method to fit the 3D model to 2D face [6], the fitting shape and texture coefficients are used for face recognition. Normalization method can be used to construct the frontal face from the probe profile face [2]. It can also be used to directly match between a probe image and a gallery image and the matching score represent the similarity between two faces [3]. These normalization methods are effective at the cost of long computation time. It is reported that two minutes is needed to normalize one face [2]. Marsico et al. proposes a FACE framework to recognize face for uncontrolled pose and illumination changes [7]. It detects some keypoints using STASM algorithm [8], and construct half face by the middle line keypoints, the rest half face is reflected from the constructed half face. This easy method is fast but not robust for it highly depends on the accuracy of keypoints detection, when the keypoints detection fails, the system performance becomes bad. A new classifier or new feature is proposed to deal with the viewpoint change problem. For the new classifier, one shot similarity (OSS) or two shot similarity (TSS) are proposed by introducing another dataset, which contains no probe and gallery images [9]. Each dataset contains different images of a single subject or different subjects viewed from a single pose. Similarity scores between two faces are calculated by the model built by one of faces and the introduced dataset using LDA or SVM. Cross-pose face recognition shares similar idea by introducing a third dataset [10]. Faces from different viewpoints are all linearly represented by the introduced dataset using subspace method, similarity between these faces is then calculated indirectly by the linear coefficients. As for new feature extraction, tied factor analysis is proposed to estimate the linear transformation and noise parameters in identity space [11]. Besides the exploration on face recognition, there are many researches on local descriptor, which is effective to deal

with affine transformation between two images. Such as Harris-Affine [12], Hessian-Affine [13], Affine SIFT [14] algorithms. These algorithms are powerful for planar object comparison, while human face is non-planar, which contains significant 3D depth. Directly using these algorithms don t work well, we propose LK-SIFT framework to deal with large viewpoint change for face recognition. We use Lucas- Kanade to generate a series of different viewpoint faces from a single frontal face. After that, SIFT is used to detect local features among all these viewpoints. Through our method, SIFT features contains enough information from all viewpoints to handle face pose variance. The rest of this paper is organized as follows. Section II reviews the SIFT algorithm. We describe the proposed LK- SIFT framework in Section III. This algorithm includes image to image alignment and LK-SIFT algorithm. Section IV applies the above algorithm to FERET database, and presents the experiment results. Finally, we conclude this paper with future work in Section V. 2 Related work Local features are effective methods for matching and recognition for it is robust to occlusion, scale, rotation or even affine transformation to some extent. Among these algorithms, Scale Invariant Feature Transform (SIFT) is an scale, rotation invariant local feature. It transforms image data into scale-invariant coordinates and localize the keypoint. Each keypoint is assigned a descriptor. The major steps for SIFT algorithm are as following [15]: (1) Scale-space extrema detection: Image is transformed into different scales and size. Extrema are searched by finding maxima and minima over all scales using a difference-of- Gaussian scheme, which are invariant to scale and orientation. (2) Keypoint localization: Extrema are refined by excluding poor localized or low contrast points by checking the refined location, scale and ratio of principal curvatures. This increases stability of keypoint localization. (3) Orientation assignment: Each keypoint is assigned to one or more orientations based on local image gradient histogram. To provide scale and rotation invariance, local image data is transformed to the corresponding orientation and scale. (4) Keypoint descriptor: Local keypoint descriptor is calculated around each keypoint by histogram of gradients. The descriptor is transformed into a representation that allows for significant levels of local shape distortion and change in illumination. There are several methods reported for image matching and recognition of SIFT algorithm, such as BBF [16], Hough transform [17]. Nearest neighbour is the original and effective matching method for SIFT features. SIFT features are first pre-extracted from gallery images and stored in a database. When matching with a probe image, each SIFT feature from the probe image is compared with all gallery features in database. Nearest neighbour and second nearest neighbour are searched based on the Euclidean distance. The ratio of these two distances is compared with a threshold. Ratio that is smaller than the threshold is considered as a matching face. The SIFT is scale and rotation invariant feature, but it is not affine invariant. Affine SIFT is the extension of SIFT algorithm. There are several parameters for affine transformation as: A = H λ R 1 (ψ)t t R 2 (φ) cosψ sinψ = λ t 0 sinφ cosφ (1) sinψ cosψ 0 1 sinφ cosφ where λ, R i and T t are a scale parameter, rotated angle, and tilted angle, respectively. Fig. 1 shows the geometric interpretation of these parameters. SIFT algorithm is just scale ( λ ) and rotation ( ψ ) invariant. The left t and φ are not invariant, Therefore, SIFT algorithm is not fully affine invariant. Affine SIFT is trying to fulfil the t and φ invariant. Fig. 1 Geometric interpretation of affine decomposition. λ and ψ are scale and rotation from camera. θ and φ is tilt and rotation of subject, which named latitude and longitude respectively. Where t = 1/cosθ. Affine SIFT transforms an image into a series of simulated images by the change of longitude φ and latitude θ [14]. These simulated images are sampled to achieve a balance between accuracy and sparsity. However, Affine SIFT generates 61 images when the number of tilts set to 7. This increases the computation time too much, which is also unnecessary for face recognition. Moreover, human face contains 3D depth, while affine transformation is effective for planar object, simple affine transformation for a holistic face is not enough to represent the pose variant of face. In this paper, we propose to use LK-SIFT algorithm to simulate different pose from a single frontal face. 3 LK-SIFT 3.1 Image to Image Alignment Lucas-Kanade algorithm is first used as an effective image alignment method [18]. Image alignment is to find correspondences between gallery and probe images, firstly we equally divide image into several subregions, for pixels in the same subregions, we assume they share the same warp parameters, Let the warp function be x, = W(x, P), where P = [p 1, p 2, p m ] T, For affine warp, m=6, and

W(x, P) = 1 + p x 1 p 3 p 5 y (2) p 2 1 + p 4 p 6 1 Fig. 2 shows two images captured at two different poses, where I represents the probe image and T represents the gallery image. We divide image T into non-overlap subregions with same size. For each subregion r in T, we try to find a warp that aligns these two images. I r is the corresponding subregion to T r after warp transformation. The main objective for alignment is to minimize the error between the T r and the warped subregions I r as: E r = I r W(x, P) T r (x) 2 x (3) The solution for Equation 3 is to iterate calculating a P and update P till P converge. Lucas-kanade gives a solution for calculating P by: P = H 1 img x I r P T (T r (x) I r (W(x, p))) (4) where I r = ( I r, I r ) is the gradient of I x y r. is the Jacobian P of the warp (shown in Eq. 2). H img is the pseudo Hessian matrix, which is given by: H img = x I r P T I r (5) P We can now update the warp parameters P P + P and iterate till the parameters P converge. This procedure is applied independently for every patch/subregion. where I j,r and T j,r are the r-th subregion of the j-th image from two set of stack images. The solution of Eq. 6 for P is then as: 1 P = H (stk) j x I j,r P T (T r (x) I r (W(x, p))) (7) Let Ω = [P 1, P 2,, P N ] be the warp parameters for all the subregions between two viewpoints. Then we can learn a series of Ω for different viewpoints from a single frontal view as shown in Fig. 3. It is wise to generate viewpoint from nearby pose to reduce the alignment error. In our experiment, each pose is generated one by one from a slight pose change to larger one. Fig. 2 Image to image alignment, image is divided into several subregions T r, a warp between two subregions T r and I r is calculated by minimizing the alignment error. 3.2 LK-SIFT framework Image to image alignment can be used to online recognition between two images. There are two kinds of online recognition methods. The first one is to calculate a match score for two images based on the warp parameters or alignment errors. Another one is to normalize images by transforming the profile face to its frontal face. However, these online alignments require long computational time, which is not good for real time applications. Another scheme is off-line alignment. Warps parameters are trained from several stack images, each stack images are from the same pose. There are two strategies for off-line alignment. First one is to average two set of stack images, and learn the warp parameters between two average images. Another strategy is to find warp parameters that minimize all images from the two set of stack images as [5]: E r(stk) = I j,r W(x, P) T j,r (x) 2 j x (6) Fig. 3 Offline warps learning of different poses from frontal face. The warps are learned from nearby stacks. In general, our proposed LK-SIFT can be summarized as following: LK-SIFT: Lucas-Kanade Scale Invariant Feature Transformation Pre-computed: 1. Learn a set of warp parameters Ω 1, Ω 2, Ω N of N different poses from frontal face using N stack images, each stack images are from the same pose as shown in Fig. 3. 2. For each gallery (frontal) face, we generate N poses using the learned sets of warp parameters. 3. Compute the SIFT keypoints of these N pose faces and stored as a keypoint database. Recognition: 1. For each probe face, compute its SIFT keypoints and compare these keypoints with keypoints of each subject in the keypoint database. 2. The subject that has the maximum number of matching keypoints with the probe face is considered as recognized subject.

LK-SIFT computes SIFT keypoints from several points of view of the gallery face. It can handle the pose change of probe face to some extent. The number of pose generated from the gallery face should be chosen wisely to achieve a balance between accuracy and sparsity. The step between two nearby pose should not be either too big or small. A big step is not enough to guarantee the recognition accuracy, and a small step result in too many redundant keypoints, which increase the computational time. From our experiment, we conclude that pose change from -15 degree to 15 degree, SIFT can achieve high accuracy. Therefore, choose a step of 15 degree for nearby pose is reasonable to achieve a high accuracy. 4 Results In our experiments, we used FERET [19] grey database to evaluate our algorithm. This database contains 200 subjects, each subject contains 9 images captured from different poses. For each subject, we use frontal image as gallery, and other 8 pose images as probe images, the pose angle of which are -60, -40,-25, -15, 15, 25, 40 and 60 degrees, respectively. Figure 4 shows the different pose generated from frontal face using LK algorithm. First and third lines are the original database images of different pose; second and fourth lines are the generated pose faces using learned warps. (a) (b) (c) (d) Original Pose Simulated Pose Original Pose Simulated Pose Figure 4. Different poses generated from frontal face using LK algorithm. First and third lines are the original database images of different pose; second and fourth lines are the generated pose faces using learned warps. The parameters used in our experiment for SIFT algorithm are: image is resized to resolution of 200*200, and the ratio of nearest neighbour for SIFT is set to 0.8. Table I shows the comparison experiment results of recognition with ASIFT [14] and SIFT method. The number of tilt for ASIFT is set to 3. The number of tilt means the affine transformation times for θ or t. When it sets to 3, ASIFT generates 10 viewpoints. For LK-SIFT algorithm, it generates 9 poses. From the table, we know that SIFT can get good results when a pose degree is between -15 to 15 degree, but LK-SIFT achieves better results than SIFT, ASIFT, especially under large pose change. TABLE I EXPERIMENT RESULTS TO RECOGNIZE FACE WITH DIFFERENT POSE ON FERET DATABASE Pose ( ) SIFT (%) ASIFT(%) LK-SIFT(%) -40 53 56 63.5-25 96.5 92.5 95.5-15 99.5 99.5 98 15 100 99 98 25 92 93 93.5 40 48 58 59.5 Average 82 83 85 5 Conclusions In this paper, a novel framework named Lucas-Kanade Scale Invariant Feature Transform (LK-SIFT) is proposed. LK-SIFT is an extension of SIFT algorithm, which is scale and rotation invariant. SIFT algorithm is powerful for small viewpoint changes in face recognition, but it fails when large viewpoint change exists. To handle this problem, we propose to use Lucas-Kanade algorithm to generate different viewpoint faces from a single frontal face. After that, SIFT is used to detect local features from these viewpoints, these SIFT features contain information from different viewpoint faces, which can deal with the problem of face viewpoint change. Finally, our framework is compared with the SIFT algorithm and other similar solutions. Experiment results show SIFT can get good results when pose degree is between -15 to 15 degree, but LK-SIFT achieve better result than SIFT, ASIFT, especially under large pose different. The computation time for LK-SIFT is smaller than ASIFT. Acknowledgement: This work (Grants No. C0112553) was supported by Business for Cooperative R&D between Industry, Academy, and Research Institute funded Korea Small and Medium Business Administration in 2013. This work was also supported by the Brain Korea 21 PLUS project, National Research Foundation of Korea. 6 References [1] G. Hua, M. H. Yang, E. L. Miller, Y. Ma, M. Turk, D.J. Kriegman and T. S. Huang, Introduction to the Special Section on Real- World Face Recognition, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 33, no. 10, pp. 1921 1924, Oct. 2011.

[2] H. T. Ho, R.Chellappa, Pose-Invariant Face Recognition Using Markov Random Fields, IEEE Trans. Image Processing, vol.22, no.4, pp.1573-1584, Apr. 2013. [3] S. R. Arashloo and J. Kittler, Energy Normalization for Pose- Invariant Face Recognition Based on MRF Model Image Matching, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 33, no.6, pp. 1274-1280, June. 2011. [4] S. Baker and I. Matthews, Lucas-kanade 20 years on: A unifying framework, International Journal of Computer Vision, vol. 56, no.3, pp.221 255, Mar. 2004. [5] A. B. Ashraf, S. Lucey and T. Chen, Learning Patch Correspondences for Improved Viewpoints Invariant Face Recognition, in Proc. IEEE conf. Computer Vision and Pattern Recognition, 2008, pp. 1-8. [6] V. Blanz and T. Vetter, Face Recognition Based on Fitting a 3D Morphable Model, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 25, no. 9, pp. 1-12, sep. 2003. [7] M. D. Marsico, M. N. D. Riccio and H. Wechsler, Robust Face Recognition for Uncontrolled Pose and Illumination Changes, IEEE Trans. Systems, Man, and Cybernetics: Systems, vol. 43, no. 1, pp. 149-162, Jan. 2013. [14] J. M. Morel and G. Yu, ASIFT, A new framework for fully affine invariant image comparison, SIAM Journal on Imaging Sciences, vol. 2, no. 2, pp.438-469, 2009. [15] D. G. Lowe, Distinctive Image Features from Scale- Invariant Keypoints, Int l J. Computer Vision, vol. 60, no. 2, pp. 91-110, Nov. 2004. [16] J. Beis, and D. G. Lowe, Shape indexing using approximate nearest-neighbour search in highdimensional spaces, In Proc. Computer Vision and Pattern Recognition, Puerto Rico, 1997, pp. 1000-1006. [17] Hough, P.V.C. 1962. Method and means for recognizing complex patterns. U.S. Patent 3069654. [18] B. Lucas and T. Kanade, An iterative image registration technique with an application to stereo vision, In Proc. International Joint Conference on Artificial Intelligence, 1981, vol. 2, pp. 674 679. [19] P. J. Phillips, H. Wechsler, J. Huang, and P. Rauss, The FERET database and evaluation procedure for facerecognition algorithms, Image Vis. Comput., vol. 16, no. 5, pp. 295 306, Apr. 1998. [8] S. Milborrow and F. Nicolls, Locating facial features with an extended active shape model, in Proc. Eur. Conf. Comput. Vis., 2008, pp. 504 513. [9] L. Wolf, T. hassner, and Y. Taigman, Effective Unconstrained Face recognition by Combining Multiple Descriptors and Learned Background Statistics, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 33, no. 10, pp. 1978-1990, Oct. 2011. [10] A. Li, S. Shan, and W. Gao, Coupled Bias Variance Tradeoff for Cross-Pose Face Recognition, IEEE Trans. Image Processing, vol. 21, no. 1, pp. 305-315, Jan. 2012. [11] S. J. D. Prince, J. H. Elder, j. Warrell, and Fatima, Tied Factor Analysis for Face Recognition across Large Pose Differences, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 30, no. 6, pp. 970-982, June. 2008. [12] K. Mikolajczyk and C. Schmid, Scale and Affine Invariant Interest Point Detectors, Int l J. Computer Vision, vol. 1, no. 60, pp. 3-86, 2004. [13] K. Mikolajczyk, T. Tuytelaars, C. Schmid, A. Zisserman, J. Matas, F. Schaffalitzky, T. Kadir, L.V. Gool, A Comparison of Affine Region Detectors, Int l J. Computer Vision, vol. 65, no. 1-2, Nov. 2005.