A 2D+3D FACE IDENTIFICATION SYSTEM FOR SURVEILLANCE APPLICATIONS

Similar documents
Exploitation of 3D Images for Face Authentication Under Pose and Illumination Variations

3D FACE RECOGNITION BY POINT SIGNATURES AND ISO-CONTOURS

3D Face Recognition Using Pose and Illumination Compensation

A NOVEL APPROACH TO ACCESS CONTROL BASED ON FACE RECOGNITION

Semi-Supervised PCA-based Face Recognition Using Self-Training

A Hierarchical Face Identification System Based on Facial Components

An Evaluation of Multimodal 2D+3D Face Biometrics

Multiple Nose Region Matching for 3D Face Recognition under Varying Facial Expression

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Haresh D. Chande #, Zankhana H. Shah *

Biometrics Technology: Multi-modal (Part 2)

3D Face Recognition. Anil K. Jain. Dept. of Computer Science & Engineering Michigan State University.

Generic Face Alignment Using an Improved Active Shape Model

The Novel Approach for 3D Face Recognition Using Simple Preprocessing Method

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model

3D Face Modelling Under Unconstrained Pose & Illumination

Spatial Frequency Domain Methods for Face and Iris Recognition

Gurmeet Kaur 1, Parikshit 2, Dr. Chander Kant 3 1 M.tech Scholar, Assistant Professor 2, 3

Digital Vision Face recognition

Detecting motion by means of 2D and 3D information

New Experiments on ICP-Based 3D Face Recognition and Authentication

MULTIVIEW REPRESENTATION OF 3D OBJECTS OF A SCENE USING VIDEO SEQUENCES

Extraction of 3D Scene Structure from a Video for the Generation of 3D Visual and Haptic Representations

K-Nearest Neighbor Classification Approach for Face and Fingerprint at Feature Level Fusion

Component-based Face Recognition with 3D Morphable Models

Segmentation and Tracking of Partial Planar Templates

Face Detection and Recognition in an Image Sequence using Eigenedginess

Recognition of Non-symmetric Faces Using Principal Component Analysis

Enhancing 3D Face Recognition By Mimics Segmentation

Illumination Invariant Face Recognition Based on Neural Network Ensemble

Three-Dimensional Face Recognition: A Fishersurface Approach

Automatic Gait Recognition. - Karthik Sridharan

Illumination invariant face recognition and impostor rejection using different MINACE filter algorithms

Computationally Efficient Serial Combination of Rotation-invariant and Rotation Compensating Iris Recognition Algorithms

FACE RECOGNITION USING INDEPENDENT COMPONENT

DTU Technical Report: ARTTS

LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM

Pose Normalization for Robust Face Recognition Based on Statistical Affine Transformation

NIST. Support Vector Machines. Applied to Face Recognition U56 QC 100 NO A OS S. P. Jonathon Phillips. Gaithersburg, MD 20899

IMPROVED FACE RECOGNITION USING ICP TECHNIQUES INCAMERA SURVEILLANCE SYSTEMS. Kirthiga, M.E-Communication system, PREC, Thanjavur

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg

An Efficient Face Recognition under Varying Image Conditions

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H.

Analysis of Local Appearance-based Face Recognition on FRGC 2.0 Database

Facial Expression Recognition Using Non-negative Matrix Factorization

Continuous User Authentication Using Temporal Information

Dynamic skin detection in color images for sign language recognition

Face Recognition using 3D Facial Shape and Color Map Information: Comparison and Combination

Tri-modal Human Body Segmentation

Robust face recognition under the polar coordinate system

EXPLOITING 3D FACES IN BIOMETRIC FORENSIC RECOGNITION

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation

Figure 1. Example sample for fabric mask. In the second column, the mask is worn on the face. The picture is taken from [5].

Performance Analysis of Face Recognition Algorithms on TMS320C64x

Gender Classification Technique Based on Facial Features using Neural Network

Online and Offline Fingerprint Template Update Using Minutiae: An Experimental Comparison

FACE RECOGNITION USING SUPPORT VECTOR MACHINES

Face Recognition using Eigenfaces SMAI Course Project

Keywords Wavelet decomposition, SIFT, Unibiometrics, Multibiometrics, Histogram Equalization.

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Lucas-Kanade Scale Invariant Feature Transform for Uncontrolled Viewpoint Face Recognition

BIOMET: A Multimodal Biometric Authentication System for Person Identification and Verification using Fingerprint and Face Recognition

Robust biometric image watermarking for fingerprint and face template protection

Image Processing and Image Representations for Face Recognition

Image-Based Face Recognition using Global Features

AN EXAMINING FACE RECOGNITION BY LOCAL DIRECTIONAL NUMBER PATTERN (Image Processing)

Biometrics Technology: Hand Geometry

Face Quality Assessment System in Video Sequences

A GENERIC FACE REPRESENTATION APPROACH FOR LOCAL APPEARANCE BASED FACE VERIFICATION

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma

I CP Fusion Techniques for 3D Face Recognition

Criminal Identification System Using Face Detection and Recognition

Distance-driven Fusion of Gait and Face for Human Identification in Video

Published in: IEEE 8th International Conference on Signal Image Technology & Internet Based Systems

Classification of Face Images for Gender, Age, Facial Expression, and Identity 1

Partial Face Recognition

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries

MULTI-VIEW FACE DETECTION AND POSE ESTIMATION EMPLOYING EDGE-BASED FEATURE VECTORS

Peg-Free Hand Geometry Verification System

Subject-Oriented Image Classification based on Face Detection and Recognition

Color Local Texture Features Based Face Recognition

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

Empirical Evaluation of Advanced Ear Biometrics

A new approach to reference point location in fingerprint recognition

Face detection and recognition. Many slides adapted from K. Grauman and D. Lowe

A Contactless Palmprint Recognition Algorithm for Mobile Phones

A Study on Similarity Computations in Template Matching Technique for Identity Verification

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

Implementation of a Face Recognition System for Interactive TV Control System

Linear Discriminant Analysis in Ottoman Alphabet Character Recognition

Dr. Enrique Cabello Pardos July

Adaptive Skin Color Classifier for Face Outline Models

Still Image Objective Segmentation Evaluation using Ground Truth

Component-based Face Recognition with 3D Morphable Models

Video Processing for Judicial Applications

Online Signature Verification Technique

Overview of the Multiple Biometrics Grand Challenge

Augmenting Reality with Projected Interactive Displays

Occlusion Robust Multi-Camera Face Tracking

Task analysis based on observing hands and objects by vision

Transcription:

A 2D+3D FACE IDENTIFICATION SYSTEM FOR SURVEILLANCE APPLICATIONS Filareti Tsalakanidou, Sotiris Malassiotis and Michael G. Strintzis Informatics and Telematics Institute Centre for Research and Technology Hellas st Km Thermi-Panorama Rd, Thessaloniki 5700, Greece filareti@iti.gr, malasiot@iti.gr, strintzi@eng.auth.gr Abstract A novel surveillance system integrating 2D and 3D facial data is presented in this paper, based on a low-cost sensor capable of real-time acquisition of 3D images and associated color images of a scene. Depth data is used for robust face detection, localization and 3D pose estimation, as well as for compensating pose and illumination variations of facial images prior to classification. The proposed system was tested under an open-set identification scenario for surveillance of humans passing through a relatively constrained area. Experimental results demonstrate the accuracy and robustness of the system under a variety of conditions usually encountered in surveillance applications.. Introduction During the last 0 years, state-of-the-art face recognition systems using 2D intensity images have advanced to be fairly accurate under controlled conditions. However, public face recognition benchmarks have shown that their performance degrades dramatically for images subject to pose, illumination or facial expressions variations []. To improve performance under these conditions, 3D face recognition has been proposed [2]. The use of 3D images for personal identification was mainly based on the fact that the 3D structure of the human face can be highly discriminatory and is inherently insensitive to illumination variations and face pigment. Moreover, the use of 3D data can aid significantly facial feature detection, facial pose estimation and pose compensation. The above characteristics are highly desirable especially for applications demanding dependable operation in uncontrolled environments, such as surveillance of public places. Nevertheless, implementation of 3D face recognition techniques for such applications is not widespread, mainly due to limitations imposed by 3D sensors. Unlike the widely-used CCTV cameras, 3D imaging technologies are not mature enough to provide high image quality at low cost. Moreover, 3D sensors usually have a limited field of view which constrains their working area to a radius of -2 meters for images of acceptable quality. In this paper a complete 3D face recognition system suitable for small-scale surveillance applications requiring unobtrusive recognition of uncooperative users within a limited area (e.g. monitoring of passengers validating their tickets in metro stations) is proposed. The system relies on a novel low-cost sensor capable of real-time acquisition of 3D images and associated color images of a scene [3]. Unlike feature-based techniques [4], that rely on high quality depth images usually acquired by 3D scanners, we propose an appearance based approach that demonstrates reliable performance even with low cost sensors. The main problem with such methods is the requirement for accurate alignment between probe and gallery images. A pose and illumination compensation scheme was proposed to achieve such geometric and photometric alignment. We also propose the combination of 2D and 3D images for reliable multimodal authentication. The combination of 2D and 3D images for PCA-based face recognition was also proposed in [5], where a database of significant size was used to produce comparative results of face recognition using eigenfaces for 2D, 3D and their combination. This test however considered only frontal images captured under constant illumination. In [6], a database containing several appearance and environmental variations was used. The Embedded Hidden Markov model method was implemented in both color and depth images and high recognition rates were reported. To cope with intrapersonal variations due to viewpoint or lighting conditions, the face database was enriched with artificially generated examples depicting variations in pose and illumination. In this paper, we describe and evaluate a complete face recognition system using a combination of 2D color and 3D

range images captured in real-time. A completely different approach compared to [6] is adopted. Variations of facial images due to changes in pose or illumination are compensated prior to classification. Several novel techniques which are capable, taking as input a pair of 2D and 3D images, to produce a pair of normalized images depicting frontal pose and illumination are proposed. Also, a novel face detection and localization method relying exclusively on depth data is presented. The efficiency and robustness of the proposed system is demonstrated on a data set of significant size depicting several variations, representative of the conditions encountered in real-world applications. The paper is organized as follows. In Section 2 the application scenario and 3D acquisition set-up are described. Pose and illumination compensation algorithms are presented in Sections 3 and 4 respectively. Face classification using 2D and 3D images is outlined in Section 5, while the performance of the system is extensively evaluated in Section 6. Finally, Section 7 concludes the paper. 2. Application scenario In this paper we propose a watch-list identification system for surveillance applications requiring unsupervised recognition of humans passing through a relatively constrained area. The proposed system employs a novel low-cost sensor capable of quasi-synchronous acquisition of 3D and color images, based on the Color Coded Light Approach [3]. The sensor is integrated using a standard CCTV color camera and a video projector both embedded in a mechanical construction. A color coded light pattern is projected on the scene and its deformation on the object surfaces is measured, thus generating a 3D image of the scene. By rapidly alternating between this pattern and a white light, a color image may be captured almost simultaneously (see Figure ). The resolution of the generated images is 580 780 pixels. The sensor was optimized for the application scenario described below, leading to an average depth accuracy of 0.5mm for objects located about m meter from the camera in an effective working space of 75cm 75cm 50cm. According to this scenario, the surveillance system comprises of a network of 3D sensors located above specific entrance or exit points. People are allowed to pass through these areas one at a time. The sensor is continuously recording pairs of depth and color images of the scene located inside its working volume. In parallel, the face recognition software processes each image pair to detect the presence of human faces. We note here, that the most important limitation of surveillance systems is caused by the fact that they have to operate in an unconstrained environment and without having the user s cooperation. This leads to environmental variations caused by changes in lighting conditions, as well as facial appearance variations due to facial pose. To cope with such variations, a pose and illumination compensation step is subsequently applied to each image pair, thus resulting in a pair of images depicting frontal pose and illumination. The normalized images are then compared against a watchlist to determine if the person appearing in these images is one of the enrolled persons (open-set :N identification). In case of a positive answer, an alarm signal is sent to the surveillance center. Although there are obviously some limitations on the working conditions under which the surveillance system operates, e.g. large facial poses (more than 35 ), the proposed system does not impose any strict constraints on user movement. 3. Face localization and pose compensation In the proposed system, face detection and localization is based solely on 3D data, thus being relatively insensitive in illumination changes and occlusion of facial features. Segmentation of the head from the body relies on modelling the head - torso points as a mixture of Gaussians, while the model parameters are estimated using the Expectation Maximization algorithm and a-priori knowledge of body and head geometry [7]. Figure : Pose compensation examples. The original pair of images and the resulting frontal views are shown. In the range image, warmer colors correspond to points closer to the camera, while black pixels correspond to undetermined depth values. To localize the position of the face and estimate the pose, first the tip of the nose is accurately detected. Then, a 3D line is fitted on the 3D points comprising the ridge of the nose, which defines two of the three degrees of freedom of the face orientation. The third degree of freedom, that 2

is the rotation angle around the nose axis, is estimated by exploiting the inherent face bilateral symmetry [7]. In order to compensate for any head rotation and generate a frontal view of this face, we first define a local coordinate system centered on the tip of the nose and aligned with the estimated face orientation. Then, we align this local frame with a reference coordinate frame using a warping transform. The resulting 3D image contains missing pixels, which can be filled either by using face symmetry or by interpolating their values using neighboring pixels (see Figure ). Gallery images are also aligned with the reference frame during the training phase of the algorithm. This last alignment is performed manually. The face detection-localization algorithm runs near to real-time (about 0fps). For each frame captured, an image quality measure is computed as follows: The estimated face pose parameters are used to create a normalized frontal depth image (dimensions 40 50) by means of 3D warping. This image is subsequently projected on a face-subspace created using a set of training depth images. Based on the distance from the face-subspace, a face-ness measure is computed and used as a quality measure [8]. Note, that if a non-face surface is in front of the camera, pose estimation will normally fail and thus the resulting quality measure will be very low. For subsequent frames of adequate quality, the best is selected and used for identification. In Figure 2 subsequent frames of an individual s face passing through the system s working volume can be seen. The green rectangular frame corresponds to the result of face detection. Figure 3: Illumination compensation examples. (a) Original image, (b) R(I D, L, u), (c) novel image relit by frontal light. an example-based regression technique is employed for estimating the non-linear relationship between the image brightness and light source direction L using a set of artificially generated bootstrap images. Given the frontal light direction L 0 and the estimate of the light source direction L, the illumination compensated image ĨC is given by multiplication of the input color image I C with a ratio image: Ĩ C (u) = I C (u) R(I D, L, u) R(I D, L 0, u) I C, I D are the pose compensated color and depth images and R is a rendering of the surface with constant albedo [3]. An important advantage of this algorithm is its flexibility in coping with complex illumination conditions by adaptation of the rendering function R. Figure 3 illustrates the relighting of side illuminated images. 5. Face classification Figure 2: Subsequent frames of an individual s face passing through the system s working volume. The green rectangular frame corresponds to the result of face detection. 4. Illumination compensation Inspired by recent work on scene relighting for realistic image renderings [9], the proposed illumination compensation method is based on the generation of a new face image relit from a frontal direction. First, we recover the scene illumination from a pair of pose compensated color and depth images. Assuming that the scene is illuminated by a single light source, A multimodal classification scheme integrating 2D and 3D face images is proposed. Two independent classifiers, one for color and one for depth images are used. Among several state-of-the-art face recognition algorithms examined, the Probabilistic Matching (PM) algorithm [8] gave the lowest error rates, while being computationally efficient. The PM algorithm is applied to both depth and color pose and illumination compensated images, independently. The scores computed by each classifier are subsequently normalized using the quadric-line-quadric normalization technique (QLQ), proposed in [0]. QLQ is an adaptive normalization procedure aiming to decrease the overlap of the genuine and impostor distributions, while still mapping the scores in [0, ]. Fusion of color and depth data is achieved 3

Detection / Recognition Face detection, localization and 3D pose estimation 3D pose compensation Illumination compensation Generation of 3D face template Watch list recognition Y/N Manual 3D pose compensation Illumination Compensation Generation of 3D face template Enrolment Figure 4: Block diagram illustrating the various steps of the proposed face identification system. Figure 5: Examples of test images depicting several appearance variations. by simply adding the two normalized scores (simple-sum fusion rule). The block diagram of Figure 4 illustrates the various steps of the face recognition algorithm. 6. Experimental results In this Section, we evaluate the efficiency of the proposed identification system using a face database of significant size, compiled in conditions that are similar to those encountered in real-world applications. The acquired images contain several facial appearance, as well as environmental variations: frontal neutral views, pose variations, facial expressions, illumination variations, images with/without wearing eyeglasses (see Figure 5). In total, more than 3000 image pairs were recorded for 73 enrolled users. The PM algorithm was trained using 2 frontal images per person. Testing was performed using the remaining images. The open-set :N identification scenario [], known as watch-list identification, was adopted for the evaluation of the proposed algorithms. According to this scenario, not every probe (unknown test face image) has a match in the gallery database (set of enrolled users), i.e. there are images in the testing set belonging to individuals not known to the system (impostors). In this case, the recognition task involves a) negative identification (rejection) of people not belonging to the gallery and b) correct identification of people that make up the watch list [2]. In surveillance applications, the watch-list of wanted individuals is much smaller than the number of people expected to be screened by the system. People not belonging to the watch-list are regarded as impostors, whose negative identification is sought after, while people in the watch-list are regarded as enroled users, whom the system should recognize correctly. For our experiments, the size of the watchlist varies from to 25 subjects for a set of 73 individuals. System performance is measured over the impostor and watch-list populations using the correct rejection rate and the correct identification rate respectively. The correct rejection rate is defined as the fraction of impostor probe images for whom the computed similarity measures, derived from their comparison to enroled users, are below threshold t. The correct identification rate is the fraction of enrolled user test images, for whom the similarity measure corresponding to their own gallery image (same id) is greater than t. In the following, we report the mean (average) correct identification and correct rejection rates obtained over 200 randomized runs of the algorithm. In each run, the watchlist subjects are randomly selected from the 73 individuals participating in our database. Table shows the mean correct rejection and recognition rates obtained for different watch-list sizes, as well as the corresponding system accuracy (detection and identification rate). For a probe set of N images (test images of both impostors and enrolled users), the accuracy is (average number of correct rejections+ average number of correct recognitions)/n. All reported rates are obtained using 4

Table : Average correct rejection and correct recognition rates and corresponding detection and identification accuracy. The threshold was set to a fixed value. Watch Correct Correct Equal List Recognition Rejection Accuracy Correct Size Rate (%) Rate (%) (%) Rate (%) 5 96.92 94.45 94.62 96.58 0 97. 90.03 9.00 96.5 5 97.07 86.22 88.45 95.82 20 96.90 83.2 86.88 95.30 25 97.0 80. 85.9 95.4 Correct Rejection Rate 5 0.75 0.7 0.65 Mean system performance for different watch list sizes 5 Correct Recognition Rate Watch list size = 5 Watch list size = 0 Watch list size = 5 Watch list size = 20 Watch list size = 25 Figure 6: Correct recognition rate vs. correct rejection rate for various watch-list sizes. a fixed threshold value. The last column of Table tabulates the equal correct rate values of the proposed system. The equal correct rate is defined as the correct rejection rate on the operating point (threshold value) for which the correct rejection and correct recognition rates are equal. Table 2 tabulates the correct recognition rates obtained for specific values of the correct rejection rate and viceversa. The corresponding detection and identification accuracy is also reported. Figure 6 illustrates the system s performance on different operating points. The average correct recognition rate is plotted against the average correct rejection rate for varying threshold values and different watch-list sizes. It can be observed that as the watch list size increases the system s performance gradually decreases, as expected. In Figure 7, the performance of the multimodal color+depth identification system, is compared to the performance of the unimodal systems for a watch-list of 5 subjects. The average correct recognition rate is plotted Table 2: Average correct recognition rates obtained for specific values of the average correct rejection rate and viceversa. Watch Correct Correct List Recognition Rejection Accuracy Size Rate (%) Rate (%) (%) 98.00 82.98 84.0 95.00 98.94 98.67 5 92.00 99.85 99.3 96.06 98.00 97.83 96.90 95.00 95.02 98.00 75.34 78.45 95.00 98.0 97.70 0 92.00 99.73 98.66 95. 98.00 97.62 96.53 95.00 95.32 98.00 65.0 72.45 95.00 97.42 96.85 5 92.00 99.58 98.03 94.6 98.00 97.3 96.00 95.00 97.4 98.00 6.03 70 95.00 96.0 95.48 20 92.00 99.33 97.36 93.92 98.00 96.87 95.56 95.00 95.83 98.00 56.70 70.60 95.00 95.48 95.50 25 92.00 99.28 96.72 93.60 98.00 98.76 95.4 95.00 95.0 against the average correct rejection rate for varying threshold values. It is clearly seen that the combined use of depth and color facial data leads to increased identification accuracy and high recognition rates. 7. Conclusions In this paper, a complete 2D+3D face recognition system based on automatic image normalization algorithms that exploit the availability of 3D information was presented. The proposed system is particularly suitable for small-scale surveillance applications requiring unobtrusive recognition of uncooperative users within a limited area, since it successfully addresses two major limitations of existing applications: pose and illumination variability. Given a pair of color and depth images, the pose of the face is first estimated using depth data only, and is then compensated, resulting in a frontal view. Finally, the scene illumination is recovered from the pair of images, and a new color image, 5

Correct Rejection Rate 5 0.75 0.7 0.65 0.6 0.55 0.5 Mean system performance using different modalities. 5 Correct Recognition Rate color depth color+depth Dimensional Image Processing, Vol., No. 5-6, pp. 358 369, Oct.-Dec. 2005. [4] T. K. Kim, S. C. Kee, and S. R. Kim, Real-time normalization and feature extraction of 3D face data using curvature characteristics, in Proc. 0th IEEE Int. Workshop on Robot and Human Interactive Communications, Sept. 200, pp. 74 79. [5] K. I. Chang, K. W. Bowyer, and P. J. Flynn, Face recognition using 2D and 3D facial data, in Proc. ACM Workshop on Multimodal User Authentication, Santa Barbara, California, Dec. 2003, pp. 25 32. [6] F. Tsalakanidou, S. Malassiotis, and M. G. Strintzis, Face localization and authentication using color and depth images, IEEE Trans. on Image Processing, Vol. 4, No. 2, pp. 52 68, Feb. 2005. Figure 7: Correct recognition rate vs. correct rejection rate for a watch-list of 5 subjects. The performance of the multimodal color+depth identification system is compared to the performance of the unimodal systems. relit from a frontal direction, is generated. Significant improvements in face classification accuracy were obtained using this scheme, as shown by experimental evaluation based on a database of significant size containing several facial appearance, as well as environmental variations. Acknowledgments This work was supported by research project PASION - Psychologically Augmented Social Interaction over Networks (contract No FP6-027654) under Information Society Technologies (IST) priority of the 6th Framework Programme of the European Community. References [] P. J. Phillips, P.J. Grother, R. J. Micheals, D. M. Blackburn, E. Tabassi, and J. M. Bone, Face recognition vendor test 2002, Evaluation report, Defense Advanced Research Projects Agency and National Institute of Justice, March 2003. [7] S. Malassiotis and M. G. Strintzis, Robust face recognition using 2D and 3D data: Pose and illumination compensation, Pattern Recognition, Vol. 38, No. 2, pp. 2537 2548, Dec. 2005. [8] B. Moghaddam, W. Wahid, and A. Pentland, Beyond eigenfaces: probabilistic matching for face recognition, in Proc. Int. Conf. on Automatic Face and Gesture Recognition (FGR 98), Nara, Japan, April 998, Vol., pp. 30 35. [9] L. Zhang and D. Samaras, Face recognition under variable lighting using harmonic image exemplars, in Proc. IEEE Int. Conf. on Computer Vision and Pattern Recognition (CVPR 03), June 2003, Vol., pp. 9 25. [0] R. Snelick, Umut Uludag, Alan Mink, Micahel Indovina, and Anil Jain, Large-scale evaluation of multimodal biometric authentication using state-of-the-art systems, IEEE Trans. on Pattern Analysis and Machine Intelligence, Vol. 27, No. 3, pp. 450 455, March 2005. [] P. Grother, Face Recognition Vendor Test 2002, Supplemental Report NISTIR 7083, National Institute of Standards and Technology, US, Feb. 2004. [2] F. Li and H. Wechsler, Open set face recognition using transduction, IEEE Trans. on Pattern Analysis and Machine Intelligence, Vol. 27, No., pp. 686 697, Nov. 2005. [2] Kevin W. Bowyer, Kyong Chang, and Patrick J. Flynn, A survey of approaches and challenges in 3D and multi-modal 3D+2D face recognition, Computer Vision and Image Understanding, Vol. 0, No., pp. 5, Jan. 2006. [3] F. Tsalakanidou, F. Forster, S. Malassiotis, and M. G. Strintzis, Real-time acquisition of depth and color images using structured light and its application to 3D face recognition, Elsevier Real-Time Imaging, Special Issue on Multi- 6