A Hierarchical Face Identification System Based on Facial Components

Similar documents
LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM

On Modeling Variations for Face Authentication

Image-Based Face Recognition using Global Features

An Integrated Face Recognition Algorithm Based on Wavelet Subspace

FACE RECOGNITION USING SUPPORT VECTOR MACHINES

Robust face recognition under the polar coordinate system

Image Processing and Image Representations for Face Recognition

The Analysis of Parameters t and k of LPP on Several Famous Face Databases

A New Multi Fractal Dimension Method for Face Recognition with Fewer Features under Expression Variations

A GENERIC FACE REPRESENTATION APPROACH FOR LOCAL APPEARANCE BASED FACE VERIFICATION

Fuzzy Bidirectional Weighted Sum for Face Recognition

Eigenfaces and Fisherfaces A comparison of face detection techniques. Abstract. Pradyumna Desale SCPD, NVIDIA

An Automatic Face Recognition System in the Near Infrared Spectrum

Local Similarity based Linear Discriminant Analysis for Face Recognition with Single Sample per Person

Restricted Nearest Feature Line with Ellipse for Face Recognition

Face Recognition Using SIFT- PCA Feature Extraction and SVM Classifier

Face recognition using Singular Value Decomposition and Hidden Markov Models

Face Biometrics Based on Principal Component Analysis and Linear Discriminant Analysis

Face Recognition Based On Granular Computing Approach and Hybrid Spatial Features

GENDER CLASSIFICATION USING SUPPORT VECTOR MACHINES

Learning to Recognize Faces in Realistic Conditions

Linear Discriminant Analysis in Ottoman Alphabet Character Recognition

Part-based Face Recognition Using Near Infrared Images

PARALLEL GABOR PCA WITH FUSION OF SVM SCORES FOR FACE VERIFICATION

Part-based Face Recognition Using Near Infrared Images

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 14, NO. 1, JANUARY Face Recognition Using LDA-Based Algorithms

Illumination Invariant Face Recognition Based on Neural Network Ensemble

Haresh D. Chande #, Zankhana H. Shah *

CHAPTER 5 GLOBAL AND LOCAL FEATURES FOR FACE RECOGNITION

IMPROVED FACE RECOGNITION USING ICP TECHNIQUES INCAMERA SURVEILLANCE SYSTEMS. Kirthiga, M.E-Communication system, PREC, Thanjavur

Semi-Supervised PCA-based Face Recognition Using Self-Training

NIST. Support Vector Machines. Applied to Face Recognition U56 QC 100 NO A OS S. P. Jonathon Phillips. Gaithersburg, MD 20899

FACE RECOGNITION USING PCA AND EIGEN FACE APPROACH

Multidirectional 2DPCA Based Face Recognition System

Recognition of Non-symmetric Faces Using Principal Component Analysis

Robust Face Recognition Using Enhanced Local Binary Pattern

Mobile Face Recognization

Local Binary LDA for Face Recognition

School of Computer and Communication, Lanzhou University of Technology, Gansu, Lanzhou,730050,P.R. China

Face Recognition using Principle Component Analysis, Eigenface and Neural Network

FACE RECOGNITION USING INDEPENDENT COMPONENT

FACE RECOGNITION BASED ON GENDER USING A MODIFIED METHOD OF 2D-LINEAR DISCRIMINANT ANALYSIS

An Efficient Face Recognition using Discriminative Robust Local Binary Pattern and Gabor Filter with Single Sample per Class

Semi-Random Subspace Method for Face Recognition

Facial Expression Recognition Using Non-negative Matrix Factorization

Aalborg Universitet. Haar-like Features for Robust Real-Time Face Recognition Nasrollahi, Kamal; Moeslund, Thomas B.

Disguised Face Identification Based Gabor Feature and SVM Classifier

An Improved Face Recognition Technique Based on Modular LPCA Approach

Linear Discriminant Analysis for 3D Face Recognition System

Principal Component Analysis and Neural Network Based Face Recognition

Applications Video Surveillance (On-line or off-line)

CHAPTER 3 PRINCIPAL COMPONENT ANALYSIS AND FISHER LINEAR DISCRIMINANT ANALYSIS

Research Article International Journals of Advanced Research in Computer Science and Software Engineering ISSN: X (Volume-7, Issue-7)

Directional Derivative and Feature Line Based Subspace Learning Algorithm for Classification

Image Based Feature Extraction Technique For Multiple Face Detection and Recognition in Color Images

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Face Recognition using Eigenfaces SMAI Course Project

Dr. Prakash B. Khanale 3 Dnyanopasak College, Parbhani, (M.S.), India

Hybrid Face Recognition and Classification System for Real Time Environment

Face Recognition using Tensor Analysis. Prahlad R. Enuganti

Decorrelated Local Binary Pattern for Robust Face Recognition

FACE RECOGNITION BASED ON LOCAL DERIVATIVE TETRA PATTERN

Selection of Location, Frequency and Orientation Parameters of 2D Gabor Wavelets for Face Recognition

Face Identification by a Cascade of Rejection Classifiers

Face Recognition Based on LDA and Improved Pairwise-Constrained Multiple Metric Learning Method

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

Optimal Feature Extraction Using Greedy Approach for Random Image Components and Subspace Approach in Face Recognition

Face Recognition via Sparse Representation

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

Small Sample Size Face Recognition using Random Quad-Tree based Ensemble Algorithm

Face Recognition by Combining Kernel Associative Memory and Gabor Transforms

Heat Kernel Based Local Binary Pattern for Face Representation

Combined Histogram-based Features of DCT Coefficients in Low-frequency Domains for Face Recognition

Robust Face Recognition via Sparse Representation

A NOVEL APPROACH TO ACCESS CONTROL BASED ON FACE RECOGNITION

A Study on Similarity Computations in Template Matching Technique for Identity Verification

II. PROPOSED FACE RECOGNITION SYSTEM

Generic Face Alignment Using an Improved Active Shape Model

Misalignment-Robust Face Recognition

Facial Expression Recognition with Emotion-Based Feature Fusion

Robust Facial Expression Classification Using Shape and Appearance Features

Illumination invariant face recognition and impostor rejection using different MINACE filter algorithms

Heeyoul (Henry) Choi. Dept. of Computer Science Texas A&M University

FACE RECOGNITION UNDER LOSSY COMPRESSION. Mustafa Ersel Kamaşak and Bülent Sankur

Face Detection and Recognition in an Image Sequence using Eigenedginess

Face Recognition Based on Multiple Facial Features

Diagonal Principal Component Analysis for Face Recognition

Dr. K. Nagabhushan Raju Professor, Dept. of Instrumentation Sri Krishnadevaraya University, Anantapuramu, Andhra Pradesh, India

Implementation of Face Recognition Using STASM and Matching Algorithm for Differing Pose and Brightness

Facial Expression Recognition Using Texture Features

COMBINING SPEEDED-UP ROBUST FEATURES WITH PRINCIPAL COMPONENT ANALYSIS IN FACE RECOGNITION SYSTEM

Object. Radiance. Viewpoint v

Compare Gabor fisher classifier and phase-based Gabor fisher classifier for face recognition

Face Recognition Using Adjacent Pixel Intensity Difference Quantization Histogram

A Novel Scheme for Face Recognition and Authentication Against Pose, Illumination and Expression Changes *

Two-View Face Recognition Using Bayesian Fusion

Enhancing Performance of Face Recognition System Using Independent Component Analysis

ROBUST LDP BASED FACE DESCRIPTOR

ROBUST FACE RECOGNITION USING TRIMMED LINEAR REGRESSION. Jian Lai and Xudong Jiang

Face Recognition Combine Generic and Specific Solutions

Transcription:

A Hierarchical Face Identification System Based on Facial Components Mehrtash T. Harandi, Majid Nili Ahmadabadi, and Babak N. Araabi Control and Intelligent Processing Center of Excellence Department of Electrical and Computer Engineering, University of Tehran, Tehran, Iran & School of cognitive science (SCS) Institute for studies in theoretical physics and mathematics (IPM), Tehran, Iran m.harandi@ece.ut.ac.ir, mnili@ut.ac.ir & araabi@ut.ac.ir Abstract It is generally agreed that faces are not recognized only by utilizing some holistic search among all learned faces, but also through a feature analysis that aimed to specify more important features of each specific face. This paper addresses a novel decision strategy that efficiently uses both holistic and facial component (left eye, right eye, nose and mouth) feature analysis to recognize faces. The proposed algorithm uses the whole face features in the first step of recognition task. If the decision machine fails to assign a class (with high confidence) then the individual facial components are processed and the resulting information are combined with those obtained from the whole face to assign the output. Simulation studies justify the superior performance of the proposed method as compared to that of Eigenface method. Experimental results also show that the proposed system is robust against small errors in facial component extractor. 1. Introduction Identifying people and facial recognition is an important and vital human ability which persists even through the passage of time, changes in appearance and partial occlusion. During the past decade different approaches for face identification are emerged, among them the most widely studied ones are based on statistical learning methods. The objective of statistical learning is to find, based on optimization of a certain separability criteria, a transformation which produce a feature representation with enhanced discriminatory power. While designing automatic face recognizers has been extensively improved in recent years, there is apparently a large gap between the current systems and the human ability of face recognition. One can even argue that while face recognition in human is a special task (with special cells), most of the current approaches in designing AFR can be considered as engineering solutions for object recognition and AFR is served as a benchmark for developing better methods of object recognition (shouldn't we apply many of the AFR methods for classifying other objects?). On the other hand, it is generally agreed that face recognition involves not only processing information about spatial layout or configuration of facial features (holistic view) but also information about individual facial features (mouth, nose, eyes, etc.). Utilizing the information from facial components is not only inspired by our biological findings but also has the following advantages Image variations (due to pose and/or illumination changes) within each component patch is more easily compensated than that of the whole image. A facial component can be weighted according to its importance. The component with a large variation can be weighted less in the matching stage. One of the first studies on the usage of facial components in the human face recognition was carried out by Brunelli and Poggio [1]. Brunelli suggest different mechanism to combine the classification results of each individual component for final decision including added score, weighted added score and the score of the most similar feature. Kim combines the facial features extracted by LDA into a single vector and uses the obtained vector for classification [2]. This paper introduced a hierarchical classifier that mimics the holistic/component behavior of human in recognizing faces. The proposed algorithm uses the whole face features in the first step of recognition task. If the decision machine fails to assign a class with high confidence then the individual facial components are processed and the resulting information are combined with those obtained from the whole face to form the final result. The novelty of our proposed methods comes from efficiently combining the individual facial components (left eye, right eye, nose and mouth) and the whole face

features and solving the recognition problem by a hierarchical decision structure. We want to emphasize that our novel approach can be considered as a practical way of combining the facial components in the decision path of most of the subspace based AFRs. So one can use PCA [3], LDA [4], ICA [5], KPCA [6] or any other statistical approach (and even a combination of them) to extract facial features and utilize our decision structure to design an AFR. The feasibility of the proposed methods has been successfully tested on ORL and YALE datasets where the images vary in expression and pose. The effectiveness of our methods is shown by comparing the recognition accuracy of the proposed methods with the well-known Eigenface method. This rest of this paper is organized as follows. In section 2 the proposed hierarchical classifier is introduced. Simulation results are described in section 3 and finally, conclusions are given in section 4. 2. Methodology Different mechanism can be utilized to use the facial component information in the recognition task. For example, one can extract the facial component features using PCA/ICA/LDA/KPCA and then combine these features into a single vector for recognition. As suggested by Brunelli, another approach is to classify each individual component and then added the score of each classifier to form the output. Intuitively, designing separate classifiers for facial components and fuse the results sound more logical than designing a classifier for the combined feature vector. Another advantage of this approach is robustness against pose and illumination because one can hope that some of the facial components are not degraded drastically by the changes in pose and illumination conditions. Biological studies also suggest the possibility of global description serving as a front end for finer, feature-based perception of faces [7]. The proposed block diagram of a recognition system that tries to mimic this behavior is shown in Fig. 1. After decomposing the input image into its components, the whole image and individual component features are extracted and fed into the decision machine. The decision machine first processes the features of the whole image in order to find a match within the stored database. If a match with high confidence is found then there is no need for further processing and the output is assigned. If the decision machine could not decide on the output class by just examining the whole image features, then the individual components are processed. Different strategies can be devised here, Whole image and facial component assign their outputs and the decision machine performs a weighted majority voting for final decision. Facial component modules only searches among the best matching subjects obtained from the whole image classifier and then the decision machine perform a majority voting or a weighted majority voting to assign the output. The strategy adopted in the reported experiment is the last one (Fig. 2). In our experiments we have utilized the weighted majority voting algorithm for assigning the output. The weights can be determined (from the training set) by using gradient based algorithms or even search algorithms with the leave one out mechanism. For extracting the whole image and facial component features, we utilized the well known Eigenfeature analysis. The whole image classifier selects k candidates for the recognition using a nearest neighbor (k-nn) classifier. If more than k of the neighbors belong to 2 the same class and this class contains the nearest neighbor too, then confidence check is passed and the classifier output is the label of that class, otherwise facial component classifiers (which are nearest neighbor classifiers too) compare the input pattern with the stored patterns of the k nearest classes obtained from the whole image classifier. In sequel the experimental results are reported. Fig. 1. Block diagram of the proposed recognition system. Whole image and facial component assign their outputs and the decision machine performs a majority voting for final decision.

In the sequel we assess the efficiency of our proposed methods against Eigenface method by comparing the recognition rate curves. 3.1. Tests on ORL dataset Fig. 2. Block diagram of the proposed hierarchical decision machine. 3. Experimental Results We assess the performance of the proposed methods using two datasets: A) ORL dataset that contains 400 images corresponding to 40 subjects and B) YALE dataset containing 165 images from 15 individuals. For both dataset we have used the k-nn classifier with k = 3 as whole image classifier. We also used the wellknown Eigenface method for feature extraction and compare the recognition curves of the proposed methods with Eigenface algorithm. We want to reemphasize that our novel approach can be considered as a practical way of combining the facial components in the decision path of most of the subspace based AFRs. Eigenfeature analysis [8], integral projection [1], multiresolutional analysis [9], and etc. are examples among different approaches devised for facial component extraction during the past years. As this paper is focused on the applications of facial components in boosting the recognition performance, the facial components are extracted manually. Later we will show how the performance of the proposed algorithm is affected by the accuracy of facial component extractor. Some examples of the extracted components for the first subject of YALE database is shown in Fig. 3. The ORL face database (developed at the Olivetti Research Laboratory, Cambridge, U.K.) is composed of 400 images with ten different images for each of the 40 distinct subjects. The variations of the images are across pose, size, time, and facial expression. All the images were taken against a dark homogeneous background with the subjects in an upright, frontal position, with tolerance for some tilting and rotation of up to about 20. There is some variation in scale of up to about 10%. The spatial and grey-level resolutions of the images are 92 112 and 256, respectively. Some of the images of this database are shown in Fig. 4. Fig. 4. Some examples of the ORL database. The first three images (from the ten available images for each subject) are chosen for training, while the remaining images are used for testing. In general, the performance of the Eigenface method varies with the number of principal components. Fig. 5 shows the recognition performance of the facial components and the whole image for the ORL database. Among the facial features, left eye component had the maximum discrimination power (near 65% correct classification). The facial components can be sorted by decreasing performance as follows 1. Left Eye 2. 3. 4. Right Eye It is interesting to note that the differences between recognition performances of the selected facial components are less than 7%, and there is a large gap between these components and whole image as the performance of whole image is always higher than the facial components by more than 20%. Fig. 3. Extracted facial components from the first subject of YALE database.

Correcr recognition rate 5 0.75 0.7 0.65 0.6 Whole Image Left Eye Right Eye Table 1. Hierarchical decision machine weights for ORL dataset Name Value Whole image 1 1 0 Left eye 3.2. Tests on YALE dataset Right eye 1 1 1 1 0.55 20 40 60 80 100 Fig. 5. Recognition performance for each facial component and the whole face on ORL database. Fig. 6 shows the results of applying our proposed method to the ORL database. The relative performance of the algorithms is self evident in this figure. Comparing the recognition curves reveals that our proposed method outperformed the Eigenface algorithm for all feature dimensions. The maximum recognition is obtained by using 90 features (the feature dimension is the identical for whole image and facial components) which is 89.29%. The maximum difference between performances is 3.6% and again obtained when 90 features are used. The Yale face database contains 165 images with 11 different images for each of the 15 distinct subjects. The 11 images per subject are taken under different facial expression or configuration: center-light, with glasses, happy, left-light, without glasses, normal, right-light, sad, sleepy, surprised, and wink. Some of the images of this database are shown in Fig. 7. The training set contains center-light, normal and without glasses of each subject and with glasses, happy, sad, sleepy, surprised and wink are used for testing. Correct recognition rate 8 6 4 2 Eigenface Method Proposed Method 0.78 20 40 60 80 100 Fig. 6. Comparison diagram of the proposed method vs. Eigenface method on ORL dataset. The Hierarchical decision machine uses the weights tabulate in table 1 for fusing the results of whole image and facial component classifiers. These weights are found by forcing the weights to be an integer in the interval [ 0,3] and brute force searching the weight space. Fig. 7. Some examples of the YALE database. Fig. 8 shows the recognition performance of the facial components and the whole image for the YALE database. Among the facial features, nose component had the maximum discrimination power (near 36% correct classification). The facial components can be sorted by decreasing performance as follows 1. 2. 3. Right Eye 4. Left Eye It is again interesting to note that the performance of whole image is always higher than the facial components by more than 45%. This is mainly because the differences between facial component images of each subject (due to lighting condition and the facial expression) and are higher than that of ORL database.

Table 2 shows the hierarchical decision machine weights which are again found by brute force searching the weight space. As it is evident from this table, the decision machine weights nose component more than the other components because it has higher recognition performance. Table 2. Hierarchical decision machine weights for YALE dataset Name Value Whole image 2 1 1 Left eye Right eye 1 1 2 1 In Fig. 9, the recognition rate of the proposed algorithm is compared to that of Eigenface method. Again for all feature dimensions the proposed method outperforms the Eigenface algorithm. The maximum recognition is obtained by using 19 features which is 93.33%. The maximum difference between performances is 4.4% and obtained when 15 features are used. Correct recognition rate 0.7 0.6 0.5 0.4 0.3 0.2 Whole Image Left Eye Right Eye 0.1 5 10 15 20 25 30 35 Fig. 8. Recognition performance for each facial component and the whole face on YALE database. 3.3. Robustness against facial component extractor In order to evaluate the dependence of the proposed algorithm on the accuracy of facial component, the following simple experiment is carried out on the YALE data-base. A uniform noise is added to the position of each facial component patch in the training set and test set. The power of the noise is dependent on the window size of facial component respectively. Fig. 10 shows the recognition rate of the proposed algorithm under noise disturbance. The performance of the hierarchical system outperforms the Eigenface method up to noises which can move the facial component window by ±20% window size randomly. If the noise power increases to ±30% of window size, the performance is converge to that of Eigenface method for higher number of features which shows that the facial components are no more useful how ever for small number of features still the proposed method outperforms the Eigenface algorithm. Correct recognition rate 4 2 8 6 4 2 Eigenface Method Proposed method with 20% disturbance in facial extractor Proposed method with 30% disturbance in facial extractor 5 10 15 20 25 30 35 Fig. 10. Comparison diagram of the proposed method with 20% and 30% noise disturbance vs. Eigenface method. Correcr recognition rate 2 8 6 4 2 Eigenface Method Proposed Method 5 10 15 20 25 30 35 Fig. 9. Comparison diagram of the proposed method vs. Eigenface method on YALE dataset. 4. Conclusion In this paper we proposed a hierarchical system which tries to mimic the human recognition behavior by efficiently combining the facial components in the recognition task. The decision system uses the weighted majority strategy to fuse the results of whole image and facial component classifier. Simulation studies justify the superior performance of the proposed method as compared to that of Eigenface method. Experimental results also show that the proposed system is robust against small errors in facial component extractor. Our next goal is to design an adaptive system which learns the weights for each subject in the training set.

5. References [1] R. Brunelli and T. Poggio, "Face recognition: features versus templates," IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 15, no. 10, pp. 1142-1052, Oct. 1993 [2] T. Kim, H. Kim, W. Hwang, S. Kee and J. Kittler, "Face description based on decomposition and combining of a facial space with LDA," Proceedings of the IEEE conference on Image Processing (ICIP'03), vol. 3, pp. 877-880, 14-17 Sep. 2003 [3] M. Turk and A. Pentland, "Eigenfaces for recognition," Journal of Cognitive Neuroscience, vol. 3, pp. 71-86, Mar. 1991 [4] P.N. Belhumeur, J.P. Hespanha and D.J. Kriegman, "Eigenfaces vs. Fisherfaces: recognition using class specific linear projection," IEEE Trans. Pattern Analysis and Machine Intelli-gence, vol. 19, no. 7, pp. 711-720, July 1997 [5] M.S. Bartlett, J.R. Movellan and T.J. Sejnowsk, "Face Recognition by Independent Component Analysis," IEEE Trans. on Neural Network, vol. 13, no. 6, pp. 1450-1464, Nov. 2002 [6] Y. Jian, A.F. Frangi, Y. Jing-Yu, D. Zhang and J. Zhong, "KPCA plus LDA: a complete kernel Fisher discriminant framework for feature extraction and recognition," IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 27,issue 2, pp. 230-244, Feb. 2005 [7] W. Zhao,, R. Chellappa, P. Phillips, A. Rosenfeld, "Face recognition: A literature survey", ACM Computing Surveys, vol. 35, issue 4, pp. 399 458, Dec. 2003 [8] A. Pentland, B. Moghaddam, and T. Starner, "View based and modular eigenspaces for face recognition," Proceedings on IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 84-91, 21-23 Jun 1994 [9] K. Kim, S. Oh and H. Choi, "Facial Feature Extraction Using PCA and Wavelet Multi-Resolution Images," Proceedings of the sixth IEEE conference on Automatic Face and Gesture Recognition (FGR'04), pp. 439-444, 17-19 May 2004