The Analysis of Parameters t and k of LPP on Several Famous Face Databases

Similar documents
Technical Report. Title: Manifold learning and Random Projections for multi-view object recognition

School of Computer and Communication, Lanzhou University of Technology, Gansu, Lanzhou,730050,P.R. China

Appearance Manifold of Facial Expression

Locality Preserving Projections (LPP) Abstract

Using Graph Model for Face Analysis

Globally and Locally Consistent Unsupervised Projection

Locality Preserving Projections (LPP) Abstract

An Integrated Face Recognition Algorithm Based on Wavelet Subspace

Fuzzy Bidirectional Weighted Sum for Face Recognition

Robust Pose Estimation using the SwissRanger SR-3000 Camera

A Hierarchical Face Identification System Based on Facial Components

Laplacian Faces: A Face Recognition Tool

Face Recognition using Laplacianfaces

Learning a Spatially Smooth Subspace for Face Recognition

A New Multi Fractal Dimension Method for Face Recognition with Fewer Features under Expression Variations

A Supervised Non-linear Dimensionality Reduction Approach for Manifold Learning

Laplacian MinMax Discriminant Projection and its Applications

Manifold Learning for Video-to-Video Face Recognition

Directional Derivative and Feature Line Based Subspace Learning Algorithm for Classification

Image-Based Face Recognition using Global Features

Local Discriminant Embedding and Its Variants

Large-Scale Face Manifold Learning

Linear Discriminant Analysis in Ottoman Alphabet Character Recognition

Local Similarity based Linear Discriminant Analysis for Face Recognition with Single Sample per Person

Dimension Reduction CS534

Misalignment-Robust Face Recognition

Linear Laplacian Discrimination for Feature Extraction

Discriminative Locality Alignment

A New Orthogonalization of Locality Preserving Projection and Applications

Head Frontal-View Identification Using Extended LLE

Image Similarities for Learning Video Manifolds. Selen Atasoy MICCAI 2011 Tutorial

Selecting Models from Videos for Appearance-Based Face Recognition

Restricted Nearest Feature Line with Ellipse for Face Recognition

Robust Face Recognition Using Enhanced Local Binary Pattern

Sparsity Preserving Canonical Correlation Analysis

FACE RECOGNITION USING SUPPORT VECTOR MACHINES

Diagonal Principal Component Analysis for Face Recognition

Linear local tangent space alignment and application to face recognition

Embedded Manifold-Based Kernel Fisher Discriminant Analysis for Face Recognition. Guoqiang Wang, Nianfeng Shi, Yunxing Shu & Dianting Liu

Face Recognition Based on LDA and Improved Pairwise-Constrained Multiple Metric Learning Method

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma

CHAPTER 3 PRINCIPAL COMPONENT ANALYSIS AND FISHER LINEAR DISCRIMINANT ANALYSIS

Semi-supervised Learning by Sparse Representation

Non-linear dimension reduction

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 14, NO. 1, JANUARY Face Recognition Using LDA-Based Algorithms

A Discriminative Non-Linear Manifold Learning Technique for Face Recognition

Stepwise Nearest Neighbor Discriminant Analysis

Dynamic transition embedding for image feature extraction and recognition

Heat Kernel Based Local Binary Pattern for Face Representation

Preserving Global and Local Features for Robust Face Recognition under Various Noisy Environments

Pattern Recognition 45 (2012) Contents lists available at SciVerse ScienceDirect. Pattern Recognition

Multidirectional 2DPCA Based Face Recognition System

Robust Face Recognition via Sparse Representation

Time Series Clustering Ensemble Algorithm Based on Locality Preserving Projection

Dimensionality Reduction and Classification through PCA and LDA

Modelling and Visualization of High Dimensional Data. Sample Examination Paper

Extended Isomap for Pattern Classification

Learning a Manifold as an Atlas Supplementary Material

Eigenfaces and Fisherfaces A comparison of face detection techniques. Abstract. Pradyumna Desale SCPD, NVIDIA

Neurocomputing. Laplacian bidirectional PCA for face recognition. Wankou Yang a,n, Changyin Sun a, Lei Zhang b, Karl Ricanek c. Letters.

Generalized Autoencoder: A Neural Network Framework for Dimensionality Reduction

SELECTION OF THE OPTIMAL PARAMETER VALUE FOR THE LOCALLY LINEAR EMBEDDING ALGORITHM. Olga Kouropteva, Oleg Okun and Matti Pietikäinen

Texture Representations Using Subspace Embeddings

GENDER CLASSIFICATION USING SUPPORT VECTOR MACHINES

Class-dependent Feature Selection for Face Recognition

Application of 2DPCA Based Techniques in DCT Domain for Face Recognition

Face Recognition Using SIFT- PCA Feature Extraction and SVM Classifier

SINGLE-SAMPLE-PER-PERSON-BASED FACE RECOGNITION USING FAST DISCRIMINATIVE MULTI-MANIFOLD ANALYSIS

On Modeling Variations for Face Authentication

A new Graph constructor for Semi-supervised Discriminant Analysis via Group Sparsity

A Direct Evolutionary Feature Extraction Algorithm for Classifying High Dimensional Data

Emotion Classification

FADA: An Efficient Dimension Reduction Scheme for Image Classification

Data fusion and multi-cue data matching using diffusion maps

Locally Linear Landmarks for large-scale manifold learning

Image Analysis & Retrieval. CS/EE 5590 Special Topics (Class Ids: 44873, 44874) Fall 2016, M/W Lec 16

Lecture Topic Projects

Semi-supervised Data Representation via Affinity Graph Learning

Facial Expression Recognition with Emotion-Based Feature Fusion

An Improved Face Recognition Technique Based on Modular LPCA Approach

Distance-driven Fusion of Gait and Face for Human Identification in Video

FACE RECOGNITION BASED ON GENDER USING A MODIFIED METHOD OF 2D-LINEAR DISCRIMINANT ANALYSIS

Robust face recognition under the polar coordinate system

Learning to Recognize Faces in Realistic Conditions

Remote Sensing Data Classification Using Combined Spectral and Spatial Local Linear Embedding (CSSLE)

Decorrelated Local Binary Pattern for Robust Face Recognition

Subspace manifold learning with sample weights

Research Article A Multifactor Extension of Linear Discriminant Analysis for Face Recognition under Varying Pose and Illumination

Feature-Aging for Age-Invariant Face Recognition

Image Processing and Image Representations for Face Recognition

Face Recognition using Tensor Analysis. Prahlad R. Enuganti

Face recognition based on improved BP neural network

Sparse Manifold Clustering and Embedding

Face Recognition with Weighted Locally Linear Embedding

Estimating Cluster Overlap on Manifolds and its Application to Neuropsychiatric Disorders

Dimension Reduction of Image Manifolds

COMBINING SPEEDED-UP ROBUST FEATURES WITH PRINCIPAL COMPONENT ANALYSIS IN FACE RECOGNITION SYSTEM

Research Article Chaotic Neural Network for Biometric Pattern Recognition

FACE RECOGNITION FROM A SINGLE SAMPLE USING RLOG FILTER AND MANIFOLD ANALYSIS

Class-Information-Incorporated Principal Component Analysis

Transcription:

The Analysis of Parameters t and k of LPP on Several Famous Face Databases Sujing Wang, Na Zhang, Mingfang Sun, and Chunguang Zhou College of Computer Science and Technology, Jilin University, Changchun 13001, China {wangsj08, nazhang08}@mails.jlu.edu.cn; cgzhou@jlu.edu.cn Abstract. The subspace transformation plays an important role in the face recognition. LPP, which is so-called the Laplacianfaces, is a very popular manifold subspace transformation for face recognition, and it aims to preserve the local structure of the samples. Recently, many variants of LPP are proposed. LPP is a baseline in their experiments. LPP uses the adjacent graph to preserve the local structure of the samples. In the original version of LPP, the local structure is determined by the parameters t (the heat kernel) and k (k-nearest neighbors) and directly influences on the performance of LPP. To the best of our knowledge, there is no report on the relation between the performance and these two parameters. The objective of this paper is to reveal this relation on several famous face databases, i.e. ORL, Yale and YaleB. Keywords: locality preserving projection; the adjacent graph; the nearest neighbors; heat kernel; parameters set 1 Introduction As one of the most important biometric techniques, face recognition has gained lots of attentions in pattern recognition and machine learning areas. The subspace transformation plays an important role in the face recognition. Feature extraction is one of the central issues for face recognition. Subspace transformation (ST) is often used as a feature extraction method. The idea of ST is to project the feature from the original high dimensional space to a low dimensional subspace, which is called projective subspace. In the projective subspace, the transformed feature is easier to be distinguished than the original one. Principal Component Analysis (PCA)[1] is a widely used subspace transformation. It attempts to find the projective directions to maximize variance of the samples. To improve classification performance, LDA[1] encodes discriminant information by maximizing the ratio between the between-class and within-class scatters. LDA can be thought of as an extension with discriminant information ofpca. BothPCAand LDAfocusonpreservingthe globalstructureofthe samples. However, Seung[10] assumed that the high dimensional visual image information in the real world lies on or is close to a smooth low dimensional manifold. Corresponding author

Inspired by this idea, multiple manifold dimensionality reduction methods that preserve the local structure of samples have been proposed, such as ISOMAP[11], LLE[9], Laplacian Eigenmaps[] etc. Locality Preserving Projections (LPP)[5] is a linear Laplacian Eigenmaps. Its performance is better than those of PCA and LDA for face recognition[6]. Recently, many variants[15][3][13][16][7][14] of LPP are proposed. LPP is a baseline in their experiments. However, the performance of LPP depends mainly on its underlying adjacent graph whose construction suffers from the following points: (1)such adjacent graph is artificially constructed; () it is generally uneasy about assigning appropriate values for the neighborhood size k and heat kernel parameter t involved in graph construction. To the best of our knowledge, there is no report on the relation between the performance and these two parameters k and t. The objective of this paper is to reveal this relation on several famous face databases. Locality Preserving Projections Given a set of N samples X = {x 1,x,...,x N },x i R D, we attempt to find a transformation matrix W of size D d to map: y i = W T x i,y i R d, such that y i easier to be distinguished in the projective subspace. Locality Preserving Projections(LPP)[5] attempts to preserve the local structure of the samples in the low-dimensional projected subspace as much as possible. The local structure of the samples is measured by constructing the adjacency graph G. There are two ways to construct G: ε neighborhoods and k nearest neighbors. The similarity matrix S is defined by the following two ways: 1. 0-1 ways S ij = { 1 nodes i and j are connected in G 0 otherwise.. Heat kernel { exp( x i x j /t ) nodes i and j are connected in G S ij = 0 otherwise. (1) () where t is a parameter that can be determined empirically. When t is large enough, exp( x i x j /t) = 1, heat kernel becomes 0-1 ways. Obviously, 0-1 ways is a special case of the heat kernel. In order to contain no any discriminant information, we do not use any label information to construct the similarity matrix S. The criterion function of LPP is as follows: min W (y i y j ) S ij (3) i,j The criterion function incurs a heavy penalty if neighboring points x i and x j are mapped far apart. Therefore, minimizing it is an attempt to ensure that if x i and x j are close, then y i and y j are close, as well. Finally, the transformation

matrix consists of the eigenvectors associated with the smallest eigenvalues of the following generalized eigenvale problem: XLX T w = λxdx T w (4) where D is a diagonal matrix; its entries D ii = j S ij measure the local density around x i. L = D S is the Laplacian matrix. We define S L = XLX T and S D = XDX T, and rewrite Eq. (4) as follows: S L w = λs D w (5) Theorem 1. Let N and D be the dimension of the sample and the number of the samples,respectively.if N > D, then the rank of S L is at most N 1 and the rank of S D is at most N. Proof. According to the definition of the Laplacian matrix and the fact that the similarity matrix is symmetrical. j w 1j w 11 w 1 w 1N w 1 j L = w j w w N. (6)..... w 1N w N j w Nj w NN we add the rd, 3nd,... Nth rows into the 1st row, and obtain L = 0. So, the rake of L is at most N 1. It is known that the maximum possible rank of the product of two matrices is smaller than or equal to the smaller of the ranks of the two matrices. Hence, rank(s L ) = rank(xlx T ) N 1. Similarly, we have rank(s L ) N. From Theorem 1, LPP like LDA also suffers from the SSS problem. Another problem is how to measure he local structure of the samples. LPP uses the similarity matrix S. If every entries are the same, the local structure of the samples is not preserved. Without loss of generality, each entry in S is set as 1/N, i.e., L = 1 N I 1 N ee T, where e ia a vector, whose entries are 1. The matrix S L is equivalent to the covariance matrix in PCA[6]. In this case, LPP degenerates into PCA. Obviously, the performance of LPP dependents on how construct the similarity matrix S. In next section, the performance of LPP with respect to the neighborhood size k and heat kernel parameter t on several famous face databases will be reported. 3 Experiment 3.1 Database and experimental set Threewell-knownfacedatabaseORL 1,Yale andtheextendedyalefacedatabase B[4] (denoted by YaleB hereafter) were used in our experiments. 1 http://www.cl.cam.ac.uk/research/dtg/attarchive/facedatabase.html http://cvc.yale.edu/projects/yalefaces/yalefaces.html

The ORL database collects images from 40 individuals, and 10 different images are captured for each individual. For each individual, the images with different facial expressions and details are obtained at different times. The face in the images may be rotated, scaled and be tilting in some degree. The sample images of one individual from the ORL database are shown in Figure 1. Fig. 1. Sample images of one individual from the ORL database. There are total of 165 gray scale images for 15 individuals where each individual has 11 images in Yale face database. The images demonstrate variations in lighting condition, facial expression (normal, happy, sad, sleepy, surprised, and wink). The sample images of one individual from the Yale database are showed in Figure. Fig.. Sample images of one individual in the YALE database. The YaleB contains 1888 images of 38 individuals under 9 poses and 64 illumination conditions. A subset containing 414 frontal pose images of 38 individuals under different illuminations per individual is extracted. The sample images of one individual from the YaleB database are showed in Figure 3. 3. the analysis of the performance with respect to t and k In our experiments, the similarity matrix S is governed by two parameters: the neighborhoodsizek andheatkernelparametert.kissearchedfrom{,3,...,n 1}. Each image vector is normalized before calculating the similarity matrix S. We randomly split the image samples so that p (for ORL and yale, p =,3,4,5,6,7,8; for YaleB, p = 5,10,0,30,40,50)images for each individual are used as the training set and the rest are used as the testing set. In yale database, p is set as i.e. 30 samples in the training set, and t is searched from {1,1.1,1.,...,1.9,,3,...,9,10,0,...,90,100}. This process is repeated 50 times. The results of three of them are plotted in Fig. 4. From the figure, little influence has been brought to the performances due to the variation of t value. Due to normalization of the image vectors, when t >, exp( x i x j /t) approximates 1. From the figure, we can also see that in

Fig. 3. Sample images of one individual from the YaleB database. 9 9 0.44 9 0.48 0.5 0.46 15 8 0.5 0.48 0.46 0.44 0.4 15 8 0.4 0.4 0.38 0.36 0.34 15 8 0.44 0.4 0.4 0.38 0.36 0.34 0.4 0.3 0.3 0.3 1 1.5 5 10 50 100 1 1.5 5 10 50 100 1 1.5 5 10 50 100 Fig.4. The performance of LPP vs. the two parameters k and t on Yale face database

319 40 160 80 1 1.5 5 0.94 0.9 0.9 0.88 0.86 0.84 0.8 0.8 0.78 0.76 0.74 189 143 95 48 1 1.5 5 0.54 0.53 0.5 0.51 0.5 0.49 0.48 the same the number of the training samples, the top performance does not also incur in the same neighborhood size. The same experiments are conducted on the ORL, Yale, YaleB face databases. t is searched from {1,1.1,1.,...,1.9,,3,4,5}. The several results are plotted in Fig. 5. From the figure, the top performance incurs on when the neighborhood size k is greater than the half of the numbers of the samples. And we can also see the fact that the performance is sensitive to the parameter k. This stems from the fact that the essential manifold structure of samples. An alternative interpretation is that facial images lie on multi-manifolds instead of a single manifold. Recently, the efforts of research on multi-manifolds for face recognition are proposed[8]. In order to verify the validation of the assumption that the performance is insensitive to the heat kernel parameter t and the top performance incurs in the case that the neighbors size k is greater than the half of the number of the samples, 50-time cross-validations are performed on Yale database. The results are illustrated in Fig. 6. (a) 30 samples on ORL (b) 190 samples on YaleB Fig.5. The performance of LPP vs. the two parameters k and t 4 Conclusion LPP is a very popular subspace transformation method for face recognition. Recently, its many variants have been proposed. However, their performances mainly depend on how to construct the adjacent graph, which artificially constructs the local structure. To the best of our knowledge, there is no report on the relation between the performance of LPP vs. the nearest neighbor size k and the heat kernel parameter t. This issue is discussed in this paper. We find that the performance is insensitive to the heat kernel parameter t and the top performance incurs in the case that the neighbors size k is greater than the half of the number of the samples. Ourfuture researcheswill focusonthe performanceofthevariantsoflppvs. the two parameters t and k. We also focus on the multi-manifold face recognition.

0.75 0.7 Recognition accuracy (%) 0.65 0.6 0.55 0.5 BASE PCA LPP 0.45 3 4 5 6 7 8 The number of training samples Fig. 6. the grid-search parameter result on yale database. References 1. Belhumeur, P.N., Hespanha, J.P., Kriegman, D.J.: Eigenfaces vs. fisherfaces: recognition using class specific linear projection. IEEE Transactions on Pattern Analysis and Machine Intelligence 19(7), 711 70 (Jul 1997). Belkin, M., Niyogi, P.: Laplacian eigenmaps and spectral techniques for embedding and clustering. Advances in neural information processing systems 1, 585 59 (00) 3. Chen, S.B., Zhao, H.F., Kong, M., Luo, B.: D-LPP: a two-dimensional extension of locality preserving projections. Neurocomputing 70(4-6), 91 91 (007) 4. Georghiades, A., Belhumeur, P., Kriegman, D.: From few to many: illumination cone models for face recognition under variable lighting and pose. IEEE Transactions on Pattern Analysis and Machine Intelligence 3(6), 643 660 (001) 5. He, X.F., Niyogi, P.: Locality preserving projections. In: Advances in neural information processing systems. vol. 16, pp. 153 160. The MIT Press (004) 6. He, X.F., Yan, S.C., Hu, Y.X., Niyogi, P., Zhang, H.J.: Face recognition using laplacianfaces. IEEE Transactions on Pattern Analysis and Machine Intelligence 7(3), 38 340 (005) 7. Liu, Y., Liu, Y., Chan, K.C.C.: Tensor distance based multilinear locality-preserved maximum information embedding. IEEE Transactions on Neural Networks 1(11), 1848 1854 (010) 8. Park, S., Savvides, M.: An extension of multifactor analysis for face recognition based on submanifold learning. In: Computer Vision and Pattern Recognition (CVPR), 010 IEEE Conference on. pp. 645 65. IEEE (010) 9. Roweis, S., Saul, L.: Nonlinear dimensionality reduction by locally linear embedding. Science 90(5500), 33 (000)

10. Seung, H.S., Lee, D.D.: The manifold ways of perception. Science 90(5500), 68 69 (000) 11. Tenenbaum, J., Silva, V., Langford, J.: A global geometric framework for nonlinear dimensionality reduction. Science 90(5500), 319 (000) 1. Turk, M., Pentland, A.: Eigenfaces for recognition. Journal of Cognitive Neuroscience 3(1), 71 86 (Jan 1991) 13. Wan, M.H., Lai, Z.H., Shao, J., Jin, Z.: Two-dimensional local graph embedding discriminant analysis(dlgeda) with its application to face and palm biometrics. Neurocomputing 73(1-3), 197 03 (Dec 009) 14. Xu, Y., Zhong, A., Yang, J., Zhang, D.: LPP solution schemes for use with face recognition. Pattern Recognition (010) 15. Yu, W.W., Teng, X.L., Liu, C.Q.: Face recognition using discriminant locality preserving projections. Image and Vision Computing 4(3), 39 48 (006) 16. Yu, W.: Two-dimensional discriminant locality preserving projections for face recognition. Pattern Recognition Letters 30(15), 1378 1383 (Nov 009)