Technical Report. Title: Manifold learning and Random Projections for multi-view object recognition

Similar documents
The Analysis of Parameters t and k of LPP on Several Famous Face Databases

School of Computer and Communication, Lanzhou University of Technology, Gansu, Lanzhou,730050,P.R. China

Robust Pose Estimation using the SwissRanger SR-3000 Camera

Globally and Locally Consistent Unsupervised Projection

Manifold Learning for Video-to-Video Face Recognition

Linear Discriminant Analysis in Ottoman Alphabet Character Recognition

Image-Based Face Recognition using Global Features

Dimension Reduction CS534

SELECTION OF THE OPTIMAL PARAMETER VALUE FOR THE LOCALLY LINEAR EMBEDDING ALGORITHM. Olga Kouropteva, Oleg Okun and Matti Pietikäinen

Isometric Mapping Hashing

Laplacian MinMax Discriminant Projection and its Applications

Appearance Manifold of Facial Expression

Locality Preserving Projections (LPP) Abstract

Locality Preserving Projections (LPP) Abstract

Head Frontal-View Identification Using Extended LLE

Extended Isomap for Pattern Classification

Using Graph Model for Face Analysis

Selecting Models from Videos for Appearance-Based Face Recognition

Non-linear dimension reduction

Image Similarities for Learning Video Manifolds. Selen Atasoy MICCAI 2011 Tutorial

Large-Scale Face Manifold Learning

Emotion Classification

Remote Sensing Data Classification Using Combined Spectral and Spatial Local Linear Embedding (CSSLE)

Recognizing Handwritten Digits Using the LLE Algorithm with Back Propagation

Face Recognition using Laplacianfaces

Learning to Recognize Faces in Realistic Conditions

Learning a Manifold as an Atlas Supplementary Material

An Integrated Face Recognition Algorithm Based on Wavelet Subspace

Modelling and Visualization of High Dimensional Data. Sample Examination Paper

Dimension Reduction of Image Manifolds

Learning a Spatially Smooth Subspace for Face Recognition

A Discriminative Non-Linear Manifold Learning Technique for Face Recognition

Sparsity Preserving Canonical Correlation Analysis

FACE RECOGNITION USING SUPPORT VECTOR MACHINES

Texture Representations Using Subspace Embeddings

A Supervised Non-linear Dimensionality Reduction Approach for Manifold Learning

Discriminative Locality Alignment

A New Multi Fractal Dimension Method for Face Recognition with Fewer Features under Expression Variations

Linear local tangent space alignment and application to face recognition

Local Discriminant Embedding and Its Variants

Multidirectional 2DPCA Based Face Recognition System

Linear Laplacian Discrimination for Feature Extraction

Face Recognition using Tensor Analysis. Prahlad R. Enuganti

Generalized Autoencoder: A Neural Network Framework for Dimensionality Reduction

CSE 6242 A / CS 4803 DVA. Feb 12, Dimension Reduction. Guest Lecturer: Jaegul Choo

Nonlinear Dimensionality Reduction Applied to the Classification of Images

Robust Face Recognition Using Enhanced Local Binary Pattern

Tensor Sparse PCA and Face Recognition: A Novel Approach

Time Series Clustering Ensemble Algorithm Based on Locality Preserving Projection

Face Recognition with Weighted Locally Linear Embedding

Data fusion and multi-cue data matching using diffusion maps

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 14, NO. 1, JANUARY Face Recognition Using LDA-Based Algorithms

Lecture Topic Projects

A New Orthogonalization of Locality Preserving Projection and Applications

Face Recognition Based on LDA and Improved Pairwise-Constrained Multiple Metric Learning Method

Dimensionality Reduction and Classification through PCA and LDA

Manifold Clustering. Abstract. 1. Introduction

Dynamic Facial Expression Recognition Using A Bayesian Temporal Manifold Model

FACE RECOGNITION BASED ON GENDER USING A MODIFIED METHOD OF 2D-LINEAR DISCRIMINANT ANALYSIS

Laplacian Faces: A Face Recognition Tool

Linear Discriminant Analysis for 3D Face Recognition System

Semi-supervised Data Representation via Affinity Graph Learning

Class-Information-Incorporated Principal Component Analysis

SINGLE-SAMPLE-PER-PERSON-BASED FACE RECOGNITION USING FAST DISCRIMINATIVE MULTI-MANIFOLD ANALYSIS

A new Graph constructor for Semi-supervised Discriminant Analysis via Group Sparsity

Diagonal Principal Component Analysis for Face Recognition

Johnson-Lindenstrauss Lemma, Random Projection, and Applications

Facial Expression Recognition using Principal Component Analysis with Singular Value Decomposition

Proximal Manifold Learning via Descriptive Neighbourhood Selection

Neurocomputing. Laplacian bidirectional PCA for face recognition. Wankou Yang a,n, Changyin Sun a, Lei Zhang b, Karl Ricanek c. Letters.

Pattern Recognition 45 (2012) Contents lists available at SciVerse ScienceDirect. Pattern Recognition

Application of 2DPCA Based Techniques in DCT Domain for Face Recognition

Restricted Nearest Feature Line with Ellipse for Face Recognition

Sensitivity to parameter and data variations in dimensionality reduction techniques

A KERNEL MACHINE BASED APPROACH FOR MULTI- VIEW FACE RECOGNITION

Embedded Manifold-Based Kernel Fisher Discriminant Analysis for Face Recognition. Guoqiang Wang, Nianfeng Shi, Yunxing Shu & Dianting Liu

Face Recognition Using SIFT- PCA Feature Extraction and SVM Classifier

Announcements. Recognition I. Gradient Space (p,q) What is the reflectance map?

A Hierarchical Face Identification System Based on Facial Components

GENDER CLASSIFICATION USING SUPPORT VECTOR MACHINES

Neighbor Line-based Locally linear Embedding

Fuzzy Bidirectional Weighted Sum for Face Recognition

Data-Dependent Kernels for High-Dimensional Data Classification

Keywords: Linear dimensionality reduction, Locally Linear Embedding.

FACE RECOGNITION FROM A SINGLE SAMPLE USING RLOG FILTER AND MANIFOLD ANALYSIS

Locality Condensation: A New Dimensionality Reduction Method for Image Retrieval

Dynamic transition embedding for image feature extraction and recognition

Deep Learning for Computer Vision

Face Recognition Using Wavelet Based Kernel Locally Discriminating Projection

Image Processing and Image Representations for Face Recognition

Similarity Estimation using Bayes Ensembles

Semi-supervised Learning by Sparse Representation

Stepwise Nearest Neighbor Discriminant Analysis

Research Article Chaotic Neural Network for Biometric Pattern Recognition

Non-Local Estimation of Manifold Structure

Non-Local Manifold Tangent Learning

Iterative random projections for high-dimensional data clustering

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma

CSE 6242 A / CX 4242 DVA. March 6, Dimension Reduction. Guest Lecturer: Jaegul Choo

An Efficient Secure Multimodal Biometric Fusion Using Palmprint and Face Image

Transcription:

Technical Report Title: Manifold learning and Random Projections for multi-view object recognition Authors: Grigorios Tsagkatakis 1 and Andreas Savakis 2 1 Center for Imaging Science, Rochester Institute of Technology, PhD Student 2 Professor and Head of Computer Engineering, Rochester Institute of Technology Purpose: discussion of the outcomes of the work supported by the CIS-Kodak micro fund. Resulted publications: 1. Grigorios Tsagkatakis and Andreas Savakis, Manifold Modeling with Learned Distance in Random Projection Space for Face Recognition, 20th International Conference on Pattern Recognition (ICPR), pp.653-656, Istanbul, Turkey, August 23-26 2010. 2. Grigorios Tsagkatakis and Andreas Savakis, Face Recognition Using Sparse Representations and Manifold Learning, Advances in Visual Computing, Springer Lecture Notes in Computer Science, pp. 509-518, Las Vegas, NV, November 29 Dec 1 2010. Abstract Recognizing objects from different viewpoints is a challenging task. One approach for handling this task is to model the appearance of an object under different viewing conditions using a low dimensional subspace. Manifold learning describes the process by which this low dimensional embedding can be generated. However, manifold learning is an unsupervised method and thus gives poor results on classification tasks. To address this impediment of manifold learning, we investigated the combination of manifold learning and distance metric learning for the generation of a representation that is both discriminative and informative. In the proposed system, initial dimensionality reduction is achieved using random projections, a computationally efficient and data independent linear transformation. Distance metric learning is then applied to increase the separation between classes and improve the accuracy of

nearest neighbor classification. Finally, a manifold learning method is used to generate a mapping between the randomly projected data and the low dimensional manifold. We demonstrate that this approach is effective for multi view face recognition. Furthermore, we show that random projections can be applied as an initial step without significantly affecting the classification accuracy. Although we provide results regarding the classification of human faces from different viewpoints, the same approach can be used for the classification of generic objects under different viewing conditions. I. Introduction Images and video convey a wealth of information, yet, extracting a higher level of understanding from data is a computationally complex task. With added constraints, such as processing power, memory and bandwidth, the problem becomes even more challenging, and the high dimensional nature of the data may not be optimal for modeling and computational purposes. Thus, dimensionality reduction becomes a necessary step before more elaborate computer vision algorithms can be applied. Dimensionality reduction is often achieved by principal components analysis (PCA). Even though PCA is a widely used method, certain limitations of the method, such as data dependence and linearity restrictions, are becoming more evident and suggest the investigation of more appropriate dimensionality reduction techniques. Manifolds offer a novel approach in signal representation that has received considerable attention in recent years [11]. Compared to linear techniques, manifold learning can identify non-linear structures and project the data in a low dimensional space, while preserving key properties such as geodesic distances or the local neighborhood structure. A variety of signals can be reliably modeled using manifold models [15]. Manifold learning was successfully applied for face recognition based on Laplacianfaces [1]. The face manifold was identified by applying locality preserving projections (LPP), and then nearest neighbor classification was used for recognition. The method achieved higher recognition accuracy compared to Eigenfaces [16] (PCA based) and Fisherfaces [17] (discriminant based) that model faces using linear structures. Figure 1: Block diagram of the proposed system

Our proposed method is composed of three modules for face representation and classification as shown in Figure 1: (a) random projections (RPs), (b) distance metric learning (DML) and (c) manifold learning (ML). RPs is a data-independent linear dimensionality reduction technique that preserves the manifold structure of the data [7] and is applied as a preprocessing step. DML is then applied in order to generate a distance metric that will increase the separability between classes and improve classification accuracy. Once the new distance metric is learned, it is used during the generation of the distance matrix in ML. During testing a nearest neighbor classifier is employed in manifold space and utilizes the learned distance for classification. II. Manifold Learning The goal of manifold learning is to reduce the dimensionality of the input data while preserving its geometric properties. Methods like Isometric Feature Embedding (Isomap) [2] and Local Linear Embedding (LLE) [3] are examples of graph embedding manifold learning approaches. In these methods, the data is described through a distance matrix (adjacency graph) and dimensionality reduction is achieved by applying eigenanalysis on this distance matrix. A problem with manifold learning methods, such as Isomap and LLE, is the lack of a direct mapping from the input space to the manifold space. Instead, they provide a mapping of the input data to the manifold space. This is called the out-of-sample extension problem and can limit the applicability of these methods for new data [12]. More recent methods try to overcome this problem by performing a linear approximation of the manifold based on the idea that even though manifolds are non-linear structures, they can be approximated reasonably well within a small neighborhood using a linear map. The benefits of linear approximations are mostly expressed in terms of savings in computational time, although in some cases, e.g. locality preserving projections, very promising results have been reported. Locality preserving projections (LPP) [1] is the optimal linear approximation of the LLE algorithm and can be described by four steps. First, the adjacency graph G is built, which is similar to LLE and Isomap. Then, the Laplacian of the graph is estimated. The Laplacian is a normalization of the graph that can provide more insights into its structure, and is given by, where and W is the affinity matrix of the graph G. Using the Laplacian, the following generalized eigenvalue problem is solved by Eq. (1) (1)

where is the data matrix. In the final step, the linear mapping that provides the embedded of the input point is generated by forming a matrix. Each column of corresponds to an eigenvector as shown in Eq. (2),,,, (2) Although manifold learning is a powerful method for unsupervised modeling, it does not take class information into account during the embedding. In other words, preserving the structure of the manifold, which is the goal in manifold learning, is independent of the classification, since neighboring points might belong to different classes, e.g. faces of different individuals. The problem was identified in [4] where it was coined classification-oriented multi-manifolds learning. A few supervised manifold learning approaches have been presented in the literature, such as supervised LLE [10], Large Margin Component Analysis [5] and Local Discriminant Embedding [6]. Supervised manifold learning algorithms exploit the class label information during training in order to increase the accuracy. However, once training is performed, the mapping and the classification are usually applied as in unsupervised methods. To handle this shortcoming, we investigate the introduction of distance learning before the application of manifold learning. The benefit of the proposed approach is that class label information is utilized in both the manifold learning and the subsequent classification. III. Distance Metric Learning Typically, the distance between input data is measured using the Euclidian distance. However, when additional information about the data is available, different distance metrics can provide a more meaningful connection between the data. Distance metric learning (DML) has received considerable attention in the past few years and a number of different DML approaches have been presented. In supervised DML, the objective is to learn a new distance that will satisfy the pairwise constrains provided by the class labels. Formally, the distance between two data points and is given by, M (3) where M is the distance metric matrix. Alternatively, M can be written as M L L in which case Eq. (3) becomes, L L (4) and equivalently, the new distance metric is the Euclidean distance between the linearly transformed data.

A well-known example is the Mahalanobis distance where the inverse of the covariance matrix is used as the matrix M in Eq. (4). In this work, we focus on a successful DML algorithm, the Large Margin Nearest Neighbor (LMNN) [8]. LMNN generates a new distance metric with the additional constraint that points from the same class are separated by a large margin from points in different classes. Formally, assume is an input point associated with class and let be a true neighbor, associated with the same class and an imposter associated with a different class. The objective of LMNN is to learn a distance matrix M, or an equivalent linear transformation L such that: L L 1 (5) The objective is achieved through a convex optimization procedure. In the proposed system, DML is applied first in order to identify a distance that satisfies the class constrains. The learned distance, or the equivalent linear transform, is then used in the generation of the adjacency matrix of the LPP algorithm. Once the manifold embedding is learned, classification of new data points is achieved by measuring the learned distance between the embedded training samples and the new data point. Experimental results suggest that the properties of the learned distance, i.e. keeping the true neighbors close and the impostures far away, are preserved through the embedding. The class is identified by the labels of the nearest neighbors. IV. Random Projections Due to the optimization nature of DML, an initial dimensionality reduction, such as PCA, is usually applied before training. We investigate the use of RPs primarily for two reasons. First, RPs is a linear technique and thus offers fast application through matrix multiplications. Furthermore, unlike PCA, RPs is data-independent, i.e. it is does not require training on a particular dataset and thus allows better generalization. RPs is based on the Johnson Lindenstrauss (JL) [9] lemma which states that, given 1>ε>0 and an integer n, select k, such that k O(ε -2 logn), then for every set P of n points in there exists a Lipchitz function f : R d R k such that for all x, y in P: 1 1 (6) As a result, the complexity of algorithms that depend on the dimensions of their input data, such as DML, can be significantly decreased. What is even more surprising is that very efficient methods exist for generating the projection matrices.

A random projection from the original d-dimensional space to the reduced k-dimensional space can be achieved by multiplying the input vectors with a matrix R, where R is a d k matrix. In [9] it was shown that the elements of R can be drawn as independent and identically distributed random variables from the following distribution: 1 with probability 1/6 3 0 with probability 2/3 1 with probability 1/6 (7) The distribution in Eq. (7) is notably efficient, since it discards 2/3 of the data that correspond to multiplication by zero. Furthermore, it can be implemented using fixed point arithmetic operations consisting of only additions and subtractions if the scaling coefficient is factored out. Even though, RPs were initially proposed for subspace dimensionality reduction, the approach was extended to manifold modeled signals in [7] where it was shown that manifold properties, such as geodesic distances, curvature of the paths, etc., are also preserved through the projection. V. Experimental Results Two sets of experiments were carried out. First we investigated how random projections affect the manifold learning algorithm. Then we tested how well manifold modeling with learned distance performed on face recognition, based on the system shown in Fig. 1. The face recognition performance was measured on the AT&T dataset [13], which consists of 40 subjects showing 10 images per subject at varying conditions. For each scenario i.e. number of images per individual, 50 different random splits of the dataset were applied and the mean recognition accuracy is reported. The only preprocessing on the images was the normalization of each face vector. Figure 2 presents the recognition accuracy of nearest neighbor classification when LPP was performed in random projections space versus the original space. The results suggest that the recognition accuracy achieved by applying LPP in the original dimensions is not compromised after dimensionality reduction via RPs when the dimensionality reduction in RP space is kept above 300.

Figure 2: LPP recognition accuracy as a function of input data dimensionality In the second experiment, we investigated the recognition accuracy of the proposed system using a learned distance in a 500-dimensional random projected space. Table 1 shows the results of our proposed method compared to standard methods for face recognition as reported in [14]. We note that LPP generates the distance matrix using same class neighbors only. The first column of the table corresponds to the number of images per individual that were used for training while the rest were used for testing. We observe that combining DML and LPP significantly increases the recognition accuracy. Table 1: Face recognition accuracy on the AT&T dataset Train Img/ PCA LDA LPP Proposed Individual 5 86.3 92.8 93.1 98.2 4 82.1 89.4 90.4 91.4 3 76.6 84 86.6 87.7 2 66.9 71.5 76.1 83.7 VI. Conclusions In this report, we describe an object recognition system that combines random projections for efficient dimensionality reduction, distance metric learning for increased classification performance, and manifold learning for embedding the data in a low dimensional manifold. Results suggest that random projections transformation does not affect classifier performance and that the combination of distance metric learning and manifold learning can increase the recognition accuracy of a classifier. The proposed scheme can be useful for face recognition and other applications, and may be extended to different combinations of distance metric learning and manifold learning approaches.

References [1] X. He, S. Yan, Y. Hu, P. Niyogi, and H.J. Zhang. Face recognition using Laplacianfaces. IEEE Trans. Pattern Analysis Machine Intelligence 27(3): 328-340, 2005. [2] J.B. Tenenbaum, V. Silva, and J.C Langford. A global geometric framework for nonlinear dimensionality reduction. Science (290): 2319-2323, 2000. [3] S.T. Roweis, and L.K. Saul. Nonlinear dimensionality reduction by locally linear embedding. Science (290): 2323-2326, 2000. [4] J. Yang, D.D. Zhang, J. Yang, and B. Niu. Globally maximizing, locally minimizing: unsupervised discriminant projection with applications to face and palm biometrics. IEEE Trans. Pattern Analysis Machine Intelligence 29(4): 650-664, 2007. [5] L.Torresani, and K. Lee. Large margin component analysis. Advances in NIPS (19): 1385-1392, 2007. [6] H. Chen, H. Chang and T. Liu. Local discriminant embedding and its variants. IEEE Computer Vision and Pattern Recognition Conference (2): 846-853, 2005. [7] R. Baraniuk and M. Wakin, Random projections of smooth manifolds. Found. of Comp. Math. 9(1): 51-77, 2009. [8] K.Q. Weinberger and L.K. Saul. Distance metric learning for large margin nearest neighbor classification. Journal of Machine Learning Research (10): 207-244, 2009. [9] D. Achlioptas. Database-friendly random projections: Johnson-Lindenstrauss with binary coins. Jour. of Comp. and Syst. Sc., 66(4): 671-687, 2003. [10] D. de Ridder, O. Kouropteva, O. Okun, M. Pietikainen, and R.P.W. Duin. Supervised locally linear embedding. Lect. Notes in Computer Science (2714): 333-341, 2003. [11] R. Pless, and R. Souvenir. A Survey of Manifold Learning for Images. IPSJ Trans. on Computer Vision and Applications (1): 83-94, 2009. [12] Y. Bengio, J. Paiement, P. Vincent, O. Delalleau, N. Roux, and M. Ouimet. Out-of-Sample Extensions for LLE, ISPMAP, MDS, Eigenmaps, and Spectral Clustering. Advances in NIPS (16), 2004. [13] AT&T dataset: http://www.cl.cam.ac.uk/research/dtg/ attarchive/facedatabase.html [14] D. Cai, X. He, J. Han, and H. Zhang. Orthogonal laplacianfaces for face recognition. IEEE Trans.on Image Processing 15(11): 3608-3613, 2006. [15] G Peyré. Manifold models for signals and images. Com. Vis. and Image Understanding 113(2): 249-- 260, 2009. [16] M. Turk, and A. Pentland. Eigenfaces for recognition. Journal of cognitive neuroscience 3(1): 71-86, 1991. [17] P.N. Belhumeur, J.P. Hespanha, and D.J. Kriegman. Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection. IEEE Trans. Pattern Analysis Machine Intelligence 19(7): 711-720, 1997.