Robust Landmark Localization and Tracking for Improved Facial Expression Analysis

Size: px
Start display at page:

Download "Robust Landmark Localization and Tracking for Improved Facial Expression Analysis"

Transcription

1 Robust Landmark Localization and Tracking for Improved Facial Expression Analysis Hamdi Dibeklioğlu, Albert Ali Salah, Theo Gevers Intelligent Systems Lab Amsterdam, University of Amsterdam {h.dibeklioglu, a.a.salah, Keywords: Facial landmarking, facial expression analysis, structural prior Abstract Automatic facial expression analysis requires localization and tracking of facial features in an accurate and robust manner. We describe a statistical method for automatic facial landmark localization, which is then used to improve a state-ofthe-art deformable tracking based facial expression recognition algorithm. Our landmarking is based on Gabor wavelet features modeled with mixtures of factor analyzers. Once the landmarks are automatically located, they are used to initialize a tracker that maintains a simplified 3D face model matched to the appearance of the face. The deformations are classified with a naïve Bayes scheme into six expression categories. We test the proposed methods extensively on cross-database experiments conducted on FRGC, Cohn-Kanade, and Bosphorus face datasets. Our results show that the statistical landmarking method we propose is robust in its generalization and increases the accuracy of expression recognition significantly. 1 Introduction Evaluation of facial expression relies on accurate face detection, face registration, localization of fiducial points in faces, and classification of shape/appearance information into expressions. When temporal information is available, tracking and temporal modeling also enter the picture. The pipeline of a facial expression analysis method starts with face detection, and often proceeds by locating several fiducial points on detected faces, also called anchor points, or landmarks. Usually, corners of eyes and eyebrows, centers of irises, nose tip, mouth corners, and the tip of the chin are used as key landmarks. These key landmarks are generally sufficient for face registration. However, more landmarks ( to points) are required in facial expression analysis. Facial surface deformations caused by expressions can be described by movements of selected facial feature points. If these points are discriminative enough, and are detected accurately, deformation analysis can classify facial expressions. Nevertheless changing pose, resolution and illumination conditions make facial landmark localization a challenging problem. Especially, statistical models can fail if the variation shown in the training set is not sufficient enough for generalization of unseen test samples. In this paper we describe a system for facial expression analysis which encompasses six basic emotional expressions (i.e. happiness, fear, surprise, anger, disgust, and sadness). Our emphasis is however on automatic landmarking, the effect of which we assess on subsequent expression categorization. We improve a recent facial landmarking algorithm by introducing a prior conditioned on face detection. We show via extensive cross-database tests that the landmarking method we use is robust. Our subsequent tests on emotion recognition establish that accurate statistical descriptions of several landmarks can be successfully used to improve expression analysis through tracking; we obtain as much as per cent classification accuracy improvement through the use of statistical landmarking. This paper is structured as follows. In Section 2, an overview of the system is presented. Section 3 describes related work in landmarking, followed by Section 4 that describes our statistical landmark localization algorithm. In Section 5, we present the algorithms for model-based facial tracking and expression analysis. The experimental results are presented in Section 6, followed by our conclusions in Section 7. 2 Overview of the System In this section, we briefly describe the overall pipeline of the system (See Figure 1). The camera 1

2 (or video) input is analyzed first for the presence of a face. We use the Viola-Jones algorithm for face detection [23]. While the system supports continuous expression evaluation, the analysis is differential and the best results are obtained by presenting a neutral, frontal face image during the initialization of the tracker. Data acquisition Face detection Histogram equilization Gabor wavelets Facial feature localization Tracker initialization Facial feature tracking Expression classification Viola-Jones algorithm IMoFA model for features GOLLUM structural analysis algorithm TPS algorithm Active appearance model PBDV tracker Motion unit extraction Naïve Bayes classifier Figure 1: Overview of the system. Once the face is located, the landmarking module is used to locate seven facial features. These are the inner and outer eye corners, the nose tip and the mouth corners. The statistical feature modeling approach via incremental mixtures of factor analyzers (IMoFA) algorithm [17], as well as the GOLLUM structural analysis algorithm to correct landmarking errors [18], will be detailed in Section 4. The located landmarks are used to initiate a piecewise Beziér volume deformation tracker [21]. An initial generic face mesh is warped to the shape indicated by the detected landmarks with the thin-plate spline algorithm [3]. The tracking module also has a general appearance model associated with it to prevent the accumulation of tracking errors [6]. This model is enhanced with a virtual structuring element to reduce the effect of different backgrounds in training and testing conditions [22]. The expression classification is performed on a set of motion units, which indicate the direction of movement of several key points on the Beziér volume with respect to the initial, neutral frame. We use a naïve Bayes classifier to compute the posterior probabilities for the six basic expression categories. 3 Related Work in Landmarking Facial landmarking is the first and indispensable step of many human-related computer vision applications especially for facial expression analysis. Expression analysis requires costly manual initialization in the absence of a reliable landmarking algorithm [5]. Automatic localization of facial feature points is a challenging problem because of varying acquisition conditions (i.e. illumination and resolution) and occlusion problems. Besides, spatial relations of facial landmarks show variations for different face images. Facial deformations based on expressions, and acquisition noise also affect patterns of landmarks. Facial landmarking techniques can be divided into two main categories as texture-based and structural methods. Texture-based methods use features based on intensity information and texture values around the related landmarks. Localization of the eye and mouth regions by vertical projection histograms of intensity values [4, 28], and eye detection by using the contrast differences in the eye region [1, 11] can be counted as examples for texture-based landmarking. Structural methods incorporate shape information by looking at the relative positions of landmarks, combining the shape features with local appearance features. Active appearance models are typical examples for structural landmarking, and many variants on this method are introduced [26]. Additionally, there are several studies that use both structural relationships between landmark locations, and local feature constraints [, 25]. In [25], Gabor feature jets are used to model local features, and the best match for each feature point is found by an exhaustive search on template library (called the bunch). In [7], Cristinacce and Cootes used an AdaBoost classifier based on Haar wavelet features [23] for coarse-level localization and an Active Shape Model (ASM) was employed for fine-tuning. ASM approaches jointly estimate all landmarks, combining local feature information with shape. For cases where local features are missing or misleading (e.g. natural occlusions, scars, accessories) joint estimation methods will experience problems. Furthermore, ASM and similar methods require greater number of landmarks for training; usually - landmarks are used per face. However, none of the major databases has this kind of ground truth. Consequently, [24] and [18] focus on independent detection of landmark points. In [24], Vukadinovic and Pantic have proposed a method that uses boosted classifiers based on gray scale texture information and Gabor wavelet features. After the detected face region is divided into regions of interest (ROI), GentleBoost templates are used to detect landmarks within the relevant ROI, independently. However, this method is not very effective for expressive faces, as it uses 2

3 heuristic location assumptions for ROI specification. Salah et al. proposed a coarse-to-fine strategy for independent feature point detection in [18]. On the coarse scale, Gabor wavelet features are statistically modeled, and complemented with a structural analysis step. Then, discrete cosine transform (DCT) features are used for fine-tuning of the coarsely localized landmarks. In this approach, the whole image is searched for each landmark. However, the search space can be drastically constrained by face detection through a shape prior. Unlike using discrete regions as defined in ROI-based methods, this kind of shape prior would estimate continuous probability values for the search space. We introduce such a prior in this paper and show that through such a prior, contrasted landmarking methods become faster and more accurate. 4 The Facial Landmarking Algorithm 4.1 Facial Model and Features In the onset of our approach, we detect the face with the Viola-Jones face detection algorithm [23]. We therefore assume that the face image is roughly frontal. Occasional detection failures do not affect the method because system works on camera input (or video), and switches to tracking once the face is detected. To reduce the computational complexity, we downsample the cropped high resolution face images into pixels and apply a histogram equalization to the cropped area for damping the illumination effects. Let (x,y) denote the relative coordinates of a particular landmark within the cropped face area. A prior distribution p(x,y l i ) is learned for each landmark i during the training stage. Here, i = 1...7, and l i denotes the selected landmark. This shape prior is well approximated with a 2-dimensional Gaussian distribution. We constrain the search for landmarks to areas delimited by the shape of this prior distribution. When compared to the approach presented in [18], this constraint provides an immediate speed-up in feature extraction (up to two orders of magnitude), as well as making use of the structural information between individual landmarks. Independent Gabor wavelet feature patches are extracted from around a 7 7 neighborhood of each landmark candidate. The resulting feature vectors (49-dimensional) are calculated as follows, for each orientation w and scale v: Ψ j ( x) = k j k T j σ e ( 2 kj = kjx, ) k jy ( k j k T j x x T 2σ 2 ) [ e (i k j x T) e ( σ 2 = (k v cosφ w,k v sinφ w ) 2 )] k v = 2 v+2 2 π,φ w = w π 8, (1) where x = (x, y) is a candidate landmark location and j is an index for different kernels. k j and x are both expressed as row vectors. The standard deviation of the Gaussian function is selected as σ = 2π. The first factor in the Gabor kernel represents the Gaussian envelope and the second factor represents the complex sinusoidal (carrier) function. The term, e σ2 /2 in the square brackets compensates for the DC value. These features are extracted in 8 orientations, w {0,1,2,3,4,5,6,7}, and at 5 different scales, v {0,1,2,3,4}. The training uses only the positive samples of each landmark class, obtained from ground truth. The features are min-max normalized before statistical modeling. 4.2 Statistical Feature Modeling We follow [18] in our model choice for statistical distribution of extracted features and use the Incremental Mixtures of Factor Analyzers (IMoFA) algorithm. This approach involves placing a number of Gaussian distributions on the data incrementally and in unsupervised fashion [17]. Through a latent variable assumption, the high-dimensional full covariance matrix of each component in the mixture distribution is expressed in a lower dimensional manifold, resulting in a locally adaptive model complexity. Such model simplification in terms of parameters improves generalization, as the number of training samples is fixed. The trade-off between accuracy and complexity is explored in a greedy and unsupervised fashion, and difficult portions of the data are modeled with more parameters automatically. The latent variable assumption reflects on the expression of the data covariance matrix as follows: Σ j = Λ j Λ T j +Ψ, (2) where j stands for the components in the mixture model,σis the covariance matrix (d d),λis the factor loading matrix (d p) that transforms the data from the p-dimensional manifold to the d-dimensional data space, Ψ is a diagonal matrix with d free parameters that stands for independent variance in each dimension. Usually p << d, which makes the model a parsimonious representation. Note that the model complexity is somewhere between a diagonal-covariance and a full-covariance mixture. Given a dataset, the IMoFA algorithm initially fits a Gaussian distribution generated from a single dimensional manifold to it, and then proceeds by increasing the complexity. The two possible operations at any iteration are splitting an existing component, or adding a factor to an existing component. A validation set is monitored for stopping the iterations. At each iteration, the Expectation-Maximization (EM) 3

4 algorithm is used to converge to a locally optimal maximum likelihood solution [9]. If we compare this approach to the popular unsupervised Gaussian mixture model learning procedure (ULF) proposed by Figueiredo and Jain [8], we can note down a few things. The ULF method proposes to fit a large number of Gaussians to data, and proceeds by merging them until a single component remains. Then all intermediate models are evaluated with a Bayesian complexity criterion, and one model is selected. The initial fitting of many components is impractical and costly for problems with high dimensionality. Also, the resulting models in ULF have either full or diagonal covariances. IMoFA explores a range of models in between these extremes, and this property is also very important for high-dimensional problems, since the gap between diagonal and full covariance models widens quadratically with the number of feature dimensions. We extract Gabor features from the neighborhood of each landmark l {1,2,3,4,5,6,7} (four eye corners, two mouth corners and the nose tip). If we denote the Gabor feature vector with u, the mixture model can be written as: p w,v (u) = j p w,v (u G j )p w,v (G j ), (3) wherew andv denote the orientation and scale of the Gabor filter, respectively. G j describes a Gaussian distributionn (µ j,σ j ), andj is an index for the components in the mixture. p w,v (G j ) is the prior probability of the component i along orientation w at scale v, and p w,v (u G j ) is the probability that u is generated by component i. Our approach in this paper is similar to [18], which also uses IMoFA models for Gabor wavelet features. However, a single scale of the Gabor wavelet was used in [18], and no shape prior was considered for conditioning the landmarking on the facial boundary. We have dispensed with the coarse-to-fine processing proposed in [18] in favor of a faster, one-shot detection, but retained the fast structural analysis procedure, which is observed to improve landmarking even with the shape prior we propose. 4.3 Structural Analysis The GOLLUM algorithm is proposed in [18] to test the consistency of a shape, expressed by a set of landmarks, by contrasting it with a set of learned models. The affine normalization of a shape, which is a regular step in shape comparison, can be erroneous if one of the landmarks is off its mark. To remove such effects from normalization, GOLLUM checks all possible normalizations based on landmark triplets, and chooses the one that has the strongest empirical support. Then, the landmarks in the chosen set (a) (b) Figure 2: (a) The Bézier volume model. (b) The motion units. can be used to detect and correct the erroneous landmarks automatically. We refer the reader to [18] for the details of the GOLLUM algorithm. On the FRGC dataset [18] reports 5 to per cent localization improvement with the use of GOLLUM although the ceiling effect prevents a complete assessment. On artificially disturbed data, GOLLUM can correct one or two corrupted landmarks out of seven with more than 98 per cent accuracy. For this reason, it is most useful against total feature occlusions. In Section 6, we report our accuracy results with the GOLLUM algorithm, and report its contribution on the average. 5 Application: Expression Classification Automatically located landmarks are used for initialization in an expression classification application. In recent years, facial expression analysis has been an active topic for researchers. Most studies focus on detecting facial actions units (AUs), and relating expressions to these AUs [2]. Despite a remarkable progress in AU detection, these systems require costly AU annotations for training, prepared by experts. We describe here a simplified approach for classifying six basic emotional expressions. This kind of systems can be trained by using video segments which are annotated for the main expression categories. For more elaborate expression analysis methods and recent developments, see [15] and [27]. Our method for expression analysis is based on [22], which employs a face model described by 16 surface patches embedded in Bézier volumes, shown in Figure 2 (a). To trace the motion of the facial features, a Piecewise Bézier Volume Deformation (PBVD) tracker is used [21]. This tracker is initialized by transforming the face model with respect to the detected face area. The alignment is originally performed by matching the facial boundary to the model. Our contribution to the system is a better initial alignment through the detected landmarks. We use the thin-plate spline (TPS) deformation algorithm to warp the generic face model to the auto- 4

5 matically detected shapes [3]. This approach transforms landmarks exactly to their target points, and interpolates all the other points. During the development of our algorithm, we have also tested Procrustes alignment [] as an alternative, but TPS was more accurate in our experiments. In general, we expect this to be the case when the landmarks are sufficiently accurate. A simple but efficient naïve Bayes classifier is used to classify facial expressions. As an advantage, the posterior probabilities allow a soft output of the system, which is usable as a continuous input to any facial affect-based system. Quantized movement vectors, extracted from a number of locations on the model, are passed as input to the classifier. These vectors are called motion units. Although the motion units are similar to AUs, they are simpler and tailored towards basic expression categories (See Figure 2 (b)). 6 Experimental Results 6.1 Experimental Setup and Data The Bosphorus, FRGC and Cohn-Kanade databases are used to train three versions of our automatic landmarking system. We do this to measure sensitivity to training conditions. Ground truth for seven landmarks on these databases are annotated manually. For testing the facial expression analysis system, the Cohn-Kanade dataset is used. The Bosphorus and FRGC sets are composed of static images, whereas the Cohn-Kanade dataset has video sequences, and is thus adequate for dynamic measurement of expressions. For each database, we use separate training, validation and test partitions, which have completely different subjects from each other. Also, there are no overlapping subjects between databases. Details of the Bosphorus, FRGC and Cohn-Kanade databases are given below. The Bosphorus database [19] for facial expression analysis has systematic expression and pose variations. There are 61 male and 44 female subjects (29 of which are professional actors and actresses) with a total number of 52 face images. The texture images are of high quality, and acquired under controlled studio light. Only frontal samples of the Bosphorus database have been used in our landmark localization experiments. For fair comparison with the results obtained on the Bosphorus dataset, we use the Spring 04 subset of FRGC database [16], which is the most challenging setting. It contains 2114 face images (neutral and uncontrolled expressions) from 465 subjects, in 6 4 resolution acquired under uncontrolled illumination conditions. The Cohn-Kanade AU-Coded Facial Expression Database [12] consists of approximately 0 image Figure 3: Samples from Cohn-Kanade, Bosphorus and FRGC databases, respectively. Table 1: Abbreviations and the size of training, validation, and test sets for each database. Database Abbr. Train Validation Test Bosphorus BOS Cohn-Kanade (Neutral) CK Cohn-Kanade (Expr.) CK-Ext FRGC FRGC sequences from subjects, each starting with a neutral face and showing a basic expression. Only frontal images are open to public use, and we only use those (249 sequences). Image resolution is6 4 pixels. Furthermore, two different datasets are prepared by taking the first (neutral) and the most extreme expression frames of these sequences, respectively. Since there are few sequences in Cohn- Kanade database, we use three-fold cross validation in our expression classifications experiments. All the databases we have used are freely available to the research community. Figure 3 illustrates the scale, pose, expression, resolution and illumination conditions across these three databases. The sizes of training, validation and test sets for each database are given in Table Automatic Landmark Localization Proposed landmarking algorithm has been tested on seven landmarks as four groups: outer eye corner(s), inner eye corner(s), nose tip, and mouth corner(s). Challenging cross-database conditions have been used to measure the accuracy of the landmarking algorithm. The state-of-the-art results in the literature vary 5

6 according to the acceptance threshold they used and the database employed for reporting. The acceptance threshold is the distance from the ground truth location beyond which a given landmark is deemed incorrect. Typically, this threshold is given in terms of inter-ocular distance. For example, in [7], Cristinacce and Cootes use per cent of inter-ocular distance as acceptance threshold, and report 95 and 92 per cent average localization accuracy for finding 22 landmarks on BIOID and XM2VTS databases, respectively. In [24], Vukadinovic and Pantic report 93 per cent average localization accuracy for Cohn- Kanade database by accepting points within per cent of inter-ocular distance to their ground truth locations. Kozakaya et al. report 95.1 per cent average localization accuracy for their weighted vector concentration approach [13] and 94.2 per cent for the modified active shape model presented in [14] on the FERET database, with per cent inter-ocular distance as threshold. In this paper, we report our results with per cent of inter-ocular distance as acceptance threshold. Table 2 shows the correct localization accuracies for different training and test sets. The reported success rates are obtained by accepting points within per cent of inter-ocular distance to the ground truth. Correct localization accuracies for different acceptance conditions are also given in Figure 4. Intersubject variation of the ground truth is about 5-7 per cent of inter-ocular distance. When the proposed system is trained and tested on the same database, the average accuracy is 93.5 per cent for Bosphorus, and 92.7 per cent for FRGC databases. These are comparable to reported state-of-the-art figures. The published implementation of the system proposed in [24] allows us to compare it with our method. Since this system has been trained on Cohn-Kanade database, we have joined neutral and extreme expressions of Cohn-Kanade (denoted as CK-Both) and trained our system for comparison. Our system has 84.5 and 79.6 per cent cross-database accuracy for FRGC and Bosphorus databases, respectively. However, the system proposed in [24] can reach only 79.8 and 77.8 per cent accuracy, respectively for the same datasets. Our experiments show that mouth corners are detected with less accuracy than the other landmarks. This is an expected result as the mouth corners are affected more under expression changes. The variation on training set naturally affects the accuracy of the system. Uncontrolled acquisition of FRGC database provides more pose variations than the frontal subset of Bosphorus. Consequently, the shape prior implicates a larger area and this results in higher accuracy with FRGC training. Although we do not show the net effect of GOLLUM, it increases the accuracy of the proposed method by one per cent on the average. Table 2: Cross-database accuracy of the landmarking algorithm. Database Landmark Type Training Test O. Eye I. Eye Nose Mouth C. BOS BOS FRGC FRGC FRGC BOS FRGC CK FRGC CK-Ext BOS CK BOS CK-Ext BOS FRGC CK-Both BOS CK-Both FRGC Since the time complexity of GOLLUM is negligible, we retain GOLLUM in our system. 6.3 Facial Expression Classification The effect of landmark localization on expression classification is given in Table 3. The results are reported without landmarking (baseline), with landmarking trained on FRGC or Bosphorus databases, and with landmark ground truth for Cohn-Kanade database. For different expressions, the accuracy improvement is different; happiness, fear and disgust greatly benefit from improved alignment. Sadness sadly does not, because it mostly relies on eyebrow movements. Eyebrows do not have consistent statistical properties, and are very difficult to detect with statistical methods. The baseline accuracy of.68 per cent is increased to 75. per cent via FRGC training. 7 Conclusions A statistical landmark localization method has been described with robust cross-database accuracy, and a state-of-the-art facial expression classification algorithm has been improved by the proposed landmarking method. The ground truth for the indicated landmarks would apparently have provided a doubled improvement, which means that there is room for improvement in landmarking, as well as in classification. While the shape prior we have introduced greatly constrains the search for landmarks, cascade-based hierarchical search is still an attractive future direction. The face model we use needs to be enhanced to take into account much greater pose variations. It is a challenge to do this without sacrificing the speed of the expression analysis algorithm, which is around 15 fps in the present setup. As a related future work, the landmarking module will be tested for initializing the tracker on arbitrary poses. 6

7 Tested on BOS Inner Eye C. Outer Eye C. Nose Tip Mouth Corners Accuracy (%) Trained on BOS Trained on FRGC Trained on CK Both Tested on FRGC Inner Eye C. Outer Eye C. Nose Tip Mouth Corners Accuracy (%) Trained on BOS Trained on FRGC Trained on CK Both Tested on CK Inner Eye C. Outer Eye C. Nose Tip Mouth Corners Accuracy (%) Trained on BOS Trained on FRGC Tested on CK-Ext Inner Eye C. Outer Eye C. Nose Tip Mouth Corners Accuracy (%) Trained on BOS Trained on FRGC Figure 4: Landmark localization accuracies on Bosphorus, FRGC and Cohn-Kanade databases with different training conditions. Acceptance threshold indicates the acceptance distance to the ground truth as a rate of interocular distance. 7

8 Table 3: Emotion recognition accuracies for the extreme frames on the Cohn-Kanade database. Lm.Support Happy Surprised Angry Disgusted Frightened Sad Avg.(Emotions) Avg.(Samples) None Auto(BOS) Auto(FRGC) Ground Truth References [1] S. Arca, P. Campadelli, and R. Lanzarotti. A face recognition system based on automatically determined facial fiducial points. Pattern recognition, 39(3): , 06. [2] M. Bartlett, G. Littlewort, M. Frank, C. Lainscsek, I. Fasel, and J. Movellan. Recognizing facial expression: machine learning and application to spontaneous behavior. In CVPR, volume 2, page 568, 05. [3] F. Bookstein. Principal warps: thin-plate splines and the decomposition of deformations. TPAMI, 11: , [4] L. Chen, L. Zhang, H. Zhang, and M. Abdel-Mottaleb. 3D shape constraint for facial feature localization using probabilistic-like output. In AFGR, 04. [5] J. Cohn, A. Zlochower, J. Lien, and T. Kanade. Automated face analysis by feature point tracking has high concurrent validity with manual FACS coding. Psychophysiology, 36(01):35 43, [6] T. Cootes, G. Edwards, and C. Taylor. Active appearance models. TPAMI, 23(6): , 01. [7] D. Cristinacce and T. Cootes. Automatic feature localisation with constrained local models. Pattern Recognition, 41:54 67, 08. [8] M. Figueiredo and A. Jain. Unsupervised Learning of Finite Mixture Models. TPAMI, pages , 02. [9] Z. Ghahramani and G. Hinton. The EM algorithm for mixtures of factor analyzers. Technical Report CRG- TR-96-1, University of Toronto, (revised), [] C. Goodall. Procrustes methods in the statistical analysis of shape. Journal of the Royal Statistical Society B, 53(2): , [11] S. Ioannou, M. Wallace, K. Karpouzis, A. Raouzaiou, and S. Kollias. Combination of multiple extraction algorithms in the detection of facial features. In ICIP, volume 2, pages , 05. [12] T. Kanade, J. Cohn, and Y. Tian. Comprehensive database for facial expression analysis. In AFGR, 00. [13] T. Kozakaya, T. Shibata, M. Yuasa, and O. Yamaguchi. Facial feature localization using weighted vector concentration approach. Image and Vision Computing, 09. [14] S. Milborrow and F. Nicolls. Locating facial features with an extended active shape model. In ECCV, pages 4 513, 08. [15] M. Pantic and L. Rothkrantz. Automatic analysis of facial expressions: The state of the art. TPAMI, 22(12): , 00. [16] P. Phillips, P. Flynn, W. Scruggs, K. Bowyer, J. Chang, K. Hoffman, J. Marques, J. Min, and W. Worek. Overview of the face recognition grand challenge. In CVPR, volume 1, pages , 05. [17] A. A. Salah and E. Alpaydın. Incremental Mixtures of Factor Analysers. In ICPR, volume 1, pages , 04. [18] A. A. Salah, H. Cinar, L. Akarun, and B. Sankur. Robust Facial Landmarking For Registration. Annals of Telecommunications, 62(1-2): , 07. [19] A. Savran, N. Alyüz, H. Dibeklioğlu, O. Çeliktutan, B. Gökberk, B. Sankur, and L. Akarun. Bosphorus database for 3D face analysis. In BIOID, 08. [] R. Senaratne and S. Halgamuge. Optimised landmark model matching for face recognition. In AFGR, page 6, 06. [21] H. Tao and T. Huang. Connected vibrations: a modal analysis approach for non-rigid motion tracking. In CVPR, pages 735 7, [22] R. Valenti, N. Sebe, and T. Gevers. Facial expression recognition: A fully integrated approach. In ICIAPW, pages 125 1, 07. [23] P. Viola and M. Jones. Rapid object detection using a boosted cascade of simple features. In CVPR, volume 1, pages , 01. [24] D. Vukadinovic and M. Pantic. Fully automatic facial feature point detection using Gabor feature based boosted classifiers. In SMC, pages , 05. [25] L. Wiskott, J. Fellous, N. Krüger, and C. Von der Malsburg. Face recognition by elastic bunch graph matching. TPAMI, 19(7): , [26] H. X. L. S. Z. H. C. Q. Yan, S. Face alignment using view-based direct appearance models. Intl J. Imaging Systems and Technology, 13:6 112, 03. [27] Z. Zeng, M. Pantic, G. Roisman, and T. Huang. A survey of affect recognition methods: Audio, visual, and spontaneous expressions. TPAMI, pages 39 58, 09. [28] X. Zhu, J. Fan, and A. Elmagarmid. Towards facial feature extraction and verification for omni-face detection in video images. Image Processing, 2: , 02. 8

UvA-DARE (Digital Academic Repository) Enabling dynamics in face analysis Dibeklioglu, H. Link to publication

UvA-DARE (Digital Academic Repository) Enabling dynamics in face analysis Dibeklioglu, H. Link to publication UvA-DARE (Digital Academic Repository) Enabling dynamics in face analysis Dibeklioglu, H. Link to publication Citation for published version (APA): Dibekliolu, H. (24). Enabling dynamics in face analysis

More information

3D Facial Feature Localization for Registration

3D Facial Feature Localization for Registration 3D Facial Feature Localization for Registration Albert Ali Salah and Lale Akarun Boğaziçi University Perceptual Intelligence Laboratory Computer Engineering Department, Turkey {salah, akarun}@boun.edu.tr

More information

Generic Face Alignment Using an Improved Active Shape Model

Generic Face Alignment Using an Improved Active Shape Model Generic Face Alignment Using an Improved Active Shape Model Liting Wang, Xiaoqing Ding, Chi Fang Electronic Engineering Department, Tsinghua University, Beijing, China {wanglt, dxq, fangchi} @ocrserv.ee.tsinghua.edu.cn

More information

Recognition of facial expressions in presence of partial occlusion

Recognition of facial expressions in presence of partial occlusion Recognition of facial expressions in presence of partial occlusion Ioan Buciu, 1 Irene Kotsia 1 and Ioannis Pitas 1 AIIA Laboratory Computer Vision and Image Processing Group Department of Informatics

More information

Multi-Attribute Robust Facial Feature Localization

Multi-Attribute Robust Facial Feature Localization Multi-Attribute Robust Facial Feature Localization Oya Çeliktutan, Hatice Çınar Akakın, Bülent Sankur Boǧaziçi University Electrical & Electronic Engineering Department 34342 Bebek, Istanbul {oya.celiktutan,

More information

Intensity-Depth Face Alignment Using Cascade Shape Regression

Intensity-Depth Face Alignment Using Cascade Shape Regression Intensity-Depth Face Alignment Using Cascade Shape Regression Yang Cao 1 and Bao-Liang Lu 1,2 1 Center for Brain-like Computing and Machine Intelligence Department of Computer Science and Engineering Shanghai

More information

HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION

HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION Dipankar Das Department of Information and Communication Engineering, University of Rajshahi, Rajshahi-6205, Bangladesh ABSTRACT Real-time

More information

Classification of Face Images for Gender, Age, Facial Expression, and Identity 1

Classification of Face Images for Gender, Age, Facial Expression, and Identity 1 Proc. Int. Conf. on Artificial Neural Networks (ICANN 05), Warsaw, LNCS 3696, vol. I, pp. 569-574, Springer Verlag 2005 Classification of Face Images for Gender, Age, Facial Expression, and Identity 1

More information

Real time facial expression recognition from image sequences using Support Vector Machines

Real time facial expression recognition from image sequences using Support Vector Machines Real time facial expression recognition from image sequences using Support Vector Machines I. Kotsia a and I. Pitas a a Aristotle University of Thessaloniki, Department of Informatics, Box 451, 54124 Thessaloniki,

More information

3D Facial Landmarking under Expression, Pose, and Occlusion Variations

3D Facial Landmarking under Expression, Pose, and Occlusion Variations 3D Facial Landmarking under Expression, Pose, and Occlusion Variations Hamdi Dibeklioğlu, Albert Ali Salah and Lale Akarun Abstract Automatic localization of 3D facial features is important for face recognition,

More information

Multiple Kernel Learning for Emotion Recognition in the Wild

Multiple Kernel Learning for Emotion Recognition in the Wild Multiple Kernel Learning for Emotion Recognition in the Wild Karan Sikka, Karmen Dykstra, Suchitra Sathyanarayana, Gwen Littlewort and Marian S. Bartlett Machine Perception Laboratory UCSD EmotiW Challenge,

More information

Learning the Deep Features for Eye Detection in Uncontrolled Conditions

Learning the Deep Features for Eye Detection in Uncontrolled Conditions 2014 22nd International Conference on Pattern Recognition Learning the Deep Features for Eye Detection in Uncontrolled Conditions Yue Wu Dept. of ECSE, Rensselaer Polytechnic Institute Troy, NY, USA 12180

More information

Evaluation of Expression Recognition Techniques

Evaluation of Expression Recognition Techniques Evaluation of Expression Recognition Techniques Ira Cohen 1, Nicu Sebe 2,3, Yafei Sun 3, Michael S. Lew 3, Thomas S. Huang 1 1 Beckman Institute, University of Illinois at Urbana-Champaign, USA 2 Faculty

More information

Face Alignment Under Various Poses and Expressions

Face Alignment Under Various Poses and Expressions Face Alignment Under Various Poses and Expressions Shengjun Xin and Haizhou Ai Computer Science and Technology Department, Tsinghua University, Beijing 100084, China ahz@mail.tsinghua.edu.cn Abstract.

More information

Classification of Upper and Lower Face Action Units and Facial Expressions using Hybrid Tracking System and Probabilistic Neural Networks

Classification of Upper and Lower Face Action Units and Facial Expressions using Hybrid Tracking System and Probabilistic Neural Networks Classification of Upper and Lower Face Action Units and Facial Expressions using Hybrid Tracking System and Probabilistic Neural Networks HADI SEYEDARABI*, WON-SOOK LEE**, ALI AGHAGOLZADEH* AND SOHRAB

More information

Multi-Attribute Robust Facial Feature Localization

Multi-Attribute Robust Facial Feature Localization Multi-Attribute Robust Facial Feature Localization Oya Çeliktutan, Hatice Çınar Akakın, Bülent Sankur Boǧaziçi University Electrical & Electronic Engineering Department 34342 Bebek, Istanbul {oya.celiktutan,

More information

Boosting Coded Dynamic Features for Facial Action Units and Facial Expression Recognition

Boosting Coded Dynamic Features for Facial Action Units and Facial Expression Recognition Boosting Coded Dynamic Features for Facial Action Units and Facial Expression Recognition Peng Yang Qingshan Liu,2 Dimitris N. Metaxas Computer Science Department, Rutgers University Frelinghuysen Road,

More information

Cross-pose Facial Expression Recognition

Cross-pose Facial Expression Recognition Cross-pose Facial Expression Recognition Abstract In real world facial expression recognition (FER) applications, it is not practical for a user to enroll his/her facial expressions under different pose

More information

Component-based Registration with Curvature Descriptors for Expression Insensitive 3D Face Recognition

Component-based Registration with Curvature Descriptors for Expression Insensitive 3D Face Recognition Component-based Registration with Curvature Descriptors for Expression Insensitive 3D Face Recognition Neşe Alyüz Boğaziçi University Computer Engineering Dept. Istanbul, Turkey nese.alyuz@boun.edu.tr

More information

Selection of Location, Frequency and Orientation Parameters of 2D Gabor Wavelets for Face Recognition

Selection of Location, Frequency and Orientation Parameters of 2D Gabor Wavelets for Face Recognition Selection of Location, Frequency and Orientation Parameters of 2D Gabor Wavelets for Face Recognition Berk Gökberk, M.O. İrfanoğlu, Lale Akarun, and Ethem Alpaydın Boğaziçi University, Department of Computer

More information

Eye Detection by Haar wavelets and cascaded Support Vector Machine

Eye Detection by Haar wavelets and cascaded Support Vector Machine Eye Detection by Haar wavelets and cascaded Support Vector Machine Vishal Agrawal B.Tech 4th Year Guide: Simant Dubey / Amitabha Mukherjee Dept of Computer Science and Engineering IIT Kanpur - 208 016

More information

Spatiotemporal Features for Effective Facial Expression Recognition

Spatiotemporal Features for Effective Facial Expression Recognition Spatiotemporal Features for Effective Facial Expression Recognition Hatice Çınar Akakın and Bülent Sankur Bogazici University, Electrical & Electronics Engineering Department, Bebek, Istanbul {hatice.cinar,bulent.sankur}@boun.edu.tr

More information

Real-time facial feature point extraction

Real-time facial feature point extraction University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2007 Real-time facial feature point extraction Ce Zhan University of Wollongong,

More information

Texture Features in Facial Image Analysis

Texture Features in Facial Image Analysis Texture Features in Facial Image Analysis Matti Pietikäinen and Abdenour Hadid Machine Vision Group Infotech Oulu and Department of Electrical and Information Engineering P.O. Box 4500, FI-90014 University

More information

Facial expression recognition using shape and texture information

Facial expression recognition using shape and texture information 1 Facial expression recognition using shape and texture information I. Kotsia 1 and I. Pitas 1 Aristotle University of Thessaloniki pitas@aiia.csd.auth.gr Department of Informatics Box 451 54124 Thessaloniki,

More information

Occluded Facial Expression Tracking

Occluded Facial Expression Tracking Occluded Facial Expression Tracking Hugo Mercier 1, Julien Peyras 2, and Patrice Dalle 1 1 Institut de Recherche en Informatique de Toulouse 118, route de Narbonne, F-31062 Toulouse Cedex 9 2 Dipartimento

More information

FITTING AND TRACKING 3D/4D FACIAL DATA USING A TEMPORAL DEFORMABLE SHAPE MODEL. Shaun Canavan, Xing Zhang, and Lijun Yin

FITTING AND TRACKING 3D/4D FACIAL DATA USING A TEMPORAL DEFORMABLE SHAPE MODEL. Shaun Canavan, Xing Zhang, and Lijun Yin FITTING AND TRACKING 3D/4D FACIAL DATA USING A TEMPORAL DEFORMABLE SHAPE MODEL Shaun Canavan, Xing Zhang, and Lijun Yin Department of Computer Science State University of New York at Binghamton ABSTRACT

More information

Facial Expression Recognition Using Non-negative Matrix Factorization

Facial Expression Recognition Using Non-negative Matrix Factorization Facial Expression Recognition Using Non-negative Matrix Factorization Symeon Nikitidis, Anastasios Tefas and Ioannis Pitas Artificial Intelligence & Information Analysis Lab Department of Informatics Aristotle,

More information

Facial Expression Recognition using Principal Component Analysis with Singular Value Decomposition

Facial Expression Recognition using Principal Component Analysis with Singular Value Decomposition ISSN: 2321-7782 (Online) Volume 1, Issue 6, November 2013 International Journal of Advance Research in Computer Science and Management Studies Research Paper Available online at: www.ijarcsms.com Facial

More information

Facial Feature Detection

Facial Feature Detection Facial Feature Detection Rainer Stiefelhagen 21.12.2009 Interactive Systems Laboratories, Universität Karlsruhe (TH) Overview Resear rch Group, Universität Karlsruhe (TH H) Introduction Review of already

More information

Facial Expression Detection Using Implemented (PCA) Algorithm

Facial Expression Detection Using Implemented (PCA) Algorithm Facial Expression Detection Using Implemented (PCA) Algorithm Dileep Gautam (M.Tech Cse) Iftm University Moradabad Up India Abstract: Facial expression plays very important role in the communication with

More information

Exploring Bag of Words Architectures in the Facial Expression Domain

Exploring Bag of Words Architectures in the Facial Expression Domain Exploring Bag of Words Architectures in the Facial Expression Domain Karan Sikka, Tingfan Wu, Josh Susskind, and Marian Bartlett Machine Perception Laboratory, University of California San Diego {ksikka,ting,josh,marni}@mplab.ucsd.edu

More information

Robust Facial Expression Classification Using Shape and Appearance Features

Robust Facial Expression Classification Using Shape and Appearance Features Robust Facial Expression Classification Using Shape and Appearance Features S L Happy and Aurobinda Routray Department of Electrical Engineering, Indian Institute of Technology Kharagpur, India Abstract

More information

A Real Time Facial Expression Classification System Using Local Binary Patterns

A Real Time Facial Expression Classification System Using Local Binary Patterns A Real Time Facial Expression Classification System Using Local Binary Patterns S L Happy, Anjith George, and Aurobinda Routray Department of Electrical Engineering, IIT Kharagpur, India Abstract Facial

More information

Facial Expression Analysis

Facial Expression Analysis Facial Expression Analysis Jeff Cohn Fernando De la Torre Human Sensing Laboratory Tutorial Looking @ People June 2012 Facial Expression Analysis F. De la Torre/J. Cohn Looking @ People (CVPR-12) 1 Outline

More information

Real-time Automatic Facial Expression Recognition in Video Sequence

Real-time Automatic Facial Expression Recognition in Video Sequence www.ijcsi.org 59 Real-time Automatic Facial Expression Recognition in Video Sequence Nivedita Singh 1 and Chandra Mani Sharma 2 1 Institute of Technology & Science (ITS) Mohan Nagar, Ghaziabad-201007,

More information

Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information

Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information Mustafa Berkay Yilmaz, Hakan Erdogan, Mustafa Unel Sabanci University, Faculty of Engineering and Natural

More information

Context-based additive logistic model for facial keypoint localization

Context-based additive logistic model for facial keypoint localization AUTHORS: CONTEXT-BASED FACIAL KEYPOINT DETECTION 1 Context-based additive logistic model for facial keypoint localization Peng Li p.li@cs.ucl.ac.uk Jonathan Warrell j.warrell@brookes.ac.uk Jania Aghajanian

More information

Enhance ASMs Based on AdaBoost-Based Salient Landmarks Localization and Confidence-Constraint Shape Modeling

Enhance ASMs Based on AdaBoost-Based Salient Landmarks Localization and Confidence-Constraint Shape Modeling Enhance ASMs Based on AdaBoost-Based Salient Landmarks Localization and Confidence-Constraint Shape Modeling Zhiheng Niu 1, Shiguang Shan 2, Xilin Chen 1,2, Bingpeng Ma 2,3, and Wen Gao 1,2,3 1 Department

More information

LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM

LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM Hazim Kemal Ekenel, Rainer Stiefelhagen Interactive Systems Labs, University of Karlsruhe Am Fasanengarten 5, 76131, Karlsruhe, Germany

More information

A Novel Approach for Face Pattern Identification and Illumination

A Novel Approach for Face Pattern Identification and Illumination A Novel Approach for Face Pattern Identification and Illumination Viniya.P 1,Peeroli.H 2 PG scholar, Applied Electronics, Mohamed sathak Engineering college,kilakarai,tamilnadu, India 1 HOD, Department

More information

TRACKING OF FACIAL FEATURE POINTS BY COMBINING SINGULAR TRACKING RESULTS WITH A 3D ACTIVE SHAPE MODEL

TRACKING OF FACIAL FEATURE POINTS BY COMBINING SINGULAR TRACKING RESULTS WITH A 3D ACTIVE SHAPE MODEL TRACKING OF FACIAL FEATURE POINTS BY COMBINING SINGULAR TRACKING RESULTS WITH A 3D ACTIVE SHAPE MODEL Moritz Kaiser, Dejan Arsić, Shamik Sural and Gerhard Rigoll Technische Universität München, Arcisstr.

More information

COMPOUND LOCAL BINARY PATTERN (CLBP) FOR PERSON-INDEPENDENT FACIAL EXPRESSION RECOGNITION

COMPOUND LOCAL BINARY PATTERN (CLBP) FOR PERSON-INDEPENDENT FACIAL EXPRESSION RECOGNITION COMPOUND LOCAL BINARY PATTERN (CLBP) FOR PERSON-INDEPENDENT FACIAL EXPRESSION RECOGNITION Priyanka Rani 1, Dr. Deepak Garg 2 1,2 Department of Electronics and Communication, ABES Engineering College, Ghaziabad

More information

Video Google faces. Josef Sivic, Mark Everingham, Andrew Zisserman. Visual Geometry Group University of Oxford

Video Google faces. Josef Sivic, Mark Everingham, Andrew Zisserman. Visual Geometry Group University of Oxford Video Google faces Josef Sivic, Mark Everingham, Andrew Zisserman Visual Geometry Group University of Oxford The objective Retrieve all shots in a video, e.g. a feature length film, containing a particular

More information

Facial Action Coding Using Multiple Visual Cues and a Hierarchy of Particle Filters

Facial Action Coding Using Multiple Visual Cues and a Hierarchy of Particle Filters Facial Action Coding Using Multiple Visual Cues and a Hierarchy of Particle Filters Joel C. McCall and Mohan M. Trivedi Computer Vision and Robotics Research Laboratory University of California, San Diego

More information

Detection of Facial Landmarks of North Eastern Indian People

Detection of Facial Landmarks of North Eastern Indian People International Journal of Information and Computation Technology. ISSN 0974-2239 Volume 3, Number 9 (2013), pp. 953-962 International Research Publications House http://www. irphouse.com /ijict.htm Detection

More information

Recognizing Partial Facial Action Units Based on 3D Dynamic Range Data for Facial Expression Recognition

Recognizing Partial Facial Action Units Based on 3D Dynamic Range Data for Facial Expression Recognition Recognizing Partial Facial Action Units Based on 3D Dynamic Range Data for Facial Expression Recognition Yi Sun, Michael Reale, and Lijun Yin Department of Computer Science, State University of New York

More information

Multi-view Facial Expression Recognition Analysis with Generic Sparse Coding Feature

Multi-view Facial Expression Recognition Analysis with Generic Sparse Coding Feature 0/19.. Multi-view Facial Expression Recognition Analysis with Generic Sparse Coding Feature Usman Tariq, Jianchao Yang, Thomas S. Huang Department of Electrical and Computer Engineering Beckman Institute

More information

Fully Automatic Facial Action Recognition in Spontaneous Behavior

Fully Automatic Facial Action Recognition in Spontaneous Behavior Fully Automatic Facial Action Recognition in Spontaneous Behavior Marian Stewart Bartlett 1, Gwen Littlewort 1, Mark Frank 2, Claudia Lainscsek 1, Ian Fasel 1, Javier Movellan 1 1 Institute for Neural

More information

Atlas Construction and Sparse Representation for the Recognition of Facial Expression

Atlas Construction and Sparse Representation for the Recognition of Facial Expression This work by IJARBEST is licensed under a Creative Commons Attribution 4.0 International License. Available at: https://www.ijarbest.com/ Atlas Construction and Sparse Representation for the Recognition

More information

Artifacts and Textured Region Detection

Artifacts and Textured Region Detection Artifacts and Textured Region Detection 1 Vishal Bangard ECE 738 - Spring 2003 I. INTRODUCTION A lot of transformations, when applied to images, lead to the development of various artifacts in them. In

More information

Combining Gabor Features: Summing vs.voting in Human Face Recognition *

Combining Gabor Features: Summing vs.voting in Human Face Recognition * Combining Gabor Features: Summing vs.voting in Human Face Recognition * Xiaoyan Mu and Mohamad H. Hassoun Department of Electrical and Computer Engineering Wayne State University Detroit, MI 4822 muxiaoyan@wayne.edu

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Feature Detection and Tracking with Constrained Local Models

Feature Detection and Tracking with Constrained Local Models Feature Detection and Tracking with Constrained Local Models David Cristinacce and Tim Cootes Dept. Imaging Science and Biomedical Engineering University of Manchester, Manchester, M3 9PT, U.K. david.cristinacce@manchester.ac.uk

More information

Appearance Manifold of Facial Expression

Appearance Manifold of Facial Expression Appearance Manifold of Facial Expression Caifeng Shan, Shaogang Gong and Peter W. McOwan Department of Computer Science Queen Mary, University of London, London E1 4NS, UK {cfshan, sgg, pmco}@dcs.qmul.ac.uk

More information

A Hierarchical Probabilistic Model for Facial Feature Detection

A Hierarchical Probabilistic Model for Facial Feature Detection A Hierarchical Probabilistic Model for Facial Feature Detection Yue Wu Ziheng Wang Qiang Ji ECSE Department, Rensselaer Polytechnic Institute {wuy9,wangz1,jiq}@rpi.edu Abstract Facial feature detection

More information

3D Facial Landmark Localisation by Matching Simple Descriptors

3D Facial Landmark Localisation by Matching Simple Descriptors 3D Facial Landmark Localisation by Matching Simple Descriptors Marcelo Romero-Huertas and Nick Pears Abstract We present our graph matching approach for 3D facial feature localisation. The work here uses

More information

Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling

Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling [DOI: 10.2197/ipsjtcva.7.99] Express Paper Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling Takayoshi Yamashita 1,a) Takaya Nakamura 1 Hiroshi Fukui 1,b) Yuji

More information

22 October, 2012 MVA ENS Cachan. Lecture 5: Introduction to generative models Iasonas Kokkinos

22 October, 2012 MVA ENS Cachan. Lecture 5: Introduction to generative models Iasonas Kokkinos Machine Learning for Computer Vision 1 22 October, 2012 MVA ENS Cachan Lecture 5: Introduction to generative models Iasonas Kokkinos Iasonas.kokkinos@ecp.fr Center for Visual Computing Ecole Centrale Paris

More information

A coarse-to-fine curvature analysis-based rotation invariant 3D face landmarking

A coarse-to-fine curvature analysis-based rotation invariant 3D face landmarking A coarse-to-fine curvature analysis-based rotation invariant 3D face landmarking Przemyslaw Szeptycki, Mohsen Ardabilian and Liming Chen Abstract Automatic 2.5D face landmarking aims at locating facial

More information

Evaluation of Face Resolution for Expression Analysis

Evaluation of Face Resolution for Expression Analysis Evaluation of Face Resolution for Expression Analysis Ying-li Tian IBM T. J. Watson Research Center, PO Box 704, Yorktown Heights, NY 10598 Email: yltian@us.ibm.com Abstract Most automatic facial expression

More information

Model Based Analysis of Face Images for Facial Feature Extraction

Model Based Analysis of Face Images for Facial Feature Extraction Model Based Analysis of Face Images for Facial Feature Extraction Zahid Riaz, Christoph Mayer, Michael Beetz, and Bernd Radig Technische Universität München, Boltzmannstr. 3, 85748 Garching, Germany {riaz,mayerc,beetz,radig}@in.tum.de

More information

Postprint.

Postprint. http://www.diva-portal.org Postprint This is the accepted version of a paper presented at 14th International Conference of the Biometrics Special Interest Group, BIOSIG, Darmstadt, Germany, 9-11 September,

More information

Disguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network. Nathan Sun CIS601

Disguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network. Nathan Sun CIS601 Disguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network Nathan Sun CIS601 Introduction Face ID is complicated by alterations to an individual s appearance Beard,

More information

A Novel LDA and HMM-based technique for Emotion Recognition from Facial Expressions

A Novel LDA and HMM-based technique for Emotion Recognition from Facial Expressions A Novel LDA and HMM-based technique for Emotion Recognition from Facial Expressions Akhil Bansal, Santanu Chaudhary, Sumantra Dutta Roy Indian Institute of Technology, Delhi, India akhil.engg86@gmail.com,

More information

A face recognition system based on local feature analysis

A face recognition system based on local feature analysis A face recognition system based on local feature analysis Stefano Arca, Paola Campadelli, Raffaella Lanzarotti Dipartimento di Scienze dell Informazione Università degli Studi di Milano Via Comelico, 39/41

More information

Facial-component-based Bag of Words and PHOG Descriptor for Facial Expression Recognition

Facial-component-based Bag of Words and PHOG Descriptor for Facial Expression Recognition Proceedings of the 2009 IEEE International Conference on Systems, Man, and Cybernetics San Antonio, TX, USA - October 2009 Facial-component-based Bag of Words and PHOG Descriptor for Facial Expression

More information

Image Processing. Image Features

Image Processing. Image Features Image Processing Image Features Preliminaries 2 What are Image Features? Anything. What they are used for? Some statements about image fragments (patches) recognition Search for similar patches matching

More information

Robust Face Alignment Based on Hierarchical Classifier Network

Robust Face Alignment Based on Hierarchical Classifier Network Robust Face Alignment Based on Hierarchical Classifier Network Li Zhang 1, Haizhou Ai 1, and Shihong Lao 2 1 Department of Computer Science, Tsinghua University, Beiing 0084, China 2 Sensing and Control

More information

Adaptive Skin Color Classifier for Face Outline Models

Adaptive Skin Color Classifier for Face Outline Models Adaptive Skin Color Classifier for Face Outline Models M. Wimmer, B. Radig, M. Beetz Informatik IX, Technische Universität München, Germany Boltzmannstr. 3, 87548 Garching, Germany [wimmerm, radig, beetz]@informatik.tu-muenchen.de

More information

Facial Processing Projects at the Intelligent Systems Lab

Facial Processing Projects at the Intelligent Systems Lab Facial Processing Projects at the Intelligent Systems Lab Qiang Ji Intelligent Systems Laboratory (ISL) Department of Electrical, Computer, and System Eng. Rensselaer Polytechnic Institute jiq@rpi.edu

More information

Facial Expression Recognition Using Gabor Motion Energy Filters

Facial Expression Recognition Using Gabor Motion Energy Filters Facial Expression Recognition Using Gabor Motion Energy Filters Tingfan Wu Marian S. Bartlett Javier R. Movellan Dept. Computer Science Engineering Institute for Neural Computation UC San Diego UC San

More information

AUTOMATIC VIDEO INDEXING

AUTOMATIC VIDEO INDEXING AUTOMATIC VIDEO INDEXING Itxaso Bustos Maite Frutos TABLE OF CONTENTS Introduction Methods Key-frame extraction Automatic visual indexing Shot boundary detection Video OCR Index in motion Image processing

More information

Image Processing Pipeline for Facial Expression Recognition under Variable Lighting

Image Processing Pipeline for Facial Expression Recognition under Variable Lighting Image Processing Pipeline for Facial Expression Recognition under Variable Lighting Ralph Ma, Amr Mohamed ralphma@stanford.edu, amr1@stanford.edu Abstract Much research has been done in the field of automated

More information

Face Analysis using Curve Edge Maps

Face Analysis using Curve Edge Maps Face Analysis using Curve Edge Maps Francis Deboeverie 1, Peter Veelaert 2 and Wilfried Philips 1 1 Ghent University - Image Processing and Interpretation/IBBT, St-Pietersnieuwstraat 41, B9000 Ghent, Belgium

More information

Robust Face Detection Based on Convolutional Neural Networks

Robust Face Detection Based on Convolutional Neural Networks Robust Face Detection Based on Convolutional Neural Networks M. Delakis and C. Garcia Department of Computer Science, University of Crete P.O. Box 2208, 71409 Heraklion, Greece {delakis, cgarcia}@csd.uoc.gr

More information

Automatic facial emotion recognition

Automatic facial emotion recognition Automatic facial emotion recognition Aitor Azcarate, Felix Hageloh, Koen van de Sande, Roberto Valenti Universiteit van Amsterdam June 2005 Abstract Recognizing human facial expression and emotion by computer

More information

Detection of asymmetric eye action units in spontaneous videos

Detection of asymmetric eye action units in spontaneous videos Detection of asymmetric eye action units in spontaneous videos The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published

More information

Learning to Recognize Faces in Realistic Conditions

Learning to Recognize Faces in Realistic Conditions 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Subject-Oriented Image Classification based on Face Detection and Recognition

Subject-Oriented Image Classification based on Face Detection and Recognition 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Computer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki

Computer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki Computer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki 2011 The MathWorks, Inc. 1 Today s Topics Introduction Computer Vision Feature-based registration Automatic image registration Object recognition/rotation

More information

Facial Emotion Recognition using Eye

Facial Emotion Recognition using Eye Facial Emotion Recognition using Eye Vishnu Priya R 1 and Muralidhar A 2 1 School of Computing Science and Engineering, VIT Chennai Campus, Tamil Nadu, India. Orcid: 0000-0002-2016-0066 2 School of Computing

More information

Lecture 16: Object recognition: Part-based generative models

Lecture 16: Object recognition: Part-based generative models Lecture 16: Object recognition: Part-based generative models Professor Stanford Vision Lab 1 What we will learn today? Introduction Constellation model Weakly supervised training One-shot learning (Problem

More information

Convolutional Neural Networks for Facial Expression Recognition

Convolutional Neural Networks for Facial Expression Recognition Convolutional Neural Networks for Facial Expression Recognition Shima Alizadeh Stanford University shima86@stanford.edu Azar Fazel Stanford University azarf@stanford.edu Abstract In this project, we have

More information

Person-Independent Facial Expression Recognition Based on Compound Local Binary Pattern (CLBP)

Person-Independent Facial Expression Recognition Based on Compound Local Binary Pattern (CLBP) The International Arab Journal of Information Technology, Vol. 11, No. 2, March 2014 195 Person-Independent Facial Expression Recognition Based on Compound Local Binary Pattern (CLBP) Faisal Ahmed 1, Hossain

More information

The Template Update Problem

The Template Update Problem The Template Update Problem Iain Matthews, Takahiro Ishikawa, and Simon Baker The Robotics Institute Carnegie Mellon University Abstract Template tracking dates back to the 1981 Lucas-Kanade algorithm.

More information

Global Gabor features for rotation invariant object classification

Global Gabor features for rotation invariant object classification Global features for rotation invariant object classification Ioan Buciu Electronics Department Faculty of Electrical Engineering and Information Technology University of Oradea 410087, Universitatii 1

More information

Sparse Shape Registration for Occluded Facial Feature Localization

Sparse Shape Registration for Occluded Facial Feature Localization Shape Registration for Occluded Facial Feature Localization Fei Yang, Junzhou Huang and Dimitris Metaxas Abstract This paper proposes a sparsity driven shape registration method for occluded facial feature

More information

Facial expression recognition for multi-player on-line games

Facial expression recognition for multi-player on-line games University of Wollongong Theses Collection University of Wollongong Theses Collection University of Wollongong Year 2008 Facial expression recognition for multi-player on-line games Ce Zhan University

More information

Face detection and recognition. Many slides adapted from K. Grauman and D. Lowe

Face detection and recognition. Many slides adapted from K. Grauman and D. Lowe Face detection and recognition Many slides adapted from K. Grauman and D. Lowe Face detection and recognition Detection Recognition Sally History Early face recognition systems: based on features and distances

More information

Automatic Detecting Neutral Face for Face Authentication and Facial Expression Analysis

Automatic Detecting Neutral Face for Face Authentication and Facial Expression Analysis From: AAAI Technical Report SS-03-08. Compilation copyright 2003, AAAI (www.aaai.org). All rights reserved. Automatic Detecting Neutral Face for Face Authentication and Facial Expression Analysis Ying-li

More information

Classifying Facial Gestures in Presence of Head Motion

Classifying Facial Gestures in Presence of Head Motion Classifying Facial Gestures in Presence of Head Motion Wei-Kai Liao and Isaac Cohen Institute for Robotics and Intelligent Systems Integrated Media Systems Center University of Southern California Los

More information

Landmark Localisation in 3D Face Data

Landmark Localisation in 3D Face Data 2009 Advanced Video and Signal Based Surveillance Landmark Localisation in 3D Face Data Marcelo Romero and Nick Pears Department of Computer Science The University of York York, UK {mromero, nep}@cs.york.ac.uk

More information

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit Augmented Reality VU Computer Vision 3D Registration (2) Prof. Vincent Lepetit Feature Point-Based 3D Tracking Feature Points for 3D Tracking Much less ambiguous than edges; Point-to-point reprojection

More information

Face Alignment Using Statistical Models and Wavelet Features

Face Alignment Using Statistical Models and Wavelet Features Face Alignment Using Statistical Models and Wavelet Features Feng Jiao 1*, Stan Li, Heung-Yeung Shum, Dale Schuurmans 1 1 Department of Computer Science, University of Waterloo Waterloo, Canada Microsoft

More information

Better than best: matching score based face registration

Better than best: matching score based face registration Better than best: based face registration Luuk Spreeuwers University of Twente Fac. EEMCS, Signals and Systems Group Hogekamp Building, 7522 NB Enschede The Netherlands l.j.spreeuwers@ewi.utwente.nl Bas

More information

Dynamic Facial Expression Recognition Using A Bayesian Temporal Manifold Model

Dynamic Facial Expression Recognition Using A Bayesian Temporal Manifold Model Dynamic Facial Expression Recognition Using A Bayesian Temporal Manifold Model Caifeng Shan, Shaogang Gong, and Peter W. McOwan Department of Computer Science Queen Mary University of London Mile End Road,

More information

Registration of three-dimensional face scans with average face models

Registration of three-dimensional face scans with average face models Journal of Electronic Imaging 17(1), 011006 (Jan Mar 2008) Registration of three-dimensional face scans with average face models Albert Ali Salah Boğaziçi University 34342 Bebek İstanbul, Turkey Centrum

More information

Large-scale Datasets: Faces with Partial Occlusions and Pose Variations in the Wild

Large-scale Datasets: Faces with Partial Occlusions and Pose Variations in the Wild Large-scale Datasets: Faces with Partial Occlusions and Pose Variations in the Wild Tarik Alafif, Zeyad Hailat, Melih Aslan and Xuewen Chen Computer Science Department, Wayne State University Detroit,

More information

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation Face Tracking Amit K. Roy-Chowdhury and Yilei Xu Department of Electrical Engineering, University of California, Riverside, CA 92521, USA {amitrc,yxu}@ee.ucr.edu Synonyms Facial Motion Estimation Definition

More information