Multiple Facial Action Unit Recognition Enhanced by Facial Expressions

Size: px
Start display at page:

Download "Multiple Facial Action Unit Recognition Enhanced by Facial Expressions"

Transcription

1 rd International Conference on Pattern Recognition (ICPR) Cancún Center, Cancún, México, December 4-8, 2016 Multiple Facial Action Unit Recognition Enhanced by Facial Expressions Jiajia Yang, Shan Wu, Shangfei Wang 1,Qiang Ji School of Computer Science and Technology University of Science and Technology of China Hefei, Anhui, China Department of Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute Troy, NY, USA Abstract Facial expressions and facial action units (AU) respectively describe facial behavior globally and locally. Therefore, the dependencies between expressions and AUs carry crucial information for facial action unit recognition, yet have not been thoroughly exploited. In this paper, we propose a novel facial action unit recognition method enhanced by facial expressions, which are only required during training. Specifically, we propose a three-layer restricted Boltzmann machine (RBM) to capture the probabilistic dependencies among expressions and AUs. The parameters of the RBM model are learned by maximizing the log conditional likelihood with gradient ascent. After that, the learned RBM model combines AU measurements with the AU-expression relations it captures to perform multiple AU recognition through probabilistic inference. Experimental results on three benchmark databases, i.e. the CK+ database, the ISL database and the BP4D database, demonstrate the effectiveness of our method on capturing the joint relations among AUs and expression to improve AU recognition. I. INTRODUCTION Automatical facial expression recognition has the great application potential in many fields, such as security [1], human-computer-interaction [2], driver safety [3], and healthcare [4]. Therefore, facial expression has attracted increasing attention in recent years. Facial expressions can be presented in two ways: expression categories and facial action units (AU). Expression categories describe facial behaviors in a global way, while facial action units depict the local variations on face. Almost all facial expressions consist of several certain combinations of AUs. For example, as shown in Table I, which lists the co-occurrence probability among four AUs and two expressions in the CK+ database, we find that both happiness and surprise occur along with at least two AUs. Happiness co-occur with AU6, AU12 and AU25, and surprise appears with AU1 and AU25. Different expressions may generate the same AUs. For example, AU25 could appear on both a happy face and a surprise face. AUs rarely appear alone. Generally they are always cooccurrence or mutual exclusion. For instance, from Table I, we find that AU1 and AU6 rarely occur simultaneously due to 1 This is the corresponding author. the structure of facial muscle, while AU6 and AU12, AU1 and AU25 always appear together. Therefore, iherent relationships among AUs as well as relations between expressions and AUs are important information for AU recognition. TABLE I PROBABILISTIC DEPENDENCIES AMONG AUS AND EXPRESSIONS. AU1 AU6 AU12 AU25 happy surprise AU AU AU AU Currently, the mainstream of AU recognition research classifies each AU independently or detects certain AU combinations. They either ignore the inherent dependencies a- mong multiple AUs as well as dependencies between AUs and expressions, or can not handle hundreds of variance in AU combinations. Only recently, several works exploit AU relations for AU recognition. For example, Tong et al. [5] adopted a Bayesian Network (BN) to model the semantic relationships among AUs through the structure and the conditional probabilities of the learned BN. Wang et al. [6] proposed a hierarchical Restricted Boltzman Machine model to capture the global semantic relationships among AUs through the connections between the hidden units and visible units of RBM. Zhu et al. [7] adopted a multi-task feature learning method to learn the shared features for each AU group, and a Bayesian network to model relations among action units from the target labels. Zhang et al. [8] utilized multi-task multiple kernel learning to detect fixed groups of multiple AUs simultaneously. The fixed AU groups are determined by a hierarchical model based on AU co-occurrences existing in AU labels and facial regions. Zhao et al. [9] proposed to select a sparse subset of facial patches and learn multiple AU classifiers simultaneously under the constrains of group /16/$ IEEE 4078

2 sparsity and local AU relations (i.e positive correlation and negative competition). Eleftheriadis et al. [10] proposed a multi-conditional latent variable model to combine AU label dependencies into latent space and classifier learning. Song et al. [11] proposed a Bayesian compressed sensing model to model AU sparsity and co-occurrence for AU recognition. All these works successfully exploit the dependencies inherent in multiple AUs to facilitate AU recognition. However, few of them consider the AU-expression relations to improve AU recognition. To the best of our knowledge, there is so far only three works that recognizes AUs assisted by expressions. Wang et al. [12] designed an AU recognition method with the help of expression labels as hidden knowledge under incomplete AU labeling. They constructed a BN model to capture the dependencies not only among AUs but also between AUs and expressions. During training, the image features and expression labels are used, while the AU labels may be missing. Structure EM is adopted to learn the structure and parameters of the BN. The traditional image-driven method is adopted to obtain the expression and AU measurements. During testing, the AUs are inferred by combining the measurements and the AU relations in the BN model. Due to the Markov assumption, the BN model can only handle local dependencies among AUs and expressions. Wang et al. [6] proposed to use a 3-way RBM to capture the global dependencies between expressions and AUs for AU recognition. The expression labels are used as privileged information, which is only required during training. The 3-way RBM model can be regarded as a mixture model, therefore, it models the relations between each expression and AUs independently, ignoring the shared dependencies among multiple expressions and AUs. Ruiz et al. [13] proposed a hidden-task-learning method to exploit prior knowledge about the relation between a set of hidden tasks (i.e. AUs) and visible-tasks ( i.e. facial expressions ). The relations between AUs and facial expressions are modelled based on empirical results obtained in psychological studies. Although Ruiz et al. method can learn to recognize AUs from the training data where AU labels are limited or even not available, the learned AU-expression relations from empirical results of psychological studies are incorrect. In this paper, we propose a novel facial action unit recognition method enhanced by facial expressions, which are only required during training. Specifically, we propose a hierarchy RBM model to capture the global relationships among AUs and facial expressions. Instead of modeling local dependencies between AUs and expressions like BN, the proposed RBM model can capture global dependencies among AUs and expressions through introducing a hidden layer. Unlike 3- way RBM, which captures the global dependencies between each expression and AUs independently, the proposed RBM model captures the dependencies among multiple expressions and AUs simultaneously by dividing its visible layer into two parts, expression nodes and AU nodes. During training, the parameters of the RBM model are learned by maximizing the log conditional likelihood with gradient ascent. During test, we infer the target AUs according to the learned RBM model without expression. Experiments on three benchmark databases demonstrate the superiority of our method to existing methods, further proving the necessity of exploiting the inherent relations among AUs and expressions. II. METHOD The framework of our method is shown in Fig. 1, which consists of two modules: measurement extraction and model learning. During the training phase, we obtain the measurement of AUs and expressions by a traditional image-based classifier. Then we learn the RBM model to capture the global relationships among AUs and expressions with the evidence of measurement. During the testing, we infer the final AUs through the learned RBM model without expressions. Assuming the training data is D = {X i, (λ i1,, λ in, λ in+1 )} N i=1, where N is the number of training samples, n is the number of AU labels. For the ith training sample, we denote X i as the feature vector of the ith training sample, and λ i1,, λ in as the n AU labels, and λ in+1 as the expression label respectively. The goal of this work is to construct an AU classifier with the help of expression knowledge that is only available during training. The details of our method is as follows. A. Measurement Extraction Fig. 1. The framework of our method The measurements mλ in Fig. 2 are preliminary estimation of the AU and expression labels using existing image-driven recognition method based on training data. In our work, the Support Vector Machine (SVM) is used as the classifier to obtain the measurements of AUs and expressions. B. Modeling AUs and Expressions Relationships by Three- Layer Restricted Boltzmann Machine In order to model the semantic relationships among AUs and expressions, a three-layer RBM model is used in this work. We first present our RBM model for AU recognition, then we 4079

3 provide the learning and inference algorithms of the threelayer RBM model. Fig. 2 shows the structure of the RBM model used to capture the global relationships among AUs and expressions. The middle layer is the binary visible units λ 1, λ 2,, λ n1 and E 1, E 2,, E n2, representing the state of n AUs and expressions respectively. In our model, we encode the expression labels with 0 and 1 in binary coding system. For k expression labels, we use {0, 1} n to represent each label, in which 2 n = k. For example, if there are 7 kinds of expression labels, we use 001 to represent the first kind expression label, and 111 for the 7th kind expression. The variables m 1, m 2,, m d in the bottom layer stand for the measurements of AUs and expressions obtained by a traditional classifier. In our model, each latent unit is connected to all AU and expression nodes in order to model the higher-order relationships among AUs and expressions. And the bottom layer s nodes are attached to each AU and expression variable, providing low-level evidence. Fig. 2. The model for AU recognition.in training phase we add expression nodes,in predicting phase without expression nodes The total energy of the model is defined in Equation 1, where b i and b j are the biases for the AUs and expression nodes respectively, and c k is the biases for latent nodes. The first set of parameters Wik 1 and W jk 1 measure the compatibility between each pair of latent units and visible units (AU and expression nodes), and the second set of parameters Wi 2 and Wj 2 measures the compatibility between each pair of measurement and AU and the compatibility between each pair of measurement and expression labels. E(m, λ, E, h; θ) = i k j λ i Wikh 1 k i k b i λ i c k h k Wi 2 λ i m i i E j Wjkh 1 k b j E j j k Wj 2 E j m j j Our model can be decomposed to two parts. The first part consisting of the hidden nodes and visible nodes is a RBM model, and each latent unit is connected to all the action units and expression units to model their dependencies. These two layers encode the high-level semantic relationship among (1) AUs and AU-expression, enabling us to infer the presence or absence of each AU in the top-down direction. The bottom two layers integrate the AU labels and AU measurements obtained by traditional classifier for AU recognition, providing evidence for AU recognition in a bottom-up direction. The obtained relationship between AUs and AU-expression can be implicitly inferred from the parameters Wik 1 and Wjk 1 (bias terms are omitted without loss of generality). To indicate how they are related, considering α th latent unit h α, its probabilistic dependency with each AU is measured by the pairwise energy E(h α, λ i ) = λ i Wiα 1 h α, i = 1,..., n1. And the probabilistic dependency with each expression unit is measured by the pairwise energy E(h α, ˆλ i ) = ˆλ i Wiα 1 h α, i = 1,..., n2. Apparently, the lager Wiα 1 is, the more likely AU λ i will be present. Conversely, the smaller W iα is, the more likely λ i will be absent. The entire parameters [Wiα 1 ]n1 1 obtain a specific presence and absence pattern of all the AUs and expressions. Take the parameter [Wi2 1 ]n1 1 for h 2 for example, Figure 3 and Figure 4 respectively depicts the pattern of h 2 with expressions and without expressions on the BP4D database, which has 6 AUs and 8 categories of expression. For each expression label, we encode with 4 binary numerical value. Comparing two figures, when we add the expression labels to the model, the relations among AUs change a little. In Figure 4 only two AUs: AU15 (a5) and AU23 (a6) tend to be present, AU15 (a5) is the most likely to be present. While in Figure 3 we add expression nodes into model, then AU4 (a3) become the most likely to be present, AU7 (a4) and AU15 (a5) tends to occur yet are less likely than AU4 (a3), and AU1, AU2 and AU23 (a1,a2,a6) are very likely to be absent. Since we added sadness expression nodes(a7 to a10:0010) to the model, AU4(Brow Lowerer) and AU7(Lid Tightener) became present, while AU23(Lip Tightener) became likely to be absent. Fig. 3. Dependencies among AUs and expressions In training phase, parameters are learned by maximizing the log conditional likelihood which is shown in Equation 2, where m (i) represents the measurements of the i-th training sample and λ (i) stands for the labels of AUs and expressions. It is maximized with the stochastic gradient ascent method and the gradient can be calculated with Equation 3. θ = arg max L(θ) θ N L(θ) = log p(λ (i), E (i) m (i) ; θ) i=1 (2) 4080

4 Fig. 4. Dependencies among AUs L(θ) N = ( E θ θ p(h,λ,e m (i),θ) E θ p(h λ (i),m (i),e (i),θ) ) i=1 (3) To obtain the gradient, we should calculate p(h λ, m, E, θ) and p(h, λ, E m, θ). p(h λ, m, E, θ) can be analytically calculated with Equation 4, where σ(x) = 1/(1+e x ) is a sigmoid function. We approximate p(h, λ, E m, θ) by sampling h with Equation 4, sampling λ with Equation 5, and sampling E with Equation 6. The detailed algorithm for learning the parameter W 1 is shown in Algorithm 1. Other parameters can be estimated in the similar way. P (h k = 1 λ, E, m) = σ( c k i P (λ i = 1 h, m) = σ( b i k P (E j = 1 h, m) = σ( b j k W 1 ikλ i j W 1 jke j ) (4) W 1 ikh k W 2 i m i ) (5) W 1 jkh k W 2 j m j ) (6) After parameters learning, we infer the target label based on the learned model. Given the measurement m of a testing sample, the final state of each AU ˆλ i can be obtained by maximizing its posterior probability given m with Equation 7. ˆλ i = arg max ˆλ i P (ˆλ i m) (7) Computing P (ˆλ i m) requires marginalizing over all the latent variables {h k } m k=1 and other action units and expression units λ ss =i which could be intractable. However it can be efficiently performed with the Gibbs sampling method by iteratively sampling h from P (h λ, m) and sampling λ from P (λ h, m). Sampled instances of each λ i are used to calculate the corresponding marginal probability. Algorithm 2 presents the detailed steps. A. Experimental Conditions III. EXPERIMENTS In our work, we use three widely used databases in the literature: the Extended Cohn-Kanade(CK+) [14], the BP4D Algorithm 1 Revised contrastive divergence algorithm for learning the proposed model [6] Input: Training data: {λ (i) R 1 n1, E (i) R 1 n2, m (i) R 1 d } N i=1 Output: Model parameters W 1, W 2, b, c 1: repeat 2: for each training sample (λ, E, m) do 3: D + 2 = λt m, D + 3 = ET m 4: Sample h + P (h E, λ, m) with Equation 4 5: Calculate the positive gradient D + 1 = λt h + 6: Sample λ P (λ h +, m) with Equation 5 7: D2 = λ T m 8: Sample E P (E h +, m) with Equation 6 9: D3 = E T m 10: Sample h P (h λ, E m) with Equation 4 11: Calculate the negative gradient D1 = λ T h 12: Update: 13: W 1 = W 1 + η(d 1 + D 1 ) 14: 15: W 2 = W 2 + η(d 2 + D 2 ) b = b + η(λ λ ) 16: c = c + η(λ λ ) 17: end for 18: until Convergence Algorithm 2 Inference of P (λ m) with Gibbs Sampling [6] Input: Test sample m; Parameters W 1, W 2, b, c Output: P (λ i m) for i = 1, 2,..., n 1: for chain = 1 C do 2: Randomly initialize a 0 3: 4: for t = 0 N do 5: Sample h t P (h λ, m) with Equation 4 6: Sample λ t+1 P (λ h t, m) with Equation 5 7: end for 8: end for 9: for i = 1 n do 10: Collect the last K samples of λ i from each chain 11: Calculate P (λ i m) based on the collected samples 12: end for database [15] and the RPI Intelligent Systems Lab (ISL) Image Database [16]. The CK+ database contains 593 samples collected from 118 subjects. There are 7 expression categories, i.e., anger, contempt, disgust, fear, happiness, sadness, surprise, and 13 AU labels for part of the samples. In our work, we select 327 samples with both expression category and AU labels. And only AUs whose frequencies are more than 10% are considered, resulting in 13 AUs (AU1, AU2, AU4, AU5, AU6, AU7, AU9, AU12, AU17, AU23, AU24, AU25, and AU27). The BP4D database consists of 6554 dynamic spontaneous facial expression data. We picked out 6 AUs (AU1, AU2,AU4, AU7, AU15, AU23) and 8 expression categories (happiness, sadness, surprise, embarrassment, fear, physical pain, anger, 4081

5 and disgust) in BP4D database[15] for our experiment. The ISL database[16] is collected under real-world conditions with uncontrolled illumination and background, as well as moderate head motion. The 19 frontal view video clips for 7 subjects displaying facial expressions are adopted in this work. It contains 10 AUs (AU1, AU2, AU5, AU6, AU12, AU17, AU23, AU25, AU27, and AU45) and two expressions (surprise and happiness). Leave one subject out cross-validation is employed in our experiment. Two types of metrics for multi-label classification are adopted, i.e. example-based assessment criteria: Hamming Loss, Example based accuracy, Example based Subset Accuracy, and label-based assessment criteria: Label based MacroF1. Detailed definitions of these evaluation measures can be found in [17]. B. Experimental Results and Analysis TABLE II EXPERIMENTAL RESULTS ON THREE DATABASES. Database method HL Acc SubAcc MacF1 CK+ BP4D ISL Image-driven Model-based our s Image-driven Model-based our s Image-driven Model-based our s HL= Hamming Loss, Acc=Example based accuracy, SubAcc=Example based Subset Accuracy, MacF1=Label based MacroF1 In this section, we compare our work with two methods: the image-driven method using SVM, and the model-based method using the proposed RBM with only AUs. The experimental results on three databases are shown in Table II. From this table, we find that: First, the model-based method with AUs only performs better than the image-driven method, since all four metrics of the model-based method on three databases are higher than the image-driven method in nearly all the cases. Specifically, on the ISL database, the improvements are most obvious, since both the accuracy and macro F1-score of image-driven method are lower than 50%, while the results of model-based method are higher than 80%. The image-driven method treats each AU as independently classes, while the model-based method using RBM captures the global semantic relations among AUs. These demonstrate that inherent dependencies among AUs are very useful to improve AU recognition. Second, our proposed method of modeling the relations among AUs and expressions achieves the best results among three methods with lowest Hamming Loss and highest accuracy and F1 score in most situations. On the CK+ database, the accuracy and the F1-score are increased by 12% and 5%; on the BP4D database, the improvement on the accuracy and the subset accuracy are about 9% and 12% separately; on the ISL database, the subset accuracy and the F1-score are raised from 75% to 79% and 95% to 96% respectively. Therefore our method enhances the performance of AU recognition by capturing relationship among AUs and expressions. Finally, the improvement of the model-based method and the proposed method is more obviously on the BP4D database and the ISL databases than that on the CK+ database. As mentioned in Section III-A, the CK+ database is a posed dataset, and other two databases are spontaneous dataset. It demonstrates that the effectiveness of our method on recognizing real-world data. C. Comparison with Related Work We compare our work with recent state-of-the-art approaches. On CK+ database, we compare our work with [12] and [6]. Both work employed expression as a privileged information to facilitate AU recognition. Wang et al. [12] proposed to use BN to capture the relations among AUs and expression. Wang et al. [6] proposed a three-way RBM model to incorporate expression during training to facilitate the estimation of the AU dependencies. On BP4D database, we compare with [18]. Gudi et al. [18] proposed a model-based method by using a deep CNN consisting of 3 convolutional layers, 1 sub-sampling layer and 1 fully connected layer to predict the occurrence and intensity of Facial AUs. The results are listed in Table III. To our knowledge, no work hasbeen done to capture AU relationship on ISL database, so we don t compare our work with any related works on this database. From Table III, we find that our method performs better than [12] and [6]. Compared with [12], our method lowers the Hamming loss by 0.03 and increases the macro F1-score by 13%. Due to the Markov assumption, the relationships captured by BN are local pairwise. In contrast, our method captures a global relations among multiple AUs and expressions through a hidden layer. When comparing with [6], the macro F1-score of our method is 93%, which is 11% higher than [6]. The relations in [6] are expression independent, yet our method captures the joint relations among AUs and expressions through the shared weights among multiple expressions. These demonstrate that our model by learning a joint relationships among AUs and expressions is stronger to model the complicated relations among multiple AUs than the state-of-the-art approaches. TABLE III COMPARISON WITH RELATED WORKS Database Method HL MacF1 CK+ BP4D [12] [6] ours [18] ours HL= Hamming Loss, MacF1=Label based MacroF1 Table IV lists some results of recent AU recognition work using AU relations only on the CK+ database. As mentioned in Section I, Song et al. [11] modeled AU sparsity and co-occurrence using a Bayesian compressed sensing model. 4082

6 Eleftheriadis et al. [10] proposed a multi-conditional latent variable model for simultaneous facial feature fusion and detection of facial action units. Zhu et al. [7] proposed to learn the shared features through a multi-task learning process and model AU relations by a Bayesian network. Zhang et al. [8] proposed to use a multi-task multiple kernel learning to detect relations among fixed groups of multiple AUs. Since the specific AUs and condition is not completely accordant, we make a rough comparison with them. From Table IV, we can see that compared with MC-LVM, MCF, H-MTMKL and BGCS, our method achieves a higher macro F1-score. However, the accuracy of our method is worse than BGCS and DBN. According to the definition of accuracy, it is easily influenced by the number of negative classes, while F1-score makes a balance between precision and recall, so it is more reliable. Furthermore, our method outperforms [7] on both macro F1-score and accuracy. Therefore, compared with the related work listed in Table IV, our method achieves a more reliable and balanced result, demonstrating the effectiveness of our method. TABLE IV REFERENCE TO RELATED WORKS Database Related Works Method MacF1 Acc CK+ MCF [11] BGCS DBN [10] MC-LVM [7] MTFL+BN [8] H-MTMKL ours our method IV. CONCLUSIONS Although the dependencies between expressions and AUs carry crucial information for facial action unit recognition, only a very few works exploited such dependencies. In this paper, we propose a novel AU recognition method assisted with expressions, which is only required during training. Specifically, we proposed a three-layer RBM to capture the global dependencies among AUs as well as among AUs and expressions by introducing a hidden layer and dividing its visible layer into two parts: expression nodes and AU nodes. Experimental results on three benchmark databases demonstrate the effectiveness of our method on capturing the joint relations among AUs and expressions and its superior performance on AU recognition. V. ACKNOWLEDGMENT This work has been supported by the National Science Foundation of China (Grant No , , ), and the project from Anhui Science and Technology Agency ( SMF223). REFERENCES [1] Andrew Ryan, Jeffery F Cohn, Simon Lucey, Jason Saragih, Patrick Lucey, Fernando De la Torre, and Adam Rossi. Automated facial expression recognition system. In Security Technology, rd Annual 2009 International Carnahan Conference on, pages IEEE, [2] Alessandro Vinciarelli, Maja Pantic, and Hervé Bourlard. Social signal processing: Survey of an emerging domain. Image and Vision Computing, 27(12): , [3] Esra Vural, Müjdat Çetin, Aytül Erçil, Gwen Littlewort, Marian Bartlett, and Javier Movellan. Automated drowsiness detection for improved driving safety [4] Patrick Lucey, Jeffrey Cohn, Simon Lucey, Iain Matthews, Sridha Sridharan, and Kenneth M Prkachin. Automatically detecting pain using facial actions. In Affective Computing and Intelligent Interaction and Workshops, ACII rd International Conference on, pages 1 8. IEEE, [5] Yan Tong and Qiang Ji. Learning bayesian networks with qualitative constraints. In Computer Vision and Pattern Recognition, CVPR IEEE Conference on, pages 1 8. IEEE, [6] Ziheng Wang, Yongqiang Li, Shangfei Wang, and Qiang Ji. Capturing global semantic relationships for facial action unit recognition. In Proceedings of the IEEE International Conference on Computer Vision, pages , [7] Yachen Zhu, Shangfei Wang, Lihua Yue, and Qiang Ji. Multiple-facial action unit recognition by shared feature learning and semantic relation modeling. In Pattern Recognition (ICPR), nd International Conference on, pages , Aug [8] Xiao Zhang and M.H. Mahoor. Simultaneous detection of multiple facial action units via hierarchical task structure learning. In Pattern Recognition (ICPR), nd International Conference on, pages , Aug [9] Kaili Zhao, Wen-Sheng Chu, Fernando De la Torre Frade, Jeffrey Cohn, and Honggang Zhang. Joint patch and multi-label learning for facial action unit detection. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June [10] Stefanos Eleftheriadis, Ognjen Rudovic, and Maja Pantic. Multiconditional latent variable model for joint facial action unit detection. In Proceedings of the IEEE International Conference on Computer Vision, pages , [11] Yale Song, Daniel McDuff, Deepak Vasisht, and Ashish Kapoor. Exploiting sparsity and co-occurrence structure for action unit recognition. In Automatic Face and Gesture Recognition (FG), th IEEE International Conference and Workshops on, volume 1, pages 1 8. IEEE, [12] Jun Wang, Shangfei Wang, and Qiang Ji. Facial action unit classification with hidden knowledge under incomplete annotation. In Proceedings of the 5th ACM on International Conference on Multimedia Retrieval, pages ACM, [13] Adria Ruiz, Joost Van de Weijer, and Xavier Binefa. From emotions to action units with hidden and semi-hidden-task learning. In Proceedings of the IEEE International Conference on Computer Vision, pages , [14] Patrick Lucey, Jeffrey F Cohn, Takeo Kanade, Jason Saragih, Zara Ambadar, and Iain Matthews. The extended cohn-kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In Computer Vision and Pattern Recognition Workshops (CVPRW), 2010 IEEE Computer Society Conference on, pages IEEE, [15] Xing Zhang, Lijun Yin, Jeffrey F Cohn, Shaun Canavan, Michael Reale, Andy Horowitz, Peng Liu, and Jeffrey M Girard. Bp4d-spontaneous: a high-resolution spontaneous 3d dynamic facial expression database. Image and Vision Computing, 32(10): , [16] Qiang Ji. Rpi intelligent systems lab (isl) image databases. ecse.rpi.edu/homepages/cvrl/database/database.html. [17] Mohammad S Sorower. A literature survey on algorithms for multi-label learning. Oregon State University, Corvallis, [18] Amogh Gudi, H Emrah Tasli, Tim M den Uyl, and Andreas Maroulis. Deep learning based facs action unit occurrence and intensity estimation. In Automatic Face and Gesture Recognition (FG), th IEEE International Conference and Workshops on, volume 6, pages 1 5. IEEE,

Learning the Deep Features for Eye Detection in Uncontrolled Conditions

Learning the Deep Features for Eye Detection in Uncontrolled Conditions 2014 22nd International Conference on Pattern Recognition Learning the Deep Features for Eye Detection in Uncontrolled Conditions Yue Wu Dept. of ECSE, Rensselaer Polytechnic Institute Troy, NY, USA 12180

More information

Constrained Joint Cascade Regression Framework for Simultaneous Facial Action Unit Recognition and Facial Landmark Detection

Constrained Joint Cascade Regression Framework for Simultaneous Facial Action Unit Recognition and Facial Landmark Detection Constrained Joint Cascade Regression Framework for Simultaneous Facial Action Unit Recognition and Facial Landmark Detection Yue Wu Qiang Ji ECSE Department, Rensselaer Polytechnic Institute 110 8th street,

More information

Multi-Instance Hidden Markov Model For Facial Expression Recognition

Multi-Instance Hidden Markov Model For Facial Expression Recognition Multi-Instance Hidden Markov Model For Facial Expression Recognition Chongliang Wu 1, Shangfei Wang 1 and Qiang Ji 2 1 School of Computer Science and Technology, University of Science and Technology of

More information

Capturing Dependencies among Labels and Features for Multiple Emotion Tagging of Multimedia Data

Capturing Dependencies among Labels and Features for Multiple Emotion Tagging of Multimedia Data Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17) Capturing Dependencies among Labels and Features for Multiple Emotion Tagging of Multimedia Data Shan Wu, 1 Shangfei

More information

Action Unit Based Facial Expression Recognition Using Deep Learning

Action Unit Based Facial Expression Recognition Using Deep Learning Action Unit Based Facial Expression Recognition Using Deep Learning Salah Al-Darraji 1, Karsten Berns 1, and Aleksandar Rodić 2 1 Robotics Research Lab, Department of Computer Science, University of Kaiserslautern,

More information

A Hierarchical Probabilistic Model for Facial Feature Detection

A Hierarchical Probabilistic Model for Facial Feature Detection A Hierarchical Probabilistic Model for Facial Feature Detection Yue Wu Ziheng Wang Qiang Ji ECSE Department, Rensselaer Polytechnic Institute {wuy9,wangz1,jiq}@rpi.edu Abstract Facial feature detection

More information

EAC-Net: A Region-based Deep Enhancing and Cropping Approach for Facial Action Unit Detection

EAC-Net: A Region-based Deep Enhancing and Cropping Approach for Facial Action Unit Detection EAC-Net: A Region-based Deep Enhancing and Cropping Approach for Facial Action Unit Detection Wei Li 1 Farnaz Abtahi 2 Zhigang Zhu 1 2 Lijun Yin 3 1 Grove School of Engineering, The CUNY City College,

More information

Facial Expression Analysis

Facial Expression Analysis Facial Expression Analysis Jeff Cohn Fernando De la Torre Human Sensing Laboratory Tutorial Looking @ People June 2012 Facial Expression Analysis F. De la Torre/J. Cohn Looking @ People (CVPR-12) 1 Outline

More information

Recognition of Facial Action Units with Action Unit Classifiers and An Association Network

Recognition of Facial Action Units with Action Unit Classifiers and An Association Network Recognition of Facial Action Units with Action Unit Classifiers and An Association Network Junkai Chen 1, Zenghai Chen 1, Zheru Chi 1 and Hong Fu 1,2 1 Department of Electronic and Information Engineering,

More information

Facial Feature Tracking under Varying Facial Expressions and Face Poses based on Restricted Boltzmann Machines

Facial Feature Tracking under Varying Facial Expressions and Face Poses based on Restricted Boltzmann Machines Facial Feature Tracking under Varying Facial Expressions and Face Poses based on Restricted Boltzmann Machines Yue Wu Zuoguan Wang Qiang Ji ECSE Department, Rensselaer Polytechnic Institute {wuy9,wangz6,iq}@rpi.edu

More information

Boosting Coded Dynamic Features for Facial Action Units and Facial Expression Recognition

Boosting Coded Dynamic Features for Facial Action Units and Facial Expression Recognition Boosting Coded Dynamic Features for Facial Action Units and Facial Expression Recognition Peng Yang Qingshan Liu,2 Dimitris N. Metaxas Computer Science Department, Rutgers University Frelinghuysen Road,

More information

Real time facial expression recognition from image sequences using Support Vector Machines

Real time facial expression recognition from image sequences using Support Vector Machines Real time facial expression recognition from image sequences using Support Vector Machines I. Kotsia a and I. Pitas a a Aristotle University of Thessaloniki, Department of Informatics, Box 451, 54124 Thessaloniki,

More information

Convolutional Neural Networks for Facial Expression Recognition

Convolutional Neural Networks for Facial Expression Recognition Convolutional Neural Networks for Facial Expression Recognition Shima Alizadeh Stanford University shima86@stanford.edu Azar Fazel Stanford University azarf@stanford.edu Abstract In this project, we have

More information

Markov Random Fields and Gibbs Sampling for Image Denoising

Markov Random Fields and Gibbs Sampling for Image Denoising Markov Random Fields and Gibbs Sampling for Image Denoising Chang Yue Electrical Engineering Stanford University changyue@stanfoed.edu Abstract This project applies Gibbs Sampling based on different Markov

More information

Multiple Kernel Learning for Emotion Recognition in the Wild

Multiple Kernel Learning for Emotion Recognition in the Wild Multiple Kernel Learning for Emotion Recognition in the Wild Karan Sikka, Karmen Dykstra, Suchitra Sathyanarayana, Gwen Littlewort and Marian S. Bartlett Machine Perception Laboratory UCSD EmotiW Challenge,

More information

HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION

HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION Dipankar Das Department of Information and Communication Engineering, University of Rajshahi, Rajshahi-6205, Bangladesh ABSTRACT Real-time

More information

Efficient Feature Learning Using Perturb-and-MAP

Efficient Feature Learning Using Perturb-and-MAP Efficient Feature Learning Using Perturb-and-MAP Ke Li, Kevin Swersky, Richard Zemel Dept. of Computer Science, University of Toronto {keli,kswersky,zemel}@cs.toronto.edu Abstract Perturb-and-MAP [1] is

More information

Facial Expression Recognition Using Non-negative Matrix Factorization

Facial Expression Recognition Using Non-negative Matrix Factorization Facial Expression Recognition Using Non-negative Matrix Factorization Symeon Nikitidis, Anastasios Tefas and Ioannis Pitas Artificial Intelligence & Information Analysis Lab Department of Informatics Aristotle,

More information

Partial Least Squares Regression on Grassmannian Manifold for Emotion Recognition

Partial Least Squares Regression on Grassmannian Manifold for Emotion Recognition Emotion Recognition In The Wild Challenge and Workshop (EmotiW 2013) Partial Least Squares Regression on Grassmannian Manifold for Emotion Recognition Mengyi Liu, Ruiping Wang, Zhiwu Huang, Shiguang Shan,

More information

A Novel Feature Extraction Technique for Facial Expression Recognition

A Novel Feature Extraction Technique for Facial Expression Recognition www.ijcsi.org 9 A Novel Feature Extraction Technique for Facial Expression Recognition *Mohammad Shahidul Islam 1, Surapong Auwatanamongkol 2 1 Department of Computer Science, School of Applied Statistics,

More information

An Algorithm based on SURF and LBP approach for Facial Expression Recognition

An Algorithm based on SURF and LBP approach for Facial Expression Recognition ISSN: 2454-2377, An Algorithm based on SURF and LBP approach for Facial Expression Recognition Neha Sahu 1*, Chhavi Sharma 2, Hitesh Yadav 3 1 Assistant Professor, CSE/IT, The North Cap University, Gurgaon,

More information

Fully Automatic Facial Action Recognition in Spontaneous Behavior

Fully Automatic Facial Action Recognition in Spontaneous Behavior Fully Automatic Facial Action Recognition in Spontaneous Behavior Marian Stewart Bartlett 1, Gwen Littlewort 1, Mark Frank 2, Claudia Lainscsek 1, Ian Fasel 1, Javier Movellan 1 1 Institute for Neural

More information

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN 2016 International Conference on Artificial Intelligence: Techniques and Applications (AITA 2016) ISBN: 978-1-60595-389-2 Face Recognition Using Vector Quantization Histogram and Support Vector Machine

More information

A Framework for Joint Estimation and Guided Annotation of Facial Action Unit Intensity

A Framework for Joint Estimation and Guided Annotation of Facial Action Unit Intensity 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops A Framework for Joint Estimation and Guided Annotation of Facial Action Unit Intensity Robert Walecki, Ognjen Rudovic, Maja Pantic,

More information

Continuous AU Intensity Estimation using Localized, Sparse Facial Feature Space

Continuous AU Intensity Estimation using Localized, Sparse Facial Feature Space Continuous AU Intensity Estimation using Localized, Sparse Facial Feature Space László A. Jeni, Jeffrey M. Girard, Jeffrey F. Cohn and Fernando De La Torre Abstract Most work in automatic facial expression

More information

Image Processing Pipeline for Facial Expression Recognition under Variable Lighting

Image Processing Pipeline for Facial Expression Recognition under Variable Lighting Image Processing Pipeline for Facial Expression Recognition under Variable Lighting Ralph Ma, Amr Mohamed ralphma@stanford.edu, amr1@stanford.edu Abstract Much research has been done in the field of automated

More information

Facial Expression Recognition with PCA and LBP Features Extracting from Active Facial Patches

Facial Expression Recognition with PCA and LBP Features Extracting from Active Facial Patches Facial Expression Recognition with PCA and LBP Features Extracting from Active Facial Patches Yanpeng Liu a, Yuwen Cao a, Yibin Li a, Ming Liu, Rui Song a Yafang Wang, Zhigang Xu, Xin Ma a Abstract Facial

More information

A Novel LDA and HMM-based technique for Emotion Recognition from Facial Expressions

A Novel LDA and HMM-based technique for Emotion Recognition from Facial Expressions A Novel LDA and HMM-based technique for Emotion Recognition from Facial Expressions Akhil Bansal, Santanu Chaudhary, Sumantra Dutta Roy Indian Institute of Technology, Delhi, India akhil.engg86@gmail.com,

More information

Cross-pose Facial Expression Recognition

Cross-pose Facial Expression Recognition Cross-pose Facial Expression Recognition Abstract In real world facial expression recognition (FER) applications, it is not practical for a user to enroll his/her facial expressions under different pose

More information

Deep Boltzmann Machines

Deep Boltzmann Machines Deep Boltzmann Machines Sargur N. Srihari srihari@cedar.buffalo.edu Topics 1. Boltzmann machines 2. Restricted Boltzmann machines 3. Deep Belief Networks 4. Deep Boltzmann machines 5. Boltzmann machines

More information

Facial Expression Analysis via Transfer Learning

Facial Expression Analysis via Transfer Learning University of Denver Digital Commons @ DU Electronic Theses and Dissertations Graduate Studies 1-1-2015 Facial Expression Analysis via Transfer Learning Xiao Zhang University of Denver Follow this and

More information

Computer vision: models, learning and inference. Chapter 10 Graphical Models

Computer vision: models, learning and inference. Chapter 10 Graphical Models Computer vision: models, learning and inference Chapter 10 Graphical Models Independence Two variables x 1 and x 2 are independent if their joint probability distribution factorizes as Pr(x 1, x 2 )=Pr(x

More information

Facial Expression Recognition Based on Local Directional Pattern Using SVM Decision-level Fusion

Facial Expression Recognition Based on Local Directional Pattern Using SVM Decision-level Fusion Facial Expression Recognition Based on Local Directional Pattern Using SVM Decision-level Fusion Juxiang Zhou 1, Tianwei Xu 2, Jianhou Gan 1 1. Key Laboratory of Education Informalization for Nationalities,

More information

Discriminative Deep Face Shape Model for Facial Point Detection

Discriminative Deep Face Shape Model for Facial Point Detection Noname manuscript No. (will be inserted by the editor) Discriminative Deep Face Shape Model for Facial Point Detection Yue Wu Qiang Ji Received: 9 Feb. 24 / Accepted: Oct. 24 Abstract Facial point detection

More information

Dynamic facial expression recognition using a behavioural model

Dynamic facial expression recognition using a behavioural model Dynamic facial expression recognition using a behavioural model Thomas Robin Michel Bierlaire Javier Cruz STRC 2009 10th september The context Recent interest for emotion recognition in transportation

More information

Learning Bayesian Networks with Qualitative Constraints

Learning Bayesian Networks with Qualitative Constraints Learning Bayesian Networks with Qualitative Constraints Yan Tong GE Global Research Center Niskayuna, NY239 Qiang Ji Rensselaer Polytechnic Institute Troy, NY28 Abstract Graphical models such as Bayesian

More information

Latent Regression Bayesian Network for Data Representation

Latent Regression Bayesian Network for Data Representation 2016 23rd International Conference on Pattern Recognition (ICPR) Cancún Center, Cancún, México, December 4-8, 2016 Latent Regression Bayesian Network for Data Representation Siqi Nie Department of Electrical,

More information

Exploring Facial Expressions with Compositional Features

Exploring Facial Expressions with Compositional Features Exploring Facial Expressions with Compositional Features Peng Yang Qingshan Liu Dimitris N. Metaxas Computer Science Department, Rutgers University Frelinghuysen Road, Piscataway, NJ 88, USA peyang@cs.rutgers.edu,

More information

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 9, SEPTEMBER

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 9, SEPTEMBER IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 9, SEPTEMBER 2011 2401 Probabilistic Image Modeling With an Extended Chain Graph for Human Activity Recognition and Image Segmentation Lei Zhang, Member,

More information

3D Facial Action Units Recognition for Emotional Expression

3D Facial Action Units Recognition for Emotional Expression 3D Facial Action Units Recognition for Emotional Expression Norhaida Hussain 1, Hamimah Ujir, Irwandi Hipiny and Jacey-Lynn Minoi 1 Department of Information Technology and Communication, Politeknik Kuching,

More information

FACSCaps: Pose-Independent Facial Action Coding with Capsules

FACSCaps: Pose-Independent Facial Action Coding with Capsules FACSCaps: Pose-Independent Facial Action Coding with Capsules Itir Onal Ertugrul 1, László A. Jeni 1, Jeffrey F. Cohn 1,2 1 Robotics Institute, Carnegie Mellon University, Pittsburgh, PA, USA 2 Department

More information

Facial Emotion Recognition using Eye

Facial Emotion Recognition using Eye Facial Emotion Recognition using Eye Vishnu Priya R 1 and Muralidhar A 2 1 School of Computing Science and Engineering, VIT Chennai Campus, Tamil Nadu, India. Orcid: 0000-0002-2016-0066 2 School of Computing

More information

Facial Action Unit Detection using Active Learning and an Efficient Non-Linear Kernel Approximation

Facial Action Unit Detection using Active Learning and an Efficient Non-Linear Kernel Approximation Facial Action Unit Detection using Active Learning and an Efficient Non-Linear Kernel Approximation Thibaud Senechal, Daniel McDuff and Rana el Kaliouby Affectiva Waltham, MA, 02452, USA thibaud.senechal@gmail.com,

More information

Robust Facial Expression Classification Using Shape and Appearance Features

Robust Facial Expression Classification Using Shape and Appearance Features Robust Facial Expression Classification Using Shape and Appearance Features S L Happy and Aurobinda Routray Department of Electrical Engineering, Indian Institute of Technology Kharagpur, India Abstract

More information

A Hierarchical Face Identification System Based on Facial Components

A Hierarchical Face Identification System Based on Facial Components A Hierarchical Face Identification System Based on Facial Components Mehrtash T. Harandi, Majid Nili Ahmadabadi, and Babak N. Araabi Control and Intelligent Processing Center of Excellence Department of

More information

A Fast Learning Algorithm for Deep Belief Nets

A Fast Learning Algorithm for Deep Belief Nets A Fast Learning Algorithm for Deep Belief Nets Geoffrey E. Hinton, Simon Osindero Department of Computer Science University of Toronto, Toronto, Canada Yee-Whye Teh Department of Computer Science National

More information

arxiv: v1 [cs.cv] 29 Sep 2016

arxiv: v1 [cs.cv] 29 Sep 2016 arxiv:1609.09545v1 [cs.cv] 29 Sep 2016 Two-stage Convolutional Part Heatmap Regression for the 1st 3D Face Alignment in the Wild (3DFAW) Challenge Adrian Bulat and Georgios Tzimiropoulos Computer Vision

More information

Emotion Detection System using Facial Action Coding System

Emotion Detection System using Facial Action Coding System International Journal of Engineering and Technical Research (IJETR) Emotion Detection System using Facial Action Coding System Vedant Chauhan, Yash Agrawal, Vinay Bhutada Abstract Behaviors, poses, actions,

More information

Unifying Geometric Features and Facial Action Units for Improved Performance of Facial Expression Analysis

Unifying Geometric Features and Facial Action Units for Improved Performance of Facial Expression Analysis Unifying Geometric Features and Facial Action Units for Improved Performance of Facial Expression Analysis Mehdi Ghayoumi 1, Arvind K Bansal 1 1 Computer Science Department, Kent State University, {mghayoum,akbansal}@kent.edu

More information

Building Classifiers using Bayesian Networks

Building Classifiers using Bayesian Networks Building Classifiers using Bayesian Networks Nir Friedman and Moises Goldszmidt 1997 Presented by Brian Collins and Lukas Seitlinger Paper Summary The Naive Bayes classifier has reasonable performance

More information

Shape Augmented Regression for 3D Face Alignment

Shape Augmented Regression for 3D Face Alignment Shape Augmented Regression for 3D Face Alignment Chao Gou 1,3(B),YueWu 2, Fei-Yue Wang 1,3, and Qiang Ji 2 1 Institute of Automation, Chinese Academy of Sciences, Beijing, China {gouchao2012,feiyue.wang}@ia.ac.cn

More information

arxiv: v1 [cs.cv] 23 Sep 2017

arxiv: v1 [cs.cv] 23 Sep 2017 Constrained Deep Transfer Feature Learning and its Applications Yue Wu Qiang Ji ECSE Department, Rensselaer Polytechnic Institute 0 8th street, Troy, NY, USA {wuy9, jiq}@rpi.edu arxiv:709.0828v [cs.cv]

More information

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS Kuan-Chuan Peng and Tsuhan Chen School of Electrical and Computer Engineering, Cornell University, Ithaca, NY

More information

Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling

Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling [DOI: 10.2197/ipsjtcva.7.99] Express Paper Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling Takayoshi Yamashita 1,a) Takaya Nakamura 1 Hiroshi Fukui 1,b) Yuji

More information

Partial Least Squares Regression on Grassmannian Manifold for Emotion Recognition

Partial Least Squares Regression on Grassmannian Manifold for Emotion Recognition Partial Least Squares Regression on Grassmannian Manifold for Emotion Recognition Mengyi Liu, Ruiping Wang, Zhiwu Huang, Shiguang Shan, Xilin Chen Key Lab of Intelligence Information Processing Institute

More information

Facial Expression Recognition Using a Hybrid CNN SIFT Aggregator

Facial Expression Recognition Using a Hybrid CNN SIFT Aggregator Facial Expression Recognition Using a Hybrid CNN SIFT Aggregator Mundher Al-Shabi, Wooi Ping Cheah, Tee Connie Faculty of Information Science and Technology, Multimedia University, Melaka, Malaysia Abstract.

More information

Using the Deformable Part Model with Autoencoded Feature Descriptors for Object Detection

Using the Deformable Part Model with Autoencoded Feature Descriptors for Object Detection Using the Deformable Part Model with Autoencoded Feature Descriptors for Object Detection Hyunghoon Cho and David Wu December 10, 2010 1 Introduction Given its performance in recent years' PASCAL Visual

More information

C.R VIMALCHAND ABSTRACT

C.R VIMALCHAND ABSTRACT International Journal of Scientific & Engineering Research, Volume 5, Issue 3, March-2014 1173 ANALYSIS OF FACE RECOGNITION SYSTEM WITH FACIAL EXPRESSION USING CONVOLUTIONAL NEURAL NETWORK AND EXTRACTED

More information

Facial expression recognition based on two-step feature histogram optimization Ling Gana, Sisi Sib

Facial expression recognition based on two-step feature histogram optimization Ling Gana, Sisi Sib 3rd International Conference on Materials Engineering, Manufacturing Technology and Control (ICMEMTC 201) Facial expression recognition based on two-step feature histogram optimization Ling Gana, Sisi

More information

Gradient of the lower bound

Gradient of the lower bound Weakly Supervised with Latent PhD advisor: Dr. Ambedkar Dukkipati Department of Computer Science and Automation gaurav.pandey@csa.iisc.ernet.in Objective Given a training set that comprises image and image-level

More information

Generic Face Alignment Using an Improved Active Shape Model

Generic Face Alignment Using an Improved Active Shape Model Generic Face Alignment Using an Improved Active Shape Model Liting Wang, Xiaoqing Ding, Chi Fang Electronic Engineering Department, Tsinghua University, Beijing, China {wanglt, dxq, fangchi} @ocrserv.ee.tsinghua.edu.cn

More information

Dynamic facial expression recognition by joint static and multi-time gap transition classification

Dynamic facial expression recognition by joint static and multi-time gap transition classification Dynamic facial expression recognition by joint static and multi-time gap transition classification Arnaud Dapogny 1 and Kevin Bailly 1 and Séverine Dubuisson 1 1 CNRS, UMR 7222, ISIR, F-75005, Paris, France

More information

arxiv: v1 [cs.cv] 21 Oct 2017

arxiv: v1 [cs.cv] 21 Oct 2017 arxiv:1710.07831v1 [cs.cv] 21 Oct 2017 A Generative Restricted Boltzmann Machine Based Method for High-Dimensional Motion Data Modeling Siqi Nie Ziheng Wang Qiang Ji October 24, 2017 Abstract Many computer

More information

Classification of Upper and Lower Face Action Units and Facial Expressions using Hybrid Tracking System and Probabilistic Neural Networks

Classification of Upper and Lower Face Action Units and Facial Expressions using Hybrid Tracking System and Probabilistic Neural Networks Classification of Upper and Lower Face Action Units and Facial Expressions using Hybrid Tracking System and Probabilistic Neural Networks HADI SEYEDARABI*, WON-SOOK LEE**, ALI AGHAGOLZADEH* AND SOHRAB

More information

LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM

LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM Hazim Kemal Ekenel, Rainer Stiefelhagen Interactive Systems Labs, University of Karlsruhe Am Fasanengarten 5, 76131, Karlsruhe, Germany

More information

Energy Based Models, Restricted Boltzmann Machines and Deep Networks. Jesse Eickholt

Energy Based Models, Restricted Boltzmann Machines and Deep Networks. Jesse Eickholt Energy Based Models, Restricted Boltzmann Machines and Deep Networks Jesse Eickholt ???? Who s heard of Energy Based Models (EBMs) Restricted Boltzmann Machines (RBMs) Deep Belief Networks Auto-encoders

More information

Facial expression recognition using shape and texture information

Facial expression recognition using shape and texture information 1 Facial expression recognition using shape and texture information I. Kotsia 1 and I. Pitas 1 Aristotle University of Thessaloniki pitas@aiia.csd.auth.gr Department of Informatics Box 451 54124 Thessaloniki,

More information

LBP Based Facial Expression Recognition Using k-nn Classifier

LBP Based Facial Expression Recognition Using k-nn Classifier ISSN 2395-1621 LBP Based Facial Expression Recognition Using k-nn Classifier #1 Chethan Singh. A, #2 Gowtham. N, #3 John Freddy. M, #4 Kashinath. N, #5 Mrs. Vijayalakshmi. G.V 1 chethan.singh1994@gmail.com

More information

An efficient face recognition algorithm based on multi-kernel regularization learning

An efficient face recognition algorithm based on multi-kernel regularization learning Acta Technica 61, No. 4A/2016, 75 84 c 2017 Institute of Thermomechanics CAS, v.v.i. An efficient face recognition algorithm based on multi-kernel regularization learning Bi Rongrong 1 Abstract. A novel

More information

A Novel Extreme Point Selection Algorithm in SIFT

A Novel Extreme Point Selection Algorithm in SIFT A Novel Extreme Point Selection Algorithm in SIFT Ding Zuchun School of Electronic and Communication, South China University of Technolog Guangzhou, China zucding@gmail.com Abstract. This paper proposes

More information

Classification of Face Images for Gender, Age, Facial Expression, and Identity 1

Classification of Face Images for Gender, Age, Facial Expression, and Identity 1 Proc. Int. Conf. on Artificial Neural Networks (ICANN 05), Warsaw, LNCS 3696, vol. I, pp. 569-574, Springer Verlag 2005 Classification of Face Images for Gender, Age, Facial Expression, and Identity 1

More information

Video annotation based on adaptive annular spatial partition scheme

Video annotation based on adaptive annular spatial partition scheme Video annotation based on adaptive annular spatial partition scheme Guiguang Ding a), Lu Zhang, and Xiaoxu Li Key Laboratory for Information System Security, Ministry of Education, Tsinghua National Laboratory

More information

DA Progress report 2 Multi-view facial expression. classification Nikolas Hesse

DA Progress report 2 Multi-view facial expression. classification Nikolas Hesse DA Progress report 2 Multi-view facial expression classification 16.12.2010 Nikolas Hesse Motivation Facial expressions (FE) play an important role in interpersonal communication FE recognition can help

More information

Object Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision

Object Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision Object Recognition Using Pictorial Structures Daniel Huttenlocher Computer Science Department Joint work with Pedro Felzenszwalb, MIT AI Lab In This Talk Object recognition in computer vision Brief definition

More information

Eye Detection by Haar wavelets and cascaded Support Vector Machine

Eye Detection by Haar wavelets and cascaded Support Vector Machine Eye Detection by Haar wavelets and cascaded Support Vector Machine Vishal Agrawal B.Tech 4th Year Guide: Simant Dubey / Amitabha Mukherjee Dept of Computer Science and Engineering IIT Kanpur - 208 016

More information

Spatiotemporal Features for Effective Facial Expression Recognition

Spatiotemporal Features for Effective Facial Expression Recognition Spatiotemporal Features for Effective Facial Expression Recognition Hatice Çınar Akakın and Bülent Sankur Bogazici University, Electrical & Electronics Engineering Department, Bebek, Istanbul {hatice.cinar,bulent.sankur}@boun.edu.tr

More information

IJREAT International Journal of Research in Engineering & Advanced Technology, Volume 1, Issue 5, Oct-Nov, 2013 ISSN:

IJREAT International Journal of Research in Engineering & Advanced Technology, Volume 1, Issue 5, Oct-Nov, 2013 ISSN: Semi Automatic Annotation Exploitation Similarity of Pics in i Personal Photo Albums P. Subashree Kasi Thangam 1 and R. Rosy Angel 2 1 Assistant Professor, Department of Computer Science Engineering College,

More information

Sum-Product Networks. STAT946 Deep Learning Guest Lecture by Pascal Poupart University of Waterloo October 15, 2015

Sum-Product Networks. STAT946 Deep Learning Guest Lecture by Pascal Poupart University of Waterloo October 15, 2015 Sum-Product Networks STAT946 Deep Learning Guest Lecture by Pascal Poupart University of Waterloo October 15, 2015 Introduction Outline What is a Sum-Product Network? Inference Applications In more depth

More information

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,

More information

The Computer Expression Recognition Toolbox (CERT)

The Computer Expression Recognition Toolbox (CERT) The Computer Expression Recognition Toolbox (CERT) Gwen Littlewort 1, Jacob Whitehill 1, Tingfan Wu 1, Ian Fasel 2, Mark Frank 3, Javier Movellan 1, and Marian Bartlett 1 {gwen, jake, ting, movellan}@mplab.ucsd.edu,

More information

Facial Action Unit Detection using Active Learning and an Efficient Non-Linear Kernel Approximation

Facial Action Unit Detection using Active Learning and an Efficient Non-Linear Kernel Approximation Facial Action Unit Detection using Active Learning and an Efficient Non-Linear Kernel Approximation Thibaud Senechal, Daniel McDuff and Rana el Kaliouby Affectiva Waltham, MA, 02452, USA thibaud.senechal@gmail.com,

More information

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Estimating Human Pose in Images. Navraj Singh December 11, 2009 Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks

More information

Exploring Bag of Words Architectures in the Facial Expression Domain

Exploring Bag of Words Architectures in the Facial Expression Domain Exploring Bag of Words Architectures in the Facial Expression Domain Karan Sikka, Tingfan Wu, Josh Susskind, and Marian Bartlett Machine Perception Laboratory, University of California San Diego {ksikka,ting,josh,marni}@mplab.ucsd.edu

More information

FACIAL EXPRESSION USING 3D ANIMATION

FACIAL EXPRESSION USING 3D ANIMATION Volume 1 Issue 1 May 2010 pp. 1 7 http://iaeme.com/ijcet.html I J C E T IAEME FACIAL EXPRESSION USING 3D ANIMATION Mr. K. Gnanamuthu Prakash 1, Dr. S. Balasubramanian 2 ABSTRACT Traditionally, human facial

More information

Dynamic Facial Expression Recognition Using A Bayesian Temporal Manifold Model

Dynamic Facial Expression Recognition Using A Bayesian Temporal Manifold Model Dynamic Facial Expression Recognition Using A Bayesian Temporal Manifold Model Caifeng Shan, Shaogang Gong, and Peter W. McOwan Department of Computer Science Queen Mary University of London Mile End Road,

More information

Probabilistic Abstraction Lattices: A Computationally Efficient Model for Conditional Probability Estimation

Probabilistic Abstraction Lattices: A Computationally Efficient Model for Conditional Probability Estimation Probabilistic Abstraction Lattices: A Computationally Efficient Model for Conditional Probability Estimation Daniel Lowd January 14, 2004 1 Introduction Probabilistic models have shown increasing popularity

More information

Human Motion Detection and Tracking for Video Surveillance

Human Motion Detection and Tracking for Video Surveillance Human Motion Detection and Tracking for Video Surveillance Prithviraj Banerjee and Somnath Sengupta Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur,

More information

Multi-View Face Tracking with Factorial and Switching HMM

Multi-View Face Tracking with Factorial and Switching HMM Multi-View Face Tracking with Factorial and Switching HMM Peng Wang, Qiang Ji Department of Electrical, Computer and System Engineering Rensselaer Polytechnic Institute Troy, NY 12180 Abstract Dynamic

More information

Robust Facial Landmark Detection under Significant Head Poses and Occlusion

Robust Facial Landmark Detection under Significant Head Poses and Occlusion Robust Facial Landmark Detection under Significant Head Poses and Occlusion Yue Wu Qiang Ji ECSE Department, Rensselaer Polytechnic Institute 110 8th street, Troy, NY, USA {wuy9,jiq}@rpi.edu Abstract There

More information

Emotional Expression Classification using Time-Series Kernels

Emotional Expression Classification using Time-Series Kernels Emotional Expression Classification using Time-Series Kernels András Lőrincz 1, László A. Jeni 2, Zoltán Szabó 1, Jeffrey F. Cohn 2, 3, and Takeo Kanade 2 1 Eötvös Loránd University, Budapest, Hungary,

More information

To be Bernoulli or to be Gaussian, for a Restricted Boltzmann Machine

To be Bernoulli or to be Gaussian, for a Restricted Boltzmann Machine 2014 22nd International Conference on Pattern Recognition To be Bernoulli or to be Gaussian, for a Restricted Boltzmann Machine Takayoshi Yamashita, Masayuki Tanaka, Eiji Yoshida, Yuji Yamauchi and Hironobu

More information

Human Detection and Tracking for Video Surveillance: A Cognitive Science Approach

Human Detection and Tracking for Video Surveillance: A Cognitive Science Approach Human Detection and Tracking for Video Surveillance: A Cognitive Science Approach Vandit Gajjar gajjar.vandit.381@ldce.ac.in Ayesha Gurnani gurnani.ayesha.52@ldce.ac.in Yash Khandhediya khandhediya.yash.364@ldce.ac.in

More information

Structured Models in. Dan Huttenlocher. June 2010

Structured Models in. Dan Huttenlocher. June 2010 Structured Models in Computer Vision i Dan Huttenlocher June 2010 Structured Models Problems where output variables are mutually dependent or constrained E.g., spatial or temporal relations Such dependencies

More information

3D Shape Estimation in Video Sequences Provides High Precision Evaluation of Facial Expressions

3D Shape Estimation in Video Sequences Provides High Precision Evaluation of Facial Expressions 3D Shape Estimation in Video Sequences Provides High Precision Evaluation of Facial Expressions László A. Jeni a, András Lőrincz b, Tamás Nagy b, Zsolt Palotai c, Judit Sebők b, Zoltán Szabó b, Dániel

More information

Recognizing Partial Facial Action Units Based on 3D Dynamic Range Data for Facial Expression Recognition

Recognizing Partial Facial Action Units Based on 3D Dynamic Range Data for Facial Expression Recognition Recognizing Partial Facial Action Units Based on 3D Dynamic Range Data for Facial Expression Recognition Yi Sun, Michael Reale, and Lijun Yin Department of Computer Science, State University of New York

More information

3 : Representation of Undirected GMs

3 : Representation of Undirected GMs 0-708: Probabilistic Graphical Models 0-708, Spring 202 3 : Representation of Undirected GMs Lecturer: Eric P. Xing Scribes: Nicole Rafidi, Kirstin Early Last Time In the last lecture, we discussed directed

More information

Combining PGMs and Discriminative Models for Upper Body Pose Detection

Combining PGMs and Discriminative Models for Upper Body Pose Detection Combining PGMs and Discriminative Models for Upper Body Pose Detection Gedas Bertasius May 30, 2014 1 Introduction In this project, I utilized probabilistic graphical models together with discriminative

More information

A Network Intrusion Detection System Architecture Based on Snort and. Computational Intelligence

A Network Intrusion Detection System Architecture Based on Snort and. Computational Intelligence 2nd International Conference on Electronics, Network and Computer Engineering (ICENCE 206) A Network Intrusion Detection System Architecture Based on Snort and Computational Intelligence Tao Liu, a, Da

More information

AUTOMATIC VIDEO INDEXING

AUTOMATIC VIDEO INDEXING AUTOMATIC VIDEO INDEXING Itxaso Bustos Maite Frutos TABLE OF CONTENTS Introduction Methods Key-frame extraction Automatic visual indexing Shot boundary detection Video OCR Index in motion Image processing

More information

Convolutional Restricted Boltzmann Machine Features for TD Learning in Go

Convolutional Restricted Boltzmann Machine Features for TD Learning in Go ConvolutionalRestrictedBoltzmannMachineFeatures fortdlearningingo ByYanLargmanandPeterPham AdvisedbyHonglakLee 1.Background&Motivation AlthoughrecentadvancesinAIhaveallowed Go playing programs to become

More information