Multi-Instance Hidden Markov Model For Facial Expression Recognition

Similar documents
Learning the Deep Features for Eye Detection in Uncontrolled Conditions

Action Unit Based Facial Expression Recognition Using Deep Learning

Facial expression recognition using shape and texture information

Facial Expression Analysis

A Hierarchical Probabilistic Model for Facial Feature Detection

Constrained Joint Cascade Regression Framework for Simultaneous Facial Action Unit Recognition and Facial Landmark Detection

A Novel LDA and HMM-based technique for Emotion Recognition from Facial Expressions

Exemplar Hidden Markov Models for Classification of Facial Expressions in Videos

Dynamic Facial Expression Recognition Using A Bayesian Temporal Manifold Model

Boosting Coded Dynamic Features for Facial Action Units and Facial Expression Recognition

Facial Expression Recognition Using Non-negative Matrix Factorization

Recognition of Facial Action Units with Action Unit Classifiers and An Association Network

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves

Facial Feature Tracking under Varying Facial Expressions and Face Poses based on Restricted Boltzmann Machines

Research on Dynamic Facial Expressions Recognition

Multiple Facial Action Unit Recognition Enhanced by Facial Expressions

Dynamic Facial Expression Recognition with Atlas Construction and Sparse Representation

Appearance Manifold of Facial Expression

DA Progress report 2 Multi-view facial expression. classification Nikolas Hesse

Real time facial expression recognition from image sequences using Support Vector Machines

Dynamic facial expression recognition using a behavioural model

An Algorithm based on SURF and LBP approach for Facial Expression Recognition

Exploring Bag of Words Architectures in the Facial Expression Domain

Evaluation of Expression Recognition Techniques

3D Facial Action Units Recognition for Emotional Expression

Conditional Random Fields - A probabilistic graphical model. Yen-Chin Lee 指導老師 : 鮑興國

Intensity Rank Estimation of Facial Expressions Based on A Single Image

10-701/15-781, Fall 2006, Final

Facial Emotion Recognition using Eye

Facial Expression Recognition using Principal Component Analysis with Singular Value Decomposition

Classification of Face Images for Gender, Age, Facial Expression, and Identity 1

Multi-view Facial Expression Recognition Analysis with Generic Sparse Coding Feature

Emotion Detection System using Facial Action Coding System

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Facial Expression Recognition with PCA and LBP Features Extracting from Active Facial Patches

Multiple Kernel Learning for Emotion Recognition in the Wild

Multi-output Laplacian Dynamic Ordinal Regression for Facial Expression Recognition and Intensity Estimation

Recognizing Partial Facial Action Units Based on 3D Dynamic Range Data for Facial Expression Recognition

Pain Recognition using Spatiotemporal Oriented Energy of Facial Muscles Irani, Ramin; Nasrollahi, Kamal; Moeslund, Thomas B.

Cross-pose Facial Expression Recognition

A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models

Experts-Shift: Learning Active Spatial Classification Experts for Keyframe-based Video Segmentation

Improving Image Segmentation Quality Via Graph Theory

A Novel Feature Extraction Technique for Facial Expression Recognition

Recognition of facial expressions in presence of partial occlusion

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework

Generic Face Alignment Using an Improved Active Shape Model

Graphical Models, Bayesian Method, Sampling, and Variational Inference

Bayesian Spatiotemporal Modeling with Hierarchical Spatial Priors for fmri

Exploring Facial Expressions with Compositional Features

HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION

C.R VIMALCHAND ABSTRACT

Research on Emotion Recognition for Facial Expression Images Based on Hidden Markov Model

Image retrieval based on bag of images

Convolutional Neural Networks for Facial Expression Recognition

Thesis Proposal : Switching Linear Dynamic Systems with Higher-order Temporal Structure. Sang Min Oh

Capturing Dependencies among Labels and Features for Multiple Emotion Tagging of Multimedia Data

A Real Time Facial Expression Classification System Using Local Binary Patterns

Arm-hand Action Recognition Based on 3D Skeleton Joints Ling RUI 1, Shi-wei MA 1,a, *, Jia-rui WEN 1 and Li-na LIU 1,2

Conditional Random Fields and beyond D A N I E L K H A S H A B I C S U I U C,

Facial Expression Recognition Based on Local Directional Pattern Using SVM Decision-level Fusion

A Novel Multi-instance Learning Algorithm with Application to Image Classification

3D Shape Estimation in Video Sequences Provides High Precision Evaluation of Facial Expressions

An Approach for Face Recognition System Using Convolutional Neural Network and Extracted Geometric Features

Facial Expression Recognition Using a Hybrid CNN SIFT Aggregator

Emotional Expression Classification using Time-Series Kernels

Short Survey on Static Hand Gesture Recognition

k-means demo Administrative Machine learning: Unsupervised learning" Assignment 5 out

arxiv: v1 [cs.cv] 21 Oct 2017

Natural Facial Expression Recognition Using Dynamic and Static Schemes

K-Means and Gaussian Mixture Models

EMOTIONAL BASED FACIAL EXPRESSION RECOGNITION USING SUPPORT VECTOR MACHINE

Emotional Expression Classification using Time-Series Kernels

A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation. Kwanyong Lee 1 and Hyeyoung Park 2

Enhanced Active Shape Models with Global Texture Constraints for Image Analysis

Facial-component-based Bag of Words and PHOG Descriptor for Facial Expression Recognition

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Partial Least Squares Regression on Grassmannian Manifold for Emotion Recognition

Normalized Texture Motifs and Their Application to Statistical Object Modeling

Robust False Positive Detection for Real-Time Multi-Target Tracking

Facial Expression Recognition Using Gabor Motion Energy Filters

MULTI-POSE FACE HALLUCINATION VIA NEIGHBOR EMBEDDING FOR FACIAL COMPONENTS. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo

Partial Least Squares Regression on Grassmannian Manifold for Emotion Recognition

Machine Learning A W 1sst KU. b) [1 P] For the probability distribution P (A, B, C, D) with the factorization

Deeply Learning Deformable Facial Action Parts Model for Dynamic Expression Analysis

CS325 Artificial Intelligence Ch. 20 Unsupervised Machine Learning

Lecture 21 : A Hybrid: Deep Learning and Graphical Models

LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM

Classification of Upper and Lower Face Action Units and Facial Expressions using Hybrid Tracking System and Probabilistic Neural Networks

Computing Gaussian Mixture Models with EM using Equivalence Constraints

K-means and Hierarchical Clustering

An indirect tire identification method based on a two-layered fuzzy scheme

Detecting Coarticulation in Sign Language using Conditional Random Fields

Mixture Models and the EM Algorithm

Network Traffic Measurements and Analysis

Machine Learning A WS15/16 1sst KU Version: January 11, b) [1 P] For the probability distribution P (A, B, C, D) with the factorization

EUSIPCO

Prototype Selection for Handwritten Connected Digits Classification

Facial Expression Recognition Using Encoded Dynamic Features

Image Processing Pipeline for Facial Expression Recognition under Variable Lighting

Transcription:

Multi-Instance Hidden Markov Model For Facial Expression Recognition Chongliang Wu 1, Shangfei Wang 1 and Qiang Ji 2 1 School of Computer Science and Technology, University of Science and Technology of China 2 ECSE Department, Rensselaer Polytechnic Institute Email:{clwzkd@mail.ustc.edu.cn, sfwang@ustc.edu.cn, qji@ecse.rpi.edu} Abstract This paper presents a novel method for facial expression recognition using only sequence level labeling. With facial image sequences containing multiple peaks of expression, our method aims to label these sequences and identifies expressional peaks automatically. We formulate this weakly labeled expression recognition as a multi-instance learning (MIL) problem. First, image sequences are clustered into multiple segments. After segmentation, image sequences are regarded as bags in MIL and the segments in the bags are viewed as instances. Second, bags data are used to train a discriminative classifier which combined multi-instance learning and discriminative Hidden Markov Model (HMM) learning. In our method, HMM is used to model temporal variation within segments. We conducted experiments on CK+ database and UNBC-McMaster Shoulder Pain Database. Experimental results on both databases show that our method can not only label the sequences effectively, but also locate apex frames of multi-peak sequences. Besides, the experiments demonstrate that our method outperforms state of the art. I. INTRODUCTION Facial expression recognition has attracted increasing attention due to its wide applications in human-computer interaction. Current research of facial expression recognition can be divided into two categories: frame-based and sequence based. The former usually recognizes expressions from apex expression frame or every frame. The latter analyzes expressions from facial image sequences. Most research of framebased expression recognition requires the pre-identified apex facial images and their corresponding expression labels, or labels for each frame, while sequence based work need the pre-segmented image sequences, usually from neutral images to apex images, and sequence level labels. Since the spontaneous facial expression videos could feature multipeak expressions, manually identifying the apex frame and segmenting expression sequence from the onset frame to the apex frame are very time consuming and prone to error. Compared with expression annotation with pre-identified apex images or pre-segmented expression sequences, weakly labeled data in the form of sequence level ground-truth without providing the onset, duration, and frequency of facial expressions within the video is much easier to annotate, and can be labeled with great accuracy. Thus, in this paper, we formulate sequence-based expression recognition as a multiinstance learning (MIL) problem, in which image sequences is the corresponding author. This paper was supported by the National Natural Science Foundation of China (Gront Nos. 61175037, 61228304, 61473270). are regarded as bags, and image segments of the sequences are viewed as instances in the bags. Multi-instance learning assumes that every positive bag contains at least one positive instance, and every negative bag contains only negative instances. Instead of an exact label for every instance, multiinstance learning algorithms only require bag-level annotations. Through multi-instance learning, we can obtain the temporal segments and their corresponding expression labels simultaneously with only sequence level ground-truth. To the best of our knowledge, there is only one work to recognize expression using multi-instance method. Sikka et al. [1] proposed multiple segment multi-instance learning method for pain recognition and localization. First, facial expression sequences are clustered into several segments. Second, features are extracted from every frame within segments and fused into one single feature vector. Third, MilBoost framework is adopted for multi-instance learning. Experiments were conducted on the UNBC-McMaster Shoulder pain database, demonstrating the advantages of there proposed approach. Although the fused feature vector in [1] captures temporal information to a certain extent, the classifier they used is static, and it hence failed to capture the temporal variation in segment, which is crucial for expression analysis. Therefore, in this paper, we propose to use a HMM to model temporal variation within a segment and develop a novel training process to combine discriminative HMM training with MIL. Although much effort has been made to develop multiinstance learning algorithms, such as diverse density, Citation-kNN, ID3-MI, RIPPER-MI, and BP-MIP [2], only few works consider temporary modeling for multiinstance learning. Yuksel [3] proposed multiple instance hidden Markov model (MI-HMM) for landmine detection. MI-HMM is accomplished via Metropolis-Hastings sampling. Although they introduced HMM in the frame of multi-instance to model the temporary information in timesequence data, Metropolis-Hastings sampling, which is a Bayesian inference, does not guarantee that we can obtain the correct posterior estimate if the class of distributions they consider do not contain the true generator of the data we are observing. In addition, learning HMM parameters numerically via sampling may not guarantee convergence depending on the number of samples collected. Different from [3], we propose to learn HMM analytically through gradient descent that guarantees the convergence. Furthermore, discriminative 978-1-4799-6026-2/15/$31.00 2015 IEEE

learning typically produces better classification results because the learning and testing criteria are the same. Thus, in this paper, we formulate expression recognition as a multi-instance learning problem, and propose a discriminative multi-instance HMM learning algorithm, in which HMMs are adopted to capture the dynamic within instances. Experiments on the Extended Cohn-Kanade (CK+) database [4]and the UNBC-McMaster Shoulder Pain Expression Archive Database [5] demonstrate the effectiveness of our proposed approach. II. MIL-HMM METHOD FOR EXPRESSION A. Method Framework RECOGNITION The framework of our approach is shown in Fig.1, which consists of image sequence segmentation and multi-instance HMM models. First, an image sequencings is clustered into several segments, and the displacements of the facial feature points between current frame and last frame are extracted as the features. After that, discriminative multi-instance HMM is proposed to capture the temporary information in facial segments. We further propose to learn HMM analytically through gradient descent. During testing, we first estimate labels of image segments using the learned HMM, and then infer the label of an image sequence through multi-instance inference. Simultaneously, we obtain the peaks of expression contained in the sequence. The details are described as follows. B. Facial image sequence segmentation Each facial image sequence is cut into several segments, which are regarded as instances in a bag. The normalized cuts (Ncut) algorithm [6] is adopted to cluster an image sequence into several segments. The whole sequence is represented as a weighted undirected graph G (V, E), where V denotes every frame, and E is formed as the links between every pair of nodes. The weight between edges, W, is a function of the similarity between nodes r and s. The object of Ncuts algorithm is to minimize the following function by removing edges between two nodes. Ncut(V 1, V 2 ) cut(v 1, V 2 ) assoc(v 1, V ) + cut(v 1, V 2 ) assoc(v 2, V ) where V 1,V 2 are two disjoint node sets of G, and satisfy the constrains of V 1 V2 V and V 1 V2 ; cut(v 1, V 1 ) is the degree of dissimilarity between V 1,V 2, and can be computed as total weight of the edges that have been removed; assoc(v 1, V ) is the total connection from nodes in V 1 to all nodes in the graph; Similarly, assoc(v 2, V ) denotes the total connection from nodes in V 2 to all nodes in the graph. cut(v 1, V 2 ) W (u, v) u V 1,v V 2 assoc(v 1, V ) u V 1,t V W (u, t) Since segments consist of continuous frames, the weight between frame fi s and frame f i r of sequence i, i.e. W i (r, s), should contain feature similarity as well as the time index of two frames, as shown in Eq.1: ( W i (r, s) exp dist(f (fi r), F (f i s)) ) 2 ( + exp δ f t r t s δ t ) (1) 2 where F (f) indicates the feature vector of frame f, the dist indicates the distance between vectors. Here, we use Euclidean distance. t is the relative time of frame in the sequence. δ f and δ t are the parameters to regularize the weight matrix. For the features, we use facial landmarks provided by the databases. First, the face images are normalized according to the binocular position. Then the displacements of feature points between current frame and last frame are used as the features. C. Multi-instance HMM for expression recognition Let D {(S, Y )} denote facial image sequences data with only sequence level expression labels, where S {S 1, S 2,..., S N } is image sequences set, and Y {Y 1, Y 2,..., Y N } is the corresponding sequence-level expression labels set. Each image sequence (i.e a bag) S i may consists of several segments: X i {X i1, X i2,..., X in }, which are regarded as instances in the bag. y i {y i1, y i2,..., y in } denote the corresponding instance labels. During training, we only have the labels for the whole image sequences. The purpose of multi-instance learning is to maximize conditional log-likelihood (CLL) as shown in Eq.2. where, CLL(Θ) S i,y i 1 + P (Y i 1 S i ) S i,y i 0 log P (Y i 1 S i, Θ) log P (Y i 0 S i, Θ) (2) n max j1 {P (y ij 1 X ij )} (3) and P (Y i 0 S i ) 1 P (Y i 1 S i ), accordingly. Here, P (y ij 1 X ij ) denotes the posterior probability of an instance X ij in S i. In this way, the probability that a bag is positive depends on the maximum probability that an instance in it is positive. Since max function is not differentiable, Noisy-Or (NOR) model is adopted as an approximation, shown in Eq.4 n P (Y i 1 S i ) 1 [1 P (y ij 1 X ij )] (4) j1

Fig. 1. The framework of our approach. First, the raw image sequences were divided in to multiple segments. Then, with facial point data, two HMMs were trained discriminatively under the framework of MIL. HMM0 denotes the Hidden Markov Model trained for negative instances and HMM1 denotes HMM for positive instances. With the trained two HMMs, label of every segment can be estimated. Image sequences label can be evaluated through the assumption of Multiple Instance. Here, HMM is adopted as the instance classifier P (y ij X ij ). As a generative model, HMM provides the joint probability of all observed variables and hidden variables. Thus, the joint probability of observed variables is shown in Eq.5: P (X ij Θ) P (π ij ) P (X ij π ij ) π ij T 1 T P (π 1,ij ) P (π t+1,ij π t,ij ) P (X t,ij π t,ij ) π ij (5) where X ij {X 1,ij, X 2,ij,..., X T 1,ij, X T,ij } is the observed variables of the jth instance in ith bag, π ij {x 1,ij, π 2,ij,..., π T 1,ij, π T,ij } express the hidden variables, and Θ are the parameters. For HMM, the local conditional probability distribution is parameterized with conditional Gaussian P (X t,ij π t,ij k) N(µ ijk, Σ ijk ). To ensure the covariance matrix be positive semidefinite, Σ ijk is parameterized as A Σ 1 2 ijk [7]. For the transition probability, θ a b P (π t+1,ij a π t,ij b), we introduce a transformed parameter β a b to incorporate the constraints θ a b 0 and a θ a b 1, where θ a b eβ a b. a eβ a b In order to maximize the conditional log-likelihood as shown in Eq.2, we should transfer the log-likelihood of HMM to conditional log-likelihood accord to Bayesian theorem: P (X ij y ij c)p (y ij c) P (y ij c X ij ) y ij c P (X ij y ij c )P (y ij c ) (6) where c {0, 1} and c {0, 1}. Thus, we train HMM discriminatively for the multiinstance learning using gradient descent as shown in Eq.7 LL(Θ) n S i,y i 1 j1 log P (yij 1 Xij) P (Y i 0 S i ) P (Y i 1 S i ) P (y ij 1 X ij ) P (y ij 0 X ij ) + n S i,y i 0 j1 log P (y ij 0 X ij) (7) The right hand side of Eq.7, the derivative, log P (y ij 1 X ij ) and log P (y ij0 X ij ), can be translated into log P (X ij y ij 1) and log P (X ij y ij 0) by applying Bayesian theorem as Eq.6 [7], and can be expressed as follows: [1 P (y ij c X ij)]. log P (X ij y ij c) c c c log P (y ij c X ij ) (8) c P (y ij c X ij) c c. log P (X ij y ij c ) c where c {0, 1} and c {0, 1}. Therefor, we obtain the derivative of P (X ij ) to parameters µ ijk, A ijk and transition probability β as shown in Eq.9, Eq.10 and Eq.11 respectively. log P (X ij ) µ ijk log P (X ij ) A ijk T P (π t,ij k X ij )A 2 ijk(x t,ij µ ijk ) (9) T P (π t,ij k X ij ) [ A 1 ijk A ijk (X t,ij µ ijk )(X t,ij µ ijk ) T ] (10) where P (π t,ij k X ij ) can be computed using the forward algorithm for HMM.

Algorithm 1 Parameter Learning input: Train bag set S with corresponding label set Y. Every bag contains multiple instances. output: Optimal parameters of HMMs ˆΘ 1. Parameter initialization: Randomly assign positive or negative label for every instances, and then learn the initialize parameters of HMM through EM algorithm; 2. Compute the parameter derivative according to Eq.7,9,10,11, and adopt BFGS [8] for optimization; 3. Using parameters returned by BFGS as output. TABLE I SEQUENCES DISTRIBUTION OVER EXPRESSION IN CK+ DATABASE Expression Number angry 45 contempt 18 disgust 59 fear 25 happiness 69 sadness 28 surprise 83 Total 327 log P (X ij ) β a b T 1 T 1 P (πt,ij b, πt+1,ij a Xij) [1 P (π t+1,ij a π t,ij b)] a a P (πt,ij b, πt+1,ij b Xij) [ P (π t+1,ij a π t,ij b)] a a (11) The detailed of parameter learning algorithm is listed in Algorithm 1. D. Testing During testing, an image sequence S i is first clustered into several segments (X i1,..., X in ) using Ncut algorithm described in Section II-B After that, the likelihood probability of every segment P (X ij y ij ) is obtained using the learned HMMs according to Eq. 5. Thus, we obtain the expression category of every segments according to Eq.12 y ij {P (X ij y ij 1) > P (X ij y ij 0)} (12) Then, the label of an image sequence is inferred using Eq.3 or Eq.4. III. EXPERIMENTS AND RESULTS A. Experimental conditions We conducted multi-instance expression recognition on two benchmark databases, i.e. the Extended Cohn-Kanade (CK+) database [4] and the UNBC-McMaster Shoulder Pain Expression Archive Database [5]. The CK+ database is a posed expression database, which consists of 327 facial expression image sequences, from neural frame to apex frame. The length of sequences ranges over 5 to 70, and each expression sequence belongs to one of seven expression categories (i.e. angry, contempt, disgust, fear, happiness, sadness, surprise). Table.I shows the distribution of image sequences over seven expressions. For this database, the whole sequence include only one instance. Therefore, we extend the image sequence to simulate multiple instances. For sequence extension, facial expression should change as natural as possible. Specifically, we add several neutral frames before the neutral frame in the sequence. Then, we let the expression after the peak (last frame) vanish naturally by copy the frame before the peak reversely. Thus, we obtain positive bags for each expression. For the negative bag, we generate negative sequences using neutral frame from TABLE II EXPRESSION RECOGNITION RESULT ON CK+ DATABASE USING OUR METHOD Expression Accuracy F1-score angry 100% 1 contempt 98.15% 0.9714 disgust 97.18% 0.9558 fear 97.33% 0.9583 happiness 99.52% 0.9927 sadness 100% 1 surprise 97.59% 0.9652 average 98.54% 0.9776 the corresponding positive sequence by simply duplication. negative sequences contains only neutral frames. The UNBC-McMaster Shoulder Pain database is a spontaneous pain expression database, including 200 image sequences from 25 subjects. The length of the sequences vary from 60 600. The Observed Pain Intensity (OPI), ranging from 0 (no-pain observed) to 5 (extreme pain observed), are assigned to each frame of all sequences by an independent observer. In our experiments, the sequences with OP I 3 are labeled as positive ( pain ) bags, and the sequences with OP I 0 are labeled as negative ( no pain ) bags[1]. Thus, we obtain 57 positive bags and 92 negative bags. Both databases provide feature points for each frame, which are used as features in our experiments. Leave-onesubject-out cross validation method is adopted. B. Experimental Result and Analysis On CK+ Database The expression recognition results on CK+ database are shown in Table.II. From Table.II, we can find that the average accuracy is 98.54%. Furthermore, the performance of anger and sadness recognition reaches 100%. This demon- TABLE III EXPERIMENT RESULT ON UNBC-MCMASTER SHOULDER PAIN DATABASE Classified as positive negative Confusion positive 39 18 Matrix negative 4 88 Accuracy 85.23% F1-score 0.78

(a) Fear expression Fig. 2. Expression localization examples on CK+ database (b) Happy expression TABLE IV COMPARISON BETWEEN OUR METHOD AND OTHER PAIN DETECTION ALGORITHMS Algorithm Accuracy # subjects # sequences MIL-HMM 85.23% 25 149 MS-MIL 83.7% 23 147 (K. Sikka (et.al) [1]) P. Lucey et.al [9] 80.99% 20 142 A.B. Ashraf et.al [10] 68.31% 20 142 A.B. Ashraf et.al [10] 81.21% 21 84 strates the effectiveness of our proposed approach. In addition to expression recognition, our method can locate the expression segments contained in the positive image sequences. The posterior probability of segments indicate the score of the segments being positive. Since the segmentation process clusters similar frames together, the score of each frame of one segment is the same as that of the segment. Fig.2 shows two samples obtained by our multi-instance HMM. According to the criterion we used for image sequence extension of the CK+ database, each positive sequence has only one positive instance in the middle. From these two samples, we can find that the inferred scores are consistent with the ground truth. C. Experimental Result and Analysis On UNBC-McMaster Shoulder pain database Expression recognition results on the UNBC-McMaster Shoulder pain database are shown in Table.III. From the table, we can find that our approach achieves 85.23% of accuracy and 0.78 of F1-score, demonstrating its effectiveness. From the confusion matrix, we may find that it is easier to recognize negative sequences than positive ones. This is probably due to the unbalanced data. Similar as CK+ database, two examples are shown in Fig.3 to demonstrate the expression localization capability of our method. The green line is the normalized PSPI value, which is a frame level ground truth. From Fig.3(a), we can find that our method can locate multi-peak pain segments, such as frames around 80 and 280. From Fig.3(b), we find that the posterior probability we predicted can locate some ambiguity segments correctly, such as frames around 175 and 450. D. Comparison with related work Till now, there is only one work that recognized expression using multi-instance learning, and conducted experiments on the the UNBC-McMaster Shoulder pain database. Therefore, we only compare our work with related work on the UNBC- McMaster Shoulder pain database as shown in Table.IV. From the table, we can find that both our approach and Sikka et.al s outperform [9] and [10]. These two works formulated pain recognition as a traditional machine learning problem, and used frame-based expression recognition method. support vector machines (SVM) are adopted as classifier to assign labels for each frame. It demonstrate the superior of multi-instance learning for facial expression recognition with weakly labeled data. Compared with Sikka et.al s approach, our method achieves better performance. Although Sikka et.al also formulats expression recognition as a multi-instance learning problem, the classifier they used is static, which can not fully capture the dynamic embedded in expression sequences. While our approach adopts HMM to model the temporary variation within segments. IV. CONCLUSION In this paper, a method for expression recognition using only weakly labeled image sequencing data by applying MIL frame work in combination with Discriminative HMM learning approach is proposed. We first demonstrate that

(a) (b) Fig. 3. Expression localization examples on UNBC-McMaster Shoulder pain database how multi-instance be created. We divide facial every image sequence into several segments using Ncut clustering algorithm. In this way, image sequences can be regarded as bags and segments in the bags are viewed as instances. Then, the detail of MIL-HMM method for expression recognition is presented. We trained a classifier under MIL framework for expression recognition. In order to better capture temporal variation contained in segments, HMM is applied. During training, we learn the parameters by combining multi-instance learning and discriminative HMM training algorithm. We experiment on CK+ database and UNBC- McMaster Shoulder Pain Database. Experimental results on both databases show that our method can not only label the sequences effectively, but also locate apex frames of multipeak sequences. Further more, the results demonstrate that our method outperforms both traditional machine learning and other multi-instance learning based methods. [6] Shi, J., Malik, J.: Normalized cuts and image segmentation. Pattern Analysis and Machine Intelligence, IEEE Transactions on 22 (2000) 888 905 [7] Wang, X., Ji, Q.: Learning dynamic bayesian network discriminatively for human activity recognition. In: Pattern Recognition (ICPR), 2012 21st International Conference on, IEEE (2012) 3553 3556 [8] Avriel, M.: Nonlinear programming: analysis and methods. Courier Dover Publications (2003) [9] Lucey, P., Howlett, J., Cohn, J.F., Lucey, S., Sridharan, S., Ambadar, Z.: Improving pain recognition through better utilisation of temporal information. In: AVSP. (2008) 167 172 [10] Ashraf, A.B., Lucey, S., Cohn, J.F., Chen, T., Ambadar, Z., Prkachin, K.M., Solomon, P.E.: The painful face pain expression recognition using active appearance models. Image and vision computing 27 (2009) 1788 1796 REFERENCES [1] Sikka, K., Dhall, A., Bartlett, M.: Weakly supervised pain localization using multiple instance learning. In: Automatic Face and Gesture Recognition (FG), 2013 10th IEEE International Conference and Workshops on, IEEE (2013) 1 8 [2] Babenko, B.: Multiple instance learning: algorithms and applications. View Article PubMed/NCBI Google Scholar (2008) [3] Yuksel, S.E., Bolton, J., Gader, P.D.: Landmine detection with multiple instance hidden markov models. In: Machine Learning for Signal Processing (MLSP), 2012 IEEE International Workshop on, IEEE (2012) 1 6 [4] Lucey, P., Cohn, J.F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The extended cohn-kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In: Computer Vision and Pattern Recognition Workshops (CVPRW), 2010 IEEE Computer Society Conference on, IEEE (2010) 94 101 [5] Lucey, P., Cohn, J.F., Prkachin, K.M., Solomon, P.E., Matthews, I.: Painful data: The unbc-mcmaster shoulder pain expression archive database. In: Automatic Face & Gesture Recognition and Workshops (FG 2011), 2011 IEEE International Conference on, IEEE (2011) 57 64