C.R VIMALCHAND ABSTRACT

Similar documents
An Approach for Face Recognition System Using Convolutional Neural Network and Extracted Geometric Features

Classification of Face Images for Gender, Age, Facial Expression, and Identity 1

Classification of Upper and Lower Face Action Units and Facial Expressions using Hybrid Tracking System and Probabilistic Neural Networks

FACIAL EXPRESSION RECOGNITION USING ARTIFICIAL NEURAL NETWORKS

A Facial Expression Classification using Histogram Based Method

Facial Expression Recognition Using Non-negative Matrix Factorization

Robust Facial Expression Classification Using Shape and Appearance Features

Emotion Detection System using Facial Action Coding System

Human Face Classification using Genetic Algorithm

Facial Expression Recognition using Principal Component Analysis with Singular Value Decomposition

A Simple Approach to Facial Expression Recognition

Facial expression recognition using shape and texture information

Emotion Recognition With Facial Expressions Classification From Geometric Facial Features

Automatic Detecting Neutral Face for Face Authentication and Facial Expression Analysis

IBM Research Report. Automatic Neutral Face Detection Using Location and Shape Features

Automatic Facial Expression Recognition based on the Salient Facial Patches

Evaluation of Gabor-Wavelet-Based Facial Action Unit Recognition in Image Sequences of Increasing Complexity

Proceedings of the 14th International Conference on Image Analysis and Processing (ICIAP 07) Year of publication:

Visual object classification by sparse convolutional neural networks

Facial Expression Detection Using Implemented (PCA) Algorithm

Machine Learning 13. week

Gender Classification Technique Based on Facial Features using Neural Network

Cross-pose Facial Expression Recognition

Face Recognition Based On Granular Computing Approach and Hybrid Spatial Features

Face Detection and Recognition in an Image Sequence using Eigenedginess

Automatic Facial Expression Recognition Based on Hybrid Approach

Facial Expression Recognition Based on Local Directional Pattern Using SVM Decision-level Fusion

Facial Feature Extraction Based On FPD and GLCM Algorithms

Facial Expression Recognition

ROBUST FACE ALIGNMENT USING CONVOLUTIONAL NEURAL NETWORKS

A HYBRID APPROACH BASED ON PCA AND LBP FOR FACIAL EXPRESSION ANALYSIS

Digital Vision Face recognition

3D Facial Action Units Recognition for Emotional Expression

Facial expression recognition is a key element in human communication.

Facial Emotion Recognition using Eye

COMPUTER AND ROBOT VISION

DA Progress report 2 Multi-view facial expression. classification Nikolas Hesse

Recognition of facial expressions in presence of partial occlusion

HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION

Dynamic Facial Expression Recognition Using A Bayesian Temporal Manifold Model

Face Detection Using Convolutional Neural Networks and Gabor Filters

OBJECT detection in general has many applications

Analysis of Image and Video Using Color, Texture and Shape Features for Object Identification

Action Unit Based Facial Expression Recognition Using Deep Learning

Face Detection by Means of Skin Detection

Meticulously Detailed Eye Model and Its Application to Analysis of Facial Image

A Real Time Facial Expression Classification System Using Local Binary Patterns

Age Invariant Face Recognition Aman Jain & Nikhil Rasiwasia Under the Guidance of Prof R M K Sinha EE372 - Computer Vision and Document Processing

Real time facial expression recognition from image sequences using Support Vector Machines

A Hierarchical Face Identification System Based on Facial Components

MULTI ORIENTATION PERFORMANCE OF FEATURE EXTRACTION FOR HUMAN HEAD RECOGNITION

NOVATEUR PUBLICATIONS INTERNATIONAL JOURNAL OF INNOVATIONS IN ENGINEERING RESEARCH AND TECHNOLOGY [IJIERT] ISSN: VOLUME 2, ISSUE 1 JAN-2015

Digital Image Processing COSC 6380/4393

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

An Integration of Face detection and Tracking for Video As Well As Images

Artifacts and Textured Region Detection

Criminal Identification System Using Face Detection and Recognition

Facial Animation System Design based on Image Processing DU Xueyan1, a

Estimation of Age Group using Histogram of Oriented gradients and Neural Network

FACIAL MOVEMENT BASED PERSON AUTHENTICATION

Edge Detection for Facial Expression Recognition

Disguised Face Identification Based Gabor Feature and SVM Classifier

CHAPTER 5 GLOBAL AND LOCAL FEATURES FOR FACE RECOGNITION

Research on Emotion Recognition for Facial Expression Images Based on Hidden Markov Model

LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM

Image Processing Pipeline for Facial Expression Recognition under Variable Lighting

Facial expression recognition based on two-step feature histogram optimization Ling Gana, Sisi Sib

Dynamic facial expression recognition using a behavioural model

Evaluation of Face Resolution for Expression Analysis

Face Alignment Under Various Poses and Expressions

A Novel LDA and HMM-based technique for Emotion Recognition from Facial Expressions

CAP 5415 Computer Vision Fall 2012

Hybrid Approach of Facial Expression Recognition

Enhanced Facial Expression Recognition using 2DPCA Principal component Analysis and Gabor Wavelets.

Face Objects Detection in still images using Viola-Jones Algorithm through MATLAB TOOLS

Exploring Bag of Words Architectures in the Facial Expression Domain

Automatic Countenance Recognition Using DCT-PCA Technique of Facial Patches with High Detection Rate

Modelling Human Perception of Facial Expressions by Discrete Choice Models

SPARSE RECONSTRUCTION OF FACIAL EXPRESSIONS WITH LOCALIZED GABOR MOMENTS. André Mourão, Pedro Borges, Nuno Correia, João Magalhães

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

Facial Expression Analysis

Automated Facial Expression Recognition Based on FACS Action Units

Facial Expression Recognition Using Local Binary Patterns

COMPOUND LOCAL BINARY PATTERN (CLBP) FOR PERSON-INDEPENDENT FACIAL EXPRESSION RECOGNITION

Countermeasure for the Protection of Face Recognition Systems Against Mask Attacks

Model Based Analysis of Face Images for Facial Feature Extraction

FACIAL EXPRESSION USING 3D ANIMATION

Recognition of Facial Action Units with Action Unit Classifiers and An Association Network

Image Based Feature Extraction Technique For Multiple Face Detection and Recognition in Color Images

Character Recognition Using Convolutional Neural Networks

Journal of Asian Scientific Research FEATURES COMPOSITION FOR PROFICIENT AND REAL TIME RETRIEVAL IN CBIR SYSTEM. Tohid Sedghi

Automatic Facial Expression Recognition Using Features of Salient Facial Patches

Computers and Mathematics with Applications. An embedded system for real-time facial expression recognition based on the extension theory

Classification of objects from Video Data (Group 30)

Recognizing Facial Actions by Combining Geometric Features and Regional Appearance Patterns

Matching Facial Composite Sketches to Police Mug-Shot Images Based on Geometric Features.

Face and Facial Expression Detection Using Viola-Jones and PCA Algorithm

Pupil Localization Algorithm based on Hough Transform and Harris Corner Detection

MULTI-VIEW FACE DETECTION AND POSE ESTIMATION EMPLOYING EDGE-BASED FEATURE VECTORS

Age Group Estimation using Face Features Ranjan Jana, Debaleena Datta, Rituparna Saha

Transcription:

International Journal of Scientific & Engineering Research, Volume 5, Issue 3, March-2014 1173 ANALYSIS OF FACE RECOGNITION SYSTEM WITH FACIAL EXPRESSION USING CONVOLUTIONAL NEURAL NETWORK AND EXTRACTED GEOMETRIC FEATURES C.R VIMALCHAND ABSTRACT Faces contain a great deal of valuable information in automatic image processing to extract semantic content for effective analysis and pattern recognition. It requires efficient algorithms that are capable to extract relevant semantic information. Recognizing a person on a photograph tells a lot about the overall content of the picture. The principal aim of facial analysis is to extract valuable information from face images such as position in the image, facial characteristics, facial expressions, the person s gender or identity. Now we consider the most important approaches to facial image analysis and present novel methods based on convolutional neural networks to detect, normalize, and recognize faces and facial features. In this paper the problems of automatic appearance-based facial analysis with machine learning techniques are described and common specific sub-problem like facial feature detection, analyzing geometric features and face recognition which are crucial parts of many applications in the context of indexation, surveillance, access-control or human machine-interaction are handled. To tackle these problems a new technique called Convolutional Neural Network (CNN) is used which is inspired by biological evidence found in the visual cortex of mammalian brains and can be applied to many different classification problems. The proposed system is a CNN- based method for automatic facial feature detection and extraction that employs a hierarchical procedure which first detects the eyes, the nose and the mouth and then refines the result by identifying 10 different facial feature points. The detection rate is 87% to 96% for different databases tolerating an error of 10% of the inter-ocular distance. Finally a novel face recognition system based on CNN architecture learning a non-linear mapping of the image space into a lower dimensional sub-space where the different classes are more easily separable. This approach produces better result for different facial databases in detection and recognition with classical face recognition approaches using PCA or LDA.Back propagation algorithm is used as learning algorithm for CNN which is a Multi-Layer Perceptrons. A Siamese CNN is used for face verification by receiving two face images and comparing to decide if they are same.. Index Terms Human-machine interaction, Convolutional neural network, automatic facial feature detection, Automatic facial feature extraction, inter-ocular distance, Siamese CNN etc. 1. Introduction Machine recognition of faces is becoming important due to its wide range of commercial and law enforcement applications, which include forensic identification, access control, border surveillance and human computer interactions. Many techniques are available which apply face recognition such as stastical projection method, Hidden Markov Models, Support Vector Machines and Neural Networks. The most common and effective approach for face recognition is based on artificial neural networks. Here one particular approach based on convolutional Neural Network is taken for face alignment and facial feature detection which shows to be crucial for many facial analysis applications. A Multi-layered feed forward NN is a powerful machine learning technique as 2014 they can be trained to approximate complex non-linear functions from high dimensional input examples. Standard Multi-layer Perceptrons (MLP) are used in pattern recognition systems to classify signatures from a separate feature extraction algorithm operating on the input data. The problem with this approach is that when the input dimension is high, as in images, the number of connections, thus the number of free parameters also high. So each of the hidden nodes are fully connected to the input layer and another disadvantage of this type of MLP comes that its input layer has fixed size and the input patterns have to be presented well aligned and/or normalized to this input window which makes it a complex task. The Convolutional Neural Networks are an approach that tries to handle these problems by automatically learn local feature extractors; they are invariant to small translations and distortions in

International Journal of Scientific & Engineering Research, Volume 5, Issue 3, March-2014 1174 the input pattern. They implement the principle of weight The pain and happy expressions are important facial sharing which drastically reduces the number of free expressions. Md.Monwar et al use location and shape parameters and thus increase their generalization capacity features to represent the pain information. These features compared to other NN. are used as the inputs for standard back propagation algorithm in the form of three-layer neural network, one 1.1 Convolutional Neural Network model and hidden layer for classification of painful and painless faces. Neocognitron This approach can be adopted for happy expressions. The Facial Action Coding System (FACS) (Ryan, Cohn, Lucey, The first implementation of a CNN was the so-called Saragih, Lucey, Torre, & Rossi, 2009) and (Lucey, Cohn, Neocognitron proposed by Fukushima which has been Lucey, Matthews, Sridharan & Prkachin, 2009) is currently originally applied for the problem of handwritten digit the most widely used method in recognizing facial recognition. The neocognitron makes use of perceptive fields, i.e. each neuron is only connected to a sub-region expressions. FACS encodes the contraction of each facial muscle (stand alone as well as in combination with other corresponding to a certain number of neighboring neurons, muscles) changes the appearance of face. It has been used in the preceding layer. Local connections have been used many times with NNs and they can be used to detect widely for the measurement of shown emotions. elementary visual features in images such as oriented Gizatdinova and Surakka used feature-based method for edges, end points or corners. Lecun s convolutional neural detecting landmarks from facial images (Gizatdinova, & network was trained by Back propagation algorithm and Surakka, 2006). The method was based on extracting applied to the problem of handwritten digit recognition. oriented edges and constructing edge maps at two resolution levels. Edge regions with characteristicedge In the world of computers and recognition it is not pattern formed landmark candidates. Cootes et. al and develop an automatic system to recognize the identification Ratliff et. al developed the Active Appearance Model AAM of faces and faces expression. In natural interpretation, the that shown strong potential in a variety of facial recognition humans can easily detect the expressions the indication of technologies (Ratliff, Patterson, 2008) and (Ratliff, & faces and facial expressions. There are many challenges in Patterson, 2008). face expression recognition such as (i) face detection and segmentation from a captured image (ii) extracting the The first implementation of a CNN was the so-called facial expression information ans (iii) the process of Neocognitron proposed by Fukushima which has been classification of the face expression in emotional state. originally applied to the problem of handwritten digit Generally there are 3 major steps in facial expression recognition. The neocognitron makes use of receptive fields recognition. The first step is to detect the face. The second i.e each neuron is only connected to a sub-region step is to extract the facial expression information (facial corresponding to a certain number of neighboring neurons, features) that convening the facial expression the third step in the preceding layer. This idea has been inspired by the is to classify the facial display conveyed by the face. The discovery of locally-sensitive, orientation-selective neurons features are given as input to CNN for classification. The in the cat s visual system by Hubel and Wiesel. Local usually extracted facial features are geometric features such connections have been used many times with NNs. They as the face of facial components (eyes, mouth etc.) and the can be used to detect elementary visual features in images, locations of facial characteristic points (corners of the eyes, such as oriented edges, end points or corners. As one mouth etc.) or appearance features representing the texture feature extractor can be useful in several parts of the input of the facial skin in specific facial areas including wrinkles, image, the weights are forced to be identical for all possible bulges and furrows. The perceptrons of Siamese CNN are locations of the receptive field, a principal called weight trained to recognize the facial expression of the same sharing. person. An important breakthrough of CNNs came with the 2. Related work widespread use of the Back-propagation learning algorithm for multi-layer feed-forward NNs. LeCun et al. presented 2014

International Journal of Scientific & Engineering Research, Volume 5, Issue 3, March-2014 1175 the first CNN that was trained by Back propagation and position inside the face, size and symmetry to the facial applied it to the problem of handwritten digit recognition. symmetry axis. Second; the facial characteristic points (FCPs) are located in each face component using mouth, Siamese Neural Networks have first been presented by nose, eyes and eyebrows FCPs extraction techniques. Bromley et al. using Time Delay Neural Networks (TDNN) Finally; certain lengths between FCPs are calculated using and applying them to the problem of signature verification, Euclidean distance D: i.e. to verify the authenticity of signatures. This idea was then adopted by Chopra et al. who used Siamese CNNs D= (x1 x2) 2 + (y1 y2) 2 (1) and employed them in the context of face verification. More precisely, the system receives two face images and has to Where (x1, y1) and (x2, y2) are the coordinates of any two decide if they belong to the same person or not. FCPs P1(x1, y1) and P2(x2, y2) respectively. Also two angles in the mouth area are computed to represent with 3. Proposed System geometric lengths the geometric features. This paper propose a system for classifying the six basic 3.1.2 Mouth FCPs Extraction: emotions (anger, disgust, fear, happy, sad, surprise) in addition to the neutral one using two types of features. It generally enters the image that contains the face under check into face detection process to segment the face image. Then the feature extraction process will applied on the face The first process after face detection is the segmentation process. The mouth area chosen by taking the portion of the face image from 25% of the face image width after the left boarder to 25% of the face image width before the right image to produce a feature vector that consists of two types edge and from 66.67% of the face image height after the top of features: geometric features and appearance features boarder to 5% of the face image height before the bottom which represents a pattern for facial expression classes. boarder. The next step is detecting the FCPs inside the Finally this feature vector used as an input into the radial mouth region. This is done by using two transformations of basis function for CNN to recognize the facial expression. the mouth image after applying Spatial sharpening filter The block diagram of the proposed system is shown in to sharpen the mouth image for more clarification of Figure1. mouth details. The first one is transform the mouth image from gray scale into binary image using certain threshold τ 3.1 Facial Features Extraction which chosen as a function of the entire mouth image threshold Τ and factor ϯ as the following equation: Generally the most important step in the field of facial expression recognition is the facial feature extraction which based on finding a set of features that conveying the facial τ = Τ / ϯ (2) expression information. This problem can be viewed as a dimensionality reduction problem (transforming the input data into a reduced representation set of features which encode the relevant information from the input data). This paper use two methods to extract the facial features: Where the value of ϯ determined by manual iterations (empirically). Due to the large diversity in mouth shape, It is recoded that best results of ϯ=3. Binary dilate morphological operation with linear 1X 5 structural element and erode morphological operation with linear geometric features and appearance features. 5 X 1 structural element are applied after the first transformation. 3.1.1 Geometric Features Extraction The second transformation is converting the mouth image In this step 19 features are extracted from the face image. First; the segmentation process is performed to divide the face image into three regions of interest: mouth, nose and two eyes and two eyebrows. This is done by taking into consideration that the detected face is frontalor near frontal and assuming certain geometric constraints such as: into binary using the canny edge detection with also static threshold value which after manual iterations chosen to be 0.5. Also close morphological operation with structural element 5 X 1 is applied after the second transformation. The two results are added together to get the final binary image of the mouth. 2014

International Journal of Scientific & Engineering Research, Volume 5, Issue 3, March-2014 1176 The previous transformations have overcome the problem the left half of the segmented area for it and the right one of mouth shape diversity. In case of angry, disgust, normal, for the other. The next step is to separate each eye-eyebrow sad and surprise classes the first transformation can get the to deal with it singly. This is done by finding a separated mouth region in good way because of the mouth often border between them. The integral projection method is represented by the two lips only unlike in the fear and used to determine the horizontal borders of the line happy classes where the teeth often appears and results to between the lips. The boarder finding algorithm consists of generate many objects after the transformations. So using five steps: canny edge detection is more accurate in the last category of classes. Since the first transformation is robust for the 1. Apply Prewitt filter that emphasizes horizontal edges first category of classes and the second for the second; the by approximating a vertical gradient. proposed approach here is using the both to get an object 2. Compute the vertical integral projection ѕ(x) for n X m which represents the mouth region in the mouth image. eye-eyebrow image by using equation (3): After that the labeling operation is performing then using the blobs analysis to determine the area and Bounding ѕ(x) = m x=1 l(x, y) Box properties of each object in the image. The object with (3) maximum area selected to be the mouth rather than other candidate objects. Using the bounding box property of the Where x {1, 2,.. n} is the row number. mouth object will be easy to determine each of the four 3. Smooth the results using a moving average filter to mouth FCPs as a mean point of none zero values of the minimize the number of peaks around global maxima. corresponding boarder (left, right, top and bottom). 4. Find global maxima 5. If the number of maxima is greater than two peaks; the 3.1.3 Nose FCPs Extraction: border is computed as the average position of the maximum two peaks position. Else it is one peak and then The nose FCPs are mainly used to measure the distance the border will be the position of this peak. from them to eyes centers and mouth FCPs. Also the first step is the segmentation process with the values 40%, 40%, The next step after determining the eye-eyebrow border is 45% and 30% for the left, right, top and bottom boarders to separate the eye and eyebrow using this border. Then the respectively to get the nose image. Unlike the mouth in FCPs of the eye and eyebrow can be determined using the varying their shapes with respect to facial expression transformation to binary with threshold technique. Also classes; the nose hole shape often have the same shape as some image enhancements are applied using Gaussian filter and image histogram equalization before getting the binary circle (hole). So an iteration process is performed to get at image. And also Close morphological operation (with least two holes from the binary image and the maximum Disk structural element in eye image and rectangle two object are selected after labeling and blob analysis structural element in brow image) and labeling operation operations. After that the Centroid property of the two are applied in eye image. Finally, by using blobs analysis holes are determined to locate the corresponding two noses the area property is calculated to choose the object with the FCPs. maximum area that corresponds to the eye. Then locate the eye center from the Centroid property in eye image. 3.2.1.3 Eyes and Eyebrows FCPs Extraction: After that the Bounding Box property is used to determine the object with maximum area that corresponding to the eyebrow. Then left, top and right FCPs in eyebrow image can located. The final step after locating all FCPs is using equation (1) to In eyes and eyebrows the proposed approach is to deal with each eye-eyebrow separately. Also starting with the segmentation process with the values 20%, 20%, 20% and 50% for the left, right, top and bottom boarders respectively is the first step to obtain eyes and eyebrows area. Taking into account the symmetry of the frontal face image the left eye-eyebrow pair is separated from the right one by taking 2014 calculate the Euclidean distance between certain FCPs to be the extracted geometric features. These lengths with the two mouth angles were chosen because they are the most lengths that convey facial expression information.

International Journal of Scientific & Engineering Research, Volume 5, Issue 3, March-2014 1177 3.2.2 Appearance Features Extraction 4. Conclusion The appearance features represent an important part of the facial expression feature. That is because of its holistic nature that deals with the whole face image. As mentioned before many approaches can be used to extract facial expression features such as edge orientation histograms that proposed in this paper. In this major step the normalized image ( 300X300 pixels ) is reduced to (250X200 Pixels) by removing a 50 pixels from left, right and top to focus on the face without the hair. Then the Canny edge detection is applied to get an edge map after some image enhancements (histogram equalization and Gaussian filter). The edge map is divided into zones. The coarsely quantized edge directions are represented as local appearance features and more global appearance features are presented as histograms of local appearance (edge directions). The edge directions are quantized into 4 angular segments. Finally, the face map is represented as feature vector of 64 components (4 histograms of 16 zones). This paper presented an automatic system for facial expression recognition. It addressed the problems of how to detect a human face in static images and how to represent and recognize facial expressions presented in those faces. A hybrid approach is used for facial features extraction as a combination between geometric and appearance facial features. CNN based neural network is used for facial expression recognition. The proposed system can be tested for various databases, while testing it is known that 93.5% accurate in recognition for person dependent dataset. The future work will be the dealing with other types of images like image sequence and 3D images. References 1. J. Bromley, I. Guyon, Y. LeCun, E. S ackinger, and R. Shah. Signature verification using a siamese time delay neural network. International Journal of Pattern Recognition and Artificial Intelligence, 7(4):669 688, 1993. Figure1. Block diagram of the facial expression recognition system Japan, pp3396, April 14-16. INPUT IMAGE 2. Cohn, J. F., Zlochower, A. J., Lien, J.J., & Kanade, T. (1998). Feature-Point Tracking by Optical Flow Discriminates subtle Difference in Facial Expression., Proc. Int'l Conf. Automatic Face and Gesture Recognition, Nara, 3. Edwards, G.J., Cootes, T.F. & Taylor, C.J. (1998). Face recognition using active appearance models. In proceedings of the European conference on computer vision. FACE DETECTION FACE FEATURE EXTRACTION CLASSIFICATION USING CNN ANGER,DISGUST,FEAR,HAPPY NORMAL,SAD,SURPRISE 2014 4. Gizatdinova, Y. & Surakka, V. (2006).Feature-Based Detection of Facial Landmarks from Neutral and Expressive Facial Images, IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 28, no. 1, pp.135-139. 5. Monwar, M. & Rezaei, S. (2006). Pain recognition using artificial neural network, Signal Processing and Information Technology, 2006 IEEE International Symposium on. 6. A. Gepperth. Visual object classification by sparse convolutional neural networks. In Proceedings of the European Symposium on Artificial Neural Networks, 2006. 7. Ratliff, M. S., and Patterson, E. (2008). Emotion recognition using facial expressions with active appearance models. International association of science and technology for development.

International Journal of Scientific & Engineering Research, Volume 5, Issue 3, March-2014 1178 8. S. Duffner and C. Garcia. An online backpropagation algorithm with validation error-based adaptive learning rate. In International Conference on Artificial Neural Networks (ICANN), volume 1, pages 249 258, Porto, Portugal, September 2007. 9. Lucey, P., Cohn, J., Lucey, S., Matthews, I. Sridharan, S. & Prkachin, K. M. (2009). Automatically detecting pain using facial actions, IEEE. 10. Ryan, A. Cohn, J. F., Lucey, S., Saragih, J. Lucey, P. Torre, F. & Rossi, A. (2009). Automated Facial Expression Recognition System, IEEE. 2014