Facial Feature Points Tracking Based on AAM with Optical Flow Constrained Initialization

Similar documents
Occluded Facial Expression Tracking

Image Coding with Active Appearance Models

Increasing the Density of Active Appearance Models

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation

3D Active Appearance Model for Aligning Faces in 2D Images

Kanade Lucas Tomasi Tracking (KLT tracker)

AAM Based Facial Feature Tracking with Kinect

Automatic Construction of Active Appearance Models as an Image Coding Problem

Face Alignment Under Various Poses and Expressions

Notes 9: Optical Flow

Generic Face Alignment Using an Improved Active Shape Model

On the Dimensionality of Deformable Face Models

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Multi-View AAM Fitting and Camera Calibration

Fitting a Single Active Appearance Model Simultaneously to Multiple Images

DISTANCE MAPS: A ROBUST ILLUMINATION PREPROCESSING FOR ACTIVE APPEARANCE MODELS

Sparse Shape Registration for Occluded Facial Feature Localization

Intensity-Depth Face Alignment Using Cascade Shape Regression

The Template Update Problem

Robust Face Alignment Based on Hierarchical Classifier Network

Vehicle Dimensions Estimation Scheme Using AAM on Stereoscopic Video

Visual Tracking (1) Pixel-intensity-based methods

2D vs. 3D Deformable Face Models: Representational Power, Construction, and Real-Time Fitting

Enhanced Active Shape Models with Global Texture Constraints for Image Analysis

In Between 3D Active Appearance Models and 3D Morphable Models

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H.

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

Factorization Method Using Interpolated Feature Tracking via Projective Geometry

TRACKING OF FACIAL FEATURE POINTS BY COMBINING SINGULAR TRACKING RESULTS WITH A 3D ACTIVE SHAPE MODEL

Human pose estimation using Active Shape Models

NIH Public Access Author Manuscript Proc Int Conf Image Proc. Author manuscript; available in PMC 2013 May 03.

3D Morphable Model Parameter Estimation

Evaluating Error Functions for Robust Active Appearance Models

The Lucas & Kanade Algorithm

Active Appearance Models

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

CS 4495 Computer Vision Motion and Optic Flow

Active Wavelet Networks for Face Alignment

Hand Gesture Extraction by Active Shape Models

Face View Synthesis Across Large Angles

Light-Invariant Fitting of Active Appearance Models

Facial Feature Detection

Dynamically Adaptive Tracking of Gestures and Facial Expressions

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Visual Tracking (1) Feature Point Tracking and Block Matching

Rapid 3D Face Modeling using a Frontal Face and a Profile Face for Accurate 2D Pose Synthesis

An Adaptive Eigenshape Model

Computer Vision II Lecture 4

Optical flow and tracking

Face Alignment Using Active Shape Model And Support Vector Machine

Passive driver gaze tracking with active appearance models

SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD

RENDERING AND ANALYSIS OF FACES USING MULTIPLE IMAGES WITH 3D GEOMETRY. Peter Eisert and Jürgen Rurainsky

Lucas-Kanade Scale Invariant Feature Transform for Uncontrolled Viewpoint Face Recognition

Real-time non-rigid driver head tracking for driver mental state estimation

Algorithm research of 3D point cloud registration based on iterative closest point 1

COMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE

Feature Detection and Tracking with Constrained Local Models

Feature Tracking and Optical Flow

Active Appearance Models

Representing Moving Images with Layers. J. Y. Wang and E. H. Adelson MIT Media Lab

Face detection and recognition. Many slides adapted from K. Grauman and D. Lowe

Operation Trajectory Control of Industrial Robots Based on Motion Simulation

Using temporal seeding to constrain the disparity search range in stereo matching

Robust 3D Face Tracking on Unknown Users with Dynamical Active Models

Classification of Face Images for Gender, Age, Facial Expression, and Identity 1

FACE RECOGNITION FROM A SINGLE SAMPLE USING RLOG FILTER AND MANIFOLD ANALYSIS

MULTI-POSE FACE HALLUCINATION VIA NEIGHBOR EMBEDDING FOR FACIAL COMPONENTS. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo

Optimization problems for fast AAM fitting in-the-wild

Head Frontal-View Identification Using Extended LLE

A Bayesian Model for Extracting Facial Features

Occlusion Robust Multi-Camera Face Tracking

Flexible Calibration of a Portable Structured Light System through Surface Plane

Tracking in image sequences

An Automatic Face Identification System Using Flexible Appearance Models

REAL-TIME FACE SWAPPING IN VIDEO SEQUENCES: MAGIC MIRROR

EE795: Computer Vision and Intelligent Systems

Full-Motion Recovery from Multiple Video Cameras Applied to Face Tracking and Recognition

Kanade Lucas Tomasi Tracking (KLT tracker)

Pose Normalization for Robust Face Recognition Based on Statistical Affine Transformation

Face Hallucination Based on Eigentransformation Learning

Motion Estimation (II) Ce Liu Microsoft Research New England

Feature Tracking and Optical Flow

Facial Expression Analysis

Fully Automatic Multi-organ Segmentation based on Multi-boost Learning and Statistical Shape Model Search

Video Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Understanding Faces. Detection, Recognition, and. Transformation of Faces 12/5/17

Computer Vision: Homework 3 Lucas-Kanade Tracking & Background Subtraction

SOME stereo image-matching methods require a user-selected

Peripheral drift illusion

CS201: Computer Vision Introduction to Tracking

Comparison between Motion Analysis and Stereo

Motion Estimation. There are three main types (or applications) of motion estimation:

Visual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania.

Face detection in a video sequence - a temporal approach

Locating Facial Landmarks Using Probabilistic Random Forest

ICA vs. PCA Active Appearance Models: Application to Cardiac MR Segmentation

Ear Recognition based on 2D Images

A Modified Mean Shift Algorithm for Visual Object Tracking

Transcription:

Journal of Pattern Recognition Research 7 (2012) 72-79 Received Oct 24, 2011. Revised Jan 16, 2012. Accepted Mar 2, 2012. Facial Feature Points Tracking Based on AAM with Optical Flow Constrained Initialization Ying Cui Zhong Jin School of Computer Science and Technology Nanjing University of Science and Technology, Nanjing, China cuiying@patternrecognition.cn jinzhong@patternrecognition.cn Abstract A facial feature points tracking method is proposed by adding Lucas-Kanade optical flow constraint on the face alignment algorithm, Active Appearance Model (AAM). The optical flow considers the inter-frame correspondence and uses it to estimate AAM initial shape more accurately with similarity preservation. Experiments show that the proposed method can successfully track the frames that general AAM tracks failed. The method achieves not only the accuracy but also the computation time improvement. Keywords: flow. Feature Point Tracking, Active Appearance Model, Lucas-Kanade optical www.jprr.org 1. Introduction Through years of research, face recognition technology has made great strides and development. With the demand for development of many applications, such as video surveillance, video conferencing, information security and access control, the study based on videos has been one of the most active points in face research areas [1]. At the same time, accurate and quick face alignment is the foundation of face recognition, facial expression recognition and 3D face modeling [2]. Both of them make face tracking and face alignment in video sequences be an important problem that must be solved [3, 4]. Active appearance model (AAM) [5] is a fast and efficient method for face alignment in still images. It can be used directly in video tracking by searching on each subsequent frame with the initialization of the final result of the previous frame [6, 7]. However, this is only suitable for small movement between frames because AAM is sensitive to the initial shape and may easily be stuck in local minima due to its gradient decent optimization. To solve the problem, in this paper, the Lucas-Kanade optical flow [8] is used to improve AAM s efficiency in video face tracking by considering the inter-frame correspondences and a robust shape initialization method is proposed. 2. Active Appearance Model AAM consists of two parts: AAM face modeling and AAM fitting. Face modeling constructs the shape model and the appearance model. AAM fitting is the process of finding the model parameters which best fit a given image. The shape S of an AAM can be described by n feature points in the image: S = (x 1, y 1, x 2, y 2,, x n, y n ) T c 2012 JPRR. All rights reserved. Permissions to make digital or hard copies of all or part of this work for personal or classroom use may be granted by JPRR provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, or to republish, requires a fee and/or special permission from JPRR.

Facial Feature Points Tracking Based on AAM AAM allows linear shape variation. This means that the shape S can be expressed as a base shape S 0 plus a linear combination of k shape vectors {S i }: S(p) = S 0 + k p i S i (1) where (p 1, p 2,, p k ) T are the shape parameters. The base shape S 0 and the k shape vectors {S i } are computed by applying Principal Component Analysis (PCA) to the training shapes (from a hand labeled training set). The base shape S 0 is the mean shape and the shape vectors {S i } are the k eigenvectors corresponding to the k largest eigenvalues. Fig.1(a) shows the mesh of the base shape. The appearance of an AAM is defined within the mesh of the base shape S 0. Let M 0 denote the set of pixels x=(x, y) T that lie inside the mesh of S 0. The appearance of an AAM is then an image A(x) defined over the pixels x M 0. Similar to the shape S, the appearance A(x) can be expressed as a base appearance A 0 (x) plus a linear combination of h appearance images {A i (x)}: A(x) = A 0 (x) + i=1 h λ i A i (x) x M 0 (2) i=1 where the coefficients {λ i } are the appearance parameters. The base appearance A 0 (x) and appearance images {A i (x)} are computed by applying PCA to the shape-normalized training images. Each training image is shape-normalized by warping the hand labeled training shape onto the mesh of S 0. The base appearance A 0 is the mean image and the appearance images {A i } are the h eigenimages corresponding to the h largest eigenvalues. Fig.1(b) shows the base appearance according to the base shape. (a) (b) Fig. 1: The mesh of the base shape (a) and the base appearance (b) To align the face shape on a given image I(x), AAM aims to find the optimal shape parameters p and the optimal appearance parameters λ to minimize the difference between the image I(x) and the model appearance A(x). If x is a pixel in M 0, then the corresponding pixel in the given image I is W(x; p). At pixel x the AAM has the appearance A(x) denoting by (2). At pixel W(x; p) the given image I has the intensity I(W(x; p)). Then we can define the difference as E(p, λ) = x M 0 [ A 0 (x) + h 2 λ i A i (x) I(W(x;p))] (3) i=1 The function W(x; p) is a piecewise affine warp defined by the pair of shapes S 0 and S(p) [9]. 73

Cui & Jin The cost function (3) can be minimized by the inverse compositional algorithm, which is well known as its efficiency [9]. 3. Feature Points Tracking Optical flow is a kind of real-time object tracking algorithm. It can be integrated with other algorithms and used for face tracking and face recognition [10, 11]. In this paper, we integrated the Lucas-Kanade optical flow with AAM for facial feature points tracking. 3.1 Lucas-Kanade Optical Flow Given two neighboring frames I t 1, I t, for a point p = (x, y) T in I t 1, if the optical flow is d = (u, v) T, then the corresponding point in I t is p + d. Lucas-Kanade optical flow aims to find the offset d to minimize the match error between the local appearances of two corresponding points. That is, we can define a cost function upon N(p), the local area of p: e(d) = x N(p) w(x) ( I t (x + d) I t 1 (x) ) 2 (4) where w(x) is the weight function. Optimize (4) and we can get the solution: d = G 1 H (5) where G = x N(p) w(x) I t ( I t ) T, H = x N(p) w(x) I t I), I = I t 1 I t, I t = di t dx. Lucas-Kanade optical flow demands feature points with salient local appearances, or it can not track accurately [12]. In this paper, Lucas-Kanade optical flow is used to track some key points with salient local appearances on the face. The key points are then used for constraining AAM shape initialization. 3.2 Inter-frame Correspondence Human face can be broadly considered as a two-dimensional graphic which meets the Similarity Transformation constraint. ( x t Let the current optical flow tracking set in frame t be F t = f1, x t f2,, ) xt fl yf1 t, yt f2,,, where yt fl l 3, the optical flow tracking set in frame t 1 be F t 1, then the Similarity Transformation between the two frames is: F t = CF t 1 + D (6) ( ) a b where C = is the scaling and rotation matrix, D = b a The detail of computing C and D can see Appendix B of [13]. ( tx t y ) is the translation. 3.3 Similarity Transformation Preservation According to the similarity preservation of the same two frames, given the shape of frame t 1, we can pre-estimate the shape of frame t by the Similarity Transformation parameters C and D. 74

Facial Feature Points Tracking Based on AAM ( x t Let the current feature points set in frame t be S t = 1, x t 2,, ) xt n y1 t, yt 2,,, and the feature yt n points set in frame t 1 be S t 1, then with the similarity preservation, S t can be estimated by parameters C and D: 3.4 AAM Tracking with Optical Flow Constraint The tracking algorithm: Ŝ t = CS t 1 + D (7) 1. Find the key points with salient local appearances for Lucas-Kanade optical flow in the last frame, frame t 1(t = 2), and get F t 1. 2. For frame t, compute the Lucas-Kanade optical flow tracking set F t. 3. Compute the Similarity Transformation parameters C and D. 4. According to the AAM tracking result S t 1 of frame t 1, estimate the current corresponding set Ŝt by parameters C and D. 5. Let Ŝt be the initial shape of AAM, use AAM to track the feature points S t. 6. Let t = t + 1, go to 2). 4. Experiments Firstly, we do experiments on still images to study the impact of initial shape on AAM fitting. Secondly, we do experiments on videos to verify the effectiveness of the proposed tracking method. 4.1 Impact of The Initial Shape The AAM model is constructed on IMM face database. The IMM database consists of 240 images of 40 people, exhibiting variations in pose, expression and lighting. Two sets of images are used for testing, each with 40 images of 40 people. One is full frontal face, and the other is profile face rotated 30 degrees to the person s right. The AAM search is initialized from a range of displacements (from -30 to +30 pixels) both in x and y, and for each displacement the point-to-point errors are evaluated by comparing the result with hand labeled ones. Some results are shown in Fig.2, Fig.3 and Fig.4. The large displacement leads to bad results, especially in profile faces. If the displacement is smaller than 10 pixels, the fitting result is good. If the displacement is larger than 10 pixels, the fitting result may lead to local optimization or even be a mess. 4.2 Face Feature Points Tracking We hand labeled ten images with different poses and expressions for one person to perform the experiment. Each image is labeled with 58 feature points. The proposed method is tested on recorded videos. And the similar experiment is taken with general AAM tracking, by initializing with the final result of the previous frame. In the proposed method, 8 key points with salient local appearances are selected for Lucas-Kanade optical flow tracking, an example is shown in Fig.5. Then the optical flow tracking result is used for AAM shape initialization. Two video frame sizes are tested: size 480 640 and size 240 320. As the test video segments are not hand labeled, we just subjectively judge the tracking is whether successful or not. The judgments are summarized in Table 1; the proposed 75

Cui & Jin -30-20 0 20 30 Fig. 2: Frontal face, the initial shape (top), the result shape (bottom) -30-20 0 20 30 Fig. 3: Profile face, the initial shape (top), the result shape (bottom) 50 Point-to-point error (pixels) 45 40 35 30 25 20 15 10 Frontal face Profile face 5 0-30 -20-10 0 10 20 Initial displacement (pixels) 30 Fig. 4: Point-to-Point errors after search from displaced positions method gives much more accurate, stable and robust results. Fig.6 shows some of the tracking results, the proposed method can successfully track the frames that general AAM tracks failed. The number of iterations of each frame is recorded during the process of 76

Facial Feature Points Tracking Based on AAM tracking. Here we define the convergence rate as the average number of iterations of the successfully tracked frames. And the comparison of convergence rate is shown in Fig.7; the proposed method has a higher convergence efficiency compared with general AAM. Although optical flow computation takes some time, the proposed method is faster than general AAM tracking, owing to its better shape initialization. The average tracking time is shown in the last column of Table 1. Fig. 5: Computed optical flow between frame 83 and 84 Table 1: Subjective Judgment of Tracking Results Frame size Method Number of frames Number of failed frames Rate of failed frames Mean tracking time per frame 480 640 240 320 Proposed method 400 6 1.5% 234.47 ms AAM 400 24 6% 304.11 ms Proposed method 400 6 1.5% 43.23 ms AAM 400 22 5.5% 46.59 ms Frame 80 Frame 103 Frame 293 Fig. 6: Tracking result comparison: general AAM (top), optical constraint AAM (bottom) 77

Cui & Jin 100 90 80 Proposed method General AAM tracking Number of frames 70 60 50 40 30 20 10 0 10 15 20 25 30 35 Number of iterations Fig. 7: Comparison of convergence rate 5. Conclusions and Future Works By adding Lucas-Kanade optical flow object tracking constraint, we have gotten a more robust AAM-based facial feature points tracking method. The inadequacy is that the transformation constraint is only with scale, translation and rotation in plane. In the future we will think about the depth rotation constraint, and make the tracker more robustly track profile views with large angles. The good tracked feature points can be used for virtual mouse controlling, face recognition and 3D face modeling. Acknowledgments This work is partially supported by the Natural Science Foundation of China under grant Nos. 60973098, 61005005 References [1] Yan Yan and Zhang Yu-Jin. State-of-the-Art on Video-Base Face Recognition. Chinese Journal of Computers, 2009, Vol.32, No.5, pp. 878-886. [2] Song Gang, Ai Haizhou and Xu Guangyou. Texture constrained facial feature point tracking. Journal of Software, 2004, 15(11), pp. 1607-1615. [3] Minyoung Kim, Kumar, S., Pavlovic, V.,and Rowley, H.. Face tracking and recognition with visual constraints in real-world videos. Computer Vision and Pattern Recognition (CVPR), 2008 IEEE Conference on, pp. 1-8. [4] Mingcai Zhou, Lin Liang, Jian Sun, and Yangsheng Wang. AAM based face tracking with temporal matching and face segmentation. Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on, pp. 701-708. [5] T. F. Cootes, G. J. Edwards, and C. J. Taylor. Active appearance models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 23(6), pp. 681-685, 2001. [6] Ahlberg J. Real-Time facial feature tracking using an active model with fast image warping. In: Kollias S, Maragos P, eds. Proc. Of the Intl Workshop on Very Low Bitrate Video (VLBV). Athens, 2001, pp. 39-43. (http://www.icg.isy.liu.se/ ahlberg/papers/conf/ VLBV-01.pdf) [7] T.F.Cootes and C.J. Taylor. An Algorithm for Tuning an Active Appearance Model to New Data. British Machine Vision Conference (BMVC), pp. 919-928, 2006. 78

Facial Feature Points Tracking Based on AAM [8] Andres Bruhn, Joachim Weichert and Christoph Schnorr. Lucas/Kanade meets Horn/Schunck: Combining Local and Global Optic Flow Methods. International Journal of Computer Vision, 61(3), pp. 211-231, 2005. [9] I. Matthews and S. Baker. Active appearance models revisited. International Journal of Computer Vision, 60(2), pp.135-164, November 2004. [10] Douglas Decarlo and Dimitris Metaxas. Optical Flow Constraints on Deformable Models with Applications to Face Tracking. International Journal of Computer Vision, vol.38, no.2, pp.99-127, July 2000. [11] Hsieh CK, Lai SH and Chen YC. An optical flow-based approach to robust face recognition under expression variations. IEEE Transactions on Image Processing, 19(1), pp. 233-240, Jan 2010. [12] Krishnan Ramnath, Simon Baker, Iain Matthews and Deva Ramanan. Increasing the density of Active Appearance Models. IEEE Conference on Computer Vision and Pattern Recognition, pp.1-8, 2008. [13] T.F. Cootes and C.J. Taylor. Statistical Models of Appearance for Computer Vision. Technical Report, University of Manchester, 125 pages, 2004. (www.face-rec.org/algorithms/aam/app models.pdf) 79