Estimation of eye and mouth corner point positions in a knowledge based coding system

Similar documents
Segmentation of a Head into Face, Ears, Neck and Hair for Knowledge Based Analysis Synthesis Coding of Videophone Sequences

FACIAL FEATURE EXTRACTION BASED ON THE SMALLEST UNIVALUE SEGMENT ASSIMILATING NUCLEUS (SUSAN) ALGORITHM. Mauricio Hess 1 Geovanni Martinez 2

Speech Driven Synthesis of Talking Head Sequences

Scalable Multiresolution Video Coding using Subband Decomposition

Image-based Fraud Detection in Automatic Teller Machine

Joint Position Estimation for Object Based Analysis Synthesis Coding

Facial Expression Recognition using Principal Component Analysis with Singular Value Decomposition

I a coder concept has been proposed by the CCITT which

A Robust and Efficient Algorithm for Eye Detection on Gray Intensity Face

Dual-state Parametric Eye Tracking

Region Segmentation for Facial Image Compression

On Modeling Variations for Face Authentication

RENDERING AND ANALYSIS OF FACES USING MULTIPLE IMAGES WITH 3D GEOMETRY. Peter Eisert and Jürgen Rurainsky

Robust Face Detection Based on Convolutional Neural Networks

Facial Expression Analysis for Model-Based Coding of Video Sequences

IJSER. Real Time Object Visual Inspection Based On Template Matching Using FPGA

A Study on Similarity Computations in Template Matching Technique for Identity Verification

VIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD. Ertem Tuncel and Levent Onural

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW

Projection Functions for Eye Detection

Facial Features Localisation

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera

Selection of Location, Frequency and Orientation Parameters of 2D Gabor Wavelets for Face Recognition

Face Detection for Skintone Images Using Wavelet and Texture Features

Face Recognition using Principle Component Analysis, Eigenface and Neural Network

Render. Predicted Pose. Pose Estimator. Feature Position. Match Templates

Face Detection and Recognition in an Image Sequence using Eigenedginess

Journal of Industrial Engineering Research

Model-Aided Coding: A New Approach to Incorporate Facial Animation into Motion-Compensated Video Coding

A face recognition system based on local feature analysis

MULTI-VIEW FACE DETECTION AND POSE ESTIMATION EMPLOYING EDGE-BASED FEATURE VECTORS

IDIAP IDIAP. Martigny ffl Valais ffl Suisse

Illumination invariant face recognition and impostor rejection using different MINACE filter algorithms

Computers and Mathematics with Applications. An embedded system for real-time facial expression recognition based on the extension theory

Robust Lip Contour Extraction using Separability of Multi-Dimensional Distributions

Face Recognition based Only on Eyes Information and Local Binary Pattern

Recognition of Non-symmetric Faces Using Principal Component Analysis

variations labeled graphs jets Gabor-wavelet transform

Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera

Detection of Mouth Movements and Its Applications to Cross-Modal Analysis of Planning Meetings

Real Time Object Visual Inspection Based On Template Matching Using FPGA

Image Matching Using Run-Length Feature

Principal Component Analysis and Neural Network Based Face Recognition

A Simple Approach to Facial Expression Recognition

Ultra-Low Bit Rate Facial Coding Hybrid Model Based on Saliency Detection

We present a method to accelerate global illumination computation in pre-rendered animations

Image Inpainting by Hyperbolic Selection of Pixels for Two Dimensional Bicubic Interpolations

Person Re-Identification Using Color Enhancing Feature

FACE DETECTION AND LOCALIZATION USING DATASET OF TINY IMAGES

Dynamic skin detection in color images for sign language recognition

A Laplacian Based Novel Approach to Efficient Text Localization in Grayscale Images

Face Objects Detection in still images using Viola-Jones Algorithm through MATLAB TOOLS

Multi-pass approach to adaptive thresholding based image segmentation

EE368 Project: Visual Code Marker Detection

A Survey of Light Source Detection Methods

Image Processing and Image Representations for Face Recognition

Face Recognition Technology Based On Image Processing Chen Xin, Yajuan Li, Zhimin Tian

IBM Research Report. Inter Mode Selection for H.264/AVC Using Time-Efficient Learning-Theoretic Algorithms

AN APPROACH OF SEMIAUTOMATED ROAD EXTRACTION FROM AERIAL IMAGE BASED ON TEMPLATE MATCHING AND NEURAL NETWORK

Model-based segmentation and recognition from range data

Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement

Text Enhancement with Asymmetric Filter for Video OCR. Datong Chen, Kim Shearer and Hervé Bourlard

Shape Similarity Estimation using Ordinal Measures

[Gaikwad *, 5(11): November 2018] ISSN DOI /zenodo Impact Factor

Fundamental Matrices from Moving Objects Using Line Motion Barcodes

Light source estimation using feature points from specular highlights and cast shadows

Eye Detection by Haar wavelets and cascaded Support Vector Machine

Edge Detection for Dental X-ray Image Segmentation using Neural Network approach

An Algorithm for Real-time Stereo Vision Implementation of Head Pose and Gaze Direction Measurement

Journal of Applied Research and Technology ISSN: Centro de Ciencias Aplicadas y Desarrollo Tecnológico.

Image matching for historical maps comparison

Gender Classification Technique Based on Facial Features using Neural Network

136 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 1, NO. 2, JUNE Augmented Reality with Automatic Illumination Control Incorporating Ellipsoidal Models

Construction of Frontal Face from Side-view Images using Face Mosaicing

Fast Mode Decision for H.264/AVC Using Mode Prediction

A deformable model driven method for handling clothes

Multidirectional 2DPCA Based Face Recognition System

A Robust and Efficient Motion Segmentation Based on Orthogonal Projection Matrix of Shape Space

AUTOMATIC EXTRACTION OF BUILDING ROOFS FROM PICTOMETRY S ORTHOGONAL AND OBLIQUE IMAGES

ADVANCED SECURITY SYSTEM USING FACIAL RECOGNITION Mahesh Karanjkar 1, Shrikrishna Jogdand* 2

LOCALIZATION OF FACIAL REGIONS AND FEATURES IN COLOR IMAGES. Karin Sobottka Ioannis Pitas

Corner Detection. Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology

Spotting Words in Latin, Devanagari and Arabic Scripts

A Robust Facial Feature Point Tracker using Graphical Models

Mapping textures on 3D geometric model using reflectance image

IBM Research Report. Automatic Neutral Face Detection Using Location and Shape Features

Facial Animation System Design based on Image Processing DU Xueyan1, a

CIRCULAR MOIRÉ PATTERNS IN 3D COMPUTER VISION APPLICATIONS

HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION

A Novel Field-source Reverse Transform for Image Structure Representation and Analysis

An Automatic Face Recognition System in the Near Infrared Spectrum

Eyes extraction from facial images using edge density

Eye Detection System using Orientation Histogram

Person Detection Using Image Covariance Descriptor

IMPROVEMENT OF BACKGROUND SUBTRACTION METHOD FOR REAL TIME MOVING OBJECT DETECTION INTRODUCTION

EFFICIENT video coding techniques usually take advantage. Three-Dimensional Motion Estimation of Objects for Video Coding

SCALE INVARIANT TEMPLATE MATCHING

Face Recognition Using K-Means and RBFN

Change detection using joint intensity histogram

Automatic Mouth Localization Using Edge Projection

Transcription:

Estimation of eye and mouth corner point positions in a knowledge based coding system Liang Institut für Theoretische Nachrichtentechnik und Informationsverarbeitung Universität Hannover, Appelstraße 9A, D 067 Hannover, F.R.Germany Phone: ++9 5 76 5, email: zhang@tnt.uni hannover.de Abstract Automatic extraction of facial feature points is one of the main problems for semantic coding of videophone sequences at very low bit rates. In this contribution, an approach for estimation of the eye and mouth corner point positions is presented. For this proposition, the location informations of the face model are exploited to define search areas for estimation of the eye and mouth corner point positions. Then, the eye and mouth corner point positions are estimated based on a template matching technique with eye and mouth corner templates. Finally, in order to verify these estimated corner point positions, some geometric conditions between the corner point positions and the center point positions of the eyes and the mouth are exploited. The proposed algorithm has been applied to test sequences Claire and Miss America with a spatial resolution corresponding to CIF and a frame rate of 0 Hz. Keywords: Facial feature points, eye and mouth corner point positions, face model, template matching.. Introduction For coding of head and shoulder videophone sequences at very low bit rates, the knowledge of the presence of a human face can be exploited by a knowledge based coder. If the knowledge about the facial expressions in the secene is available, it can be exploited by extending a knowledge based coder to a semantic coder, too. For implementation of a semantic coding system, a face model must be automatically adapted not only to a person s face in the sequence but also to its facial expressions. Some algorithms for automatically adapting the face model CANDIDE to the person s face based on the center points of the eyes and the mouth has been proposed,. For the further adaptation of the face model to facial expressions, 8 structural feature points on a person s face in the image have to be automatically estimated first. In this contribution, only estimation of the eye and mouth corner point positions is addressed. Many approaches for automatic extraction of the facial feature points have been proposed. An overview about these approaches can be found in the papers 5,6. Some methods for corner detection in a curve have been developed for computer vision applications 7,8. In the paper 8, corner detection is cast as a problem of cost optimization and the cost function captures different desirable characteristics of corners such as

edginess, curvature and region dissimilarity. In the paper 9, this corner detection approach 8 is used to detect the eye corner points in order to reduce the processing time for the deformable template 0. But, the performance of this corner detection depends on the gray level contrast around the eye. In this contribution, an approach for estimation of the eye and mouth corner point positions is presented. After the face model is tracked frame by frame in a knowledge based coding system, the center point positions of the eyes and the mouth of the face model projected onto the image plane already match those positions of the real person in the image. Therefore, the location informations of the face model are exploited to define search areas for estimation of the eye and mouth corner point positions. Then, the eye and mouth corner point positions are estimated based on a template matching technique with eye and mouth corner templates. Finally, in order to verify these estimated corner point positions, some geometric conditions between the corner point positions and the center point positions of the eyes and the mouth are exploited. The proposed algorithm has been applied to test sequences Claire and Miss America with a spatial resolution corresponding to CIF and a frame rate of 0 Hz. This contribution is organized as follows. Section discusses some geometric conditions of the facial feature points on a human face in the image plane. Section describes estimation of the eye and mouth corner point positions in the image. The experimental results using real image sequences Claire and Miss America are given in Section.. Geometric conditions of the facial feature points This algorithm is developed for the head and shoulder videophone sequences, which facial features, eyes and mouth, are visible on the image plane and the eyes are open. It is assumed, too, that the pupils of the eyes lie near the center positions of the eyes. Hence, the feature points, corner and center point positions of the eyes and the mouth (Fig. ), have the following geometric conditions, p e rl p e lr p e ll p e rr c e r c e l Eye and mouth center point positions c e r c e l c m Eye and mouth corner point positions p m r c m p m l p e rr p e rl p e lr pe ll p m r p m l Figure. Facial feature points on a human face.

. The corner point connecting lines p e ll pe lr, pe rl pe rr and p m l pm r are approximately parallel to the connecting line of both eye center points c e l ce r, respectively,. The length of the left eye p e ll pe lr is equal to that of the right eye pe rl pe rr,. The length c m p m l is equal to that cm p m r, where c e l, ce r and c m stand for the center point positions and p e lr, pe ll, pe rr, p e rl, pm l and pm r stand for the corner point positions, respectively. These geometric conditions are exploited to verify the correctness of the estimated corner point positions of the eyes and the mouth.. Estimation of the eye and mouth corner point positions The algorithm for estimation of the eye and mouth corner point positions consists of following steps: Detection of potential areas for the eye and mouth corner point positions Fig. illustrates the diagram for detection of potential areas for the eye and mouth corner point positions. After the face model is tracked frame by frame, the center point positions of the eyes and the mouth of the face model projected onto the image plane already match those positions of the real person in the image. Therefore, this location informations of the face model are exploited to define search areas for estimation of the eye and mouth corner point positions. After that, in order to further reduce the search areas, the template matching is used. The corner templates used in this contribution were created from the test sequneces Claire, Miss America and Michael. The eye and mouth corners of the person in these sequences were extracted and changed to equal size. Then, the average eye and mouth corners were computed from these eye and mouth corners, respectively. Finally, these average eye and mouth corners are used as the eye and mouth corner templates. The templates are not only roughly adapted to the size of the person s face but also inclined in accordance with the inclination of the head. In each search area, the correlation between the real image and a corresponding corner template is computed in a window centered at (x,y). Points with high values of this correlation are extracted as potential area for the corner point position. The subsequent algorithms for estimation of the eye and mouth corner point positions are applied only to these potential areas which might contain the eye and mouth corner point positions. Estimation of the eye and mouth corner point coordinates For estimation of the corner point coordinates, a probability measure f is evaluated: f c k factor (x, y), () where c k factor (x, y) are the correlations between the real image k and an artificial corner template with a different size factor of 0.8, 0.9,...,.. The point with the highest values of f in each potential area is selected as D corner point coordinates.

real image global search areas eye corner template mouth corner template template matching adapted eye corner template points with the high correlation adapted mouth corner template potential areas for corner point positions Figure. Detection of potential areas for the corner point positions.

Verification of the correctness of the estimated corner point coordinates After the eye and mouth corner point coordinates have been estimated, their correctness has to be verified. According to the geometric conditions in Section, the following criteria are used for this goal.. The inclinations of the corner point connecting lines p e ll pe lr, pe rl pe rr, p m l pm r and the inclination of the eye center point connecting line c e l ce r must respectively satisfy p e ll pe lr c e th l ce r, p e rl pe rr c e l ce r th, p m l pm r c e th l ce r, where inclination is defined as arctan Y l Y r X l X r.. Lengths of both eyes must satisfy L p L e rl pe rr p e ll lr pe L p e rl pe rr L p e ll pe lr th.. A half of the mouth length must satisfy L p m r c m L p m l cm L p m r c m L p m l cm th. The thresholds th, th, th and th are considered constant during the sequence. If the above conditions are satisfied, these estimated coordinates are taken as the eye and mouth corner point coordinates, respectively.. Experimental results The described algorithm was applied to the test sequences Claire and Miss America with a spatial resolution corresponding to CIF and a frame rate of 0 Hz. In order to measure the accuracy of the eye and mouth corner point estimates, the maximum position errors feye max corner,n, fmouth max corner,n over the frame n and their average position errors f eye corner, f mouth corner, f eye corner f mouth corner N eye n N mouth n f max eye corner,n, () f max mouth corner,n, () f max eye corner,n max X ll,n X t ll,n, Y ll,n Y t ll,n, X lr,n X t lr,n, Y lr,n Y t lr,n, 5

X rl,n X t rl,n, Y rl,n Y t rl,n, X rr,n X t rr,n, Y rr,n Y t rr,n, () f max mouth corner,n max X ml,n X t ml,n, Y ml,n Y t ml,n, X mr,n X t mr,n, Y mr,n Y t mr,n, (5) are introduced, where (X t, ll,n Yt ll,n ), (Xt lr,n, Yt lr,n ), (Xt rl,n, Yt rl,n ), (Xt rr,n, Y t rr,n), (X t ml,n, Yt ml,n )and (Xt mr,n, Y t mr,n) are the true image corner coordinates of the eyes and the mouth at the image n (manually determined), (X ll,n, Y ll,n), (X lr,n, Y lr,n), (X rl,n, Y rl,n), (X rr,n, Y rr,n), (X ml,n, Y ml,n) and (X mr,n, Y mr,n) are the estimated corner coordinates of the eyes and the mouth at the image n. N eye is the number of images in which the eye corner coordinates are estimated. N mouth is the number of images in which the mouth corner coordinates are estimated. In this experiment, the threshold th, th, th and th are 5 0, 5 0, 0.5 and 0., respectively. For the test sequences Claire (CIF, 0 Hz), the face model is adapted to the person s face in the 6th frame. 7% positions of the eye corner points and 77% positions of the mouth corner points are successfully estimated. Fig. shows the maximum position errors feye max corner,n and fmouth max corner,n over the frame n. Their average position errors f eye corner and f mouth corner are. and. pel, respectively. Fig. shows the estimated eye and mouth corner point positions in the 6 th frame. feye max corner,n pel feye max corner,n pel 5 5 f eye corner. pel f mouth corner. pel 0 0 0 0 0 0 50 frame n (a) Maximum position error for eye corner 0 0 0 0 0 0 50 frame n (b) Maximum position error for mouth corner Figure. Maximum position errors for the eye and mouth corner positions for test sequence Claire (CIF, 0Hz). For the test sequences Miss America (CIF, 0 Hz), the face model is adapted to the person s face in the second frame. 69% positions of the eye corner points and 9% positions of the mouth corner points are successfully estimated. Fig.5 shows the maximum position errors feye max corner,n and fmouth max corner,n over the frame n. Their average position errors f eye corner and f mouth corner are.5 and.09, respectively. Fig.6 shows the estimated eye and mouth corner point positions in the 7 th frame. 6

5 f max eye corner,n pel 5 f max eye corner,n pel f eye corner.5 pel f mouth corner.09 pel 0 0 0 0 0 0 50 frame n (a) Maximum position error for eye corner 0 0 0 0 0 0 50 frame n (b) Maximum position error for mouth corner Figure 5. Maximum position errors for the eye and mouth corner positions for test sequence Miss America (CIF, 0Hz). Figure. The estimated eye and mouth corner positions in the 6 th frame of the test sequence Claire (CIF, 0Hz). Figure 6. The estimated eye and mouth corner positions in the 7 th frame of the test sequence Miss America (CIF, 0Hz). 7

5. Conclusion In this contribution, an algorithm for estimation of the eye and mouth corner point coordinates has been proposed. After the face model is tracked frame by frame in a knowledge based coding system, the eye and mouth corner point coordinates of a person in the image are estimated. For both test sequences Claire and Miss America, the average position errors for the eye corner and the mouth corner are found to be. and. pel, respectively. References [] M. Kampmann, J. Ostermann, Automatic adaptation of a face model in a layered coder with an object based analysis synthesis layer and a knowledge based layer, accepted for publication in Signal Processing: Image Communication, Special Issue on Communicating Natural Scenes by Image Synthesis. [] H. Musmann, A layered coding system for very low bit rate video coding, Signal Processing: Image Communication, Vol. 7, Nos. 6, November 995, pp. 67 78. [] K. Aizawa, H. Harashima and T. Saito, Model based analysis synthesis image coding (MBASIC) system for a person s face, Signal Processing: Image Communication, Vol., No., October 989, pp.9 5. [] R. Koch, Adaptation of a D facial mask to human faces in videophone sequences using model based image analysis, Picture Coding Symposium (PCS 9), Tokyo, Japan, pp. 85 88, September 99. [5] A. Samal and P. Iyengar, Automatic Recognition and analysis of human faces and facial expressions: A Survey, Pattern Recognition, Vol. 5, No., January 99, pp.65 77. [6] R. Chellappa, C.L. Wilson and S. Sirohey, Human and Machine Recognition of Faces: A Survey, Proceedings of IEEE, Vol. 8, No. 5, May 995, pp. 705 70. [7] C. Teh and R. Chin, On the detection of dominant points on digital curves, IEEE Trans. Pattern Anal. Mach. Comput. Vol., No. 8, August 989, pp. 859 87. [8] X. Xie, R. Sudhakar and H. Zhuang, Corner Detection by a cost minimization approach, Pattern Recognition, Vol. 6, No. 8, August 99, pp. 5. [9] Kin Man Lam and H. Yan, Locating and extracting the eye in human face images, Pattern Recognition, Vol. 9, No. 5, May, 996, pp. 77 779. [0] A. Yuille, P. Hallinan and D. Cohen, Feature extraction from faces using deformable templates, International Journal of Computer Vision, Vol. 8, No., 99, pp. 99. [] L., Tracking a face in a knowledge based analysis synthesis coder, International Workshop on Coding Techniques for Very Low Bit Rate Video (VLBV 95), Tokyo, Japan, paper A 6, November 995. 8