Algorithm Optimizations for Low-Complexity Eye Tracking

Similar documents
Eye Detection by Haar wavelets and cascaded Support Vector Machine

A Hybrid Face Detection System using combination of Appearance-based and Feature-based methods

Computationally Efficient Serial Combination of Rotation-invariant and Rotation Compensating Iris Recognition Algorithms

Dynamic skin detection in color images for sign language recognition

Robust Real-Time Eye Detection and Tracking Under Variable Lighting Conditions and Various Face Orientations

Task analysis based on observing hands and objects by vision

Adaptive Skin Color Classifier for Face Outline Models

A Real-Time Hand Gesture Recognition for Dynamic Applications

A High Speed Face Measurement System

Vehicle Detection Method using Haar-like Feature on Real Time System

MULTI ORIENTATION PERFORMANCE OF FEATURE EXTRACTION FOR HUMAN HEAD RECOGNITION

Eye tracking by image processing for helping disabled people. Alireza Rahimpour

A REAL-TIME FACIAL FEATURE BASED HEAD TRACKER

Component-based Face Recognition with 3D Morphable Models

Last week. Multi-Frame Structure from Motion: Multi-View Stereo. Unknown camera viewpoints

Introduction to Medical Imaging (5XSA0) Module 5

Face Tracking : An implementation of the Kanade-Lucas-Tomasi Tracking algorithm

A Novel Smoke Detection Method Using Support Vector Machine

OCR For Handwritten Marathi Script

Video-Rate Hair Tracking System Using Kinect

Robust color segmentation algorithms in illumination variation conditions

A Method of Annotation Extraction from Paper Documents Using Alignment Based on Local Arrangements of Feature Points

A new approach to reference point location in fingerprint recognition

Mouse Pointer Tracking with Eyes

MOVING OBJECT DETECTION USING BACKGROUND SUBTRACTION ALGORITHM USING SIMULINK

Learning based face hallucination techniques: A survey

Keywords Binary Linked Object, Binary silhouette, Fingertip Detection, Hand Gesture Recognition, k-nn algorithm.

Progress Report of Final Year Project

Face Detection and Recognition in an Image Sequence using Eigenedginess

Effective Features of Remote Sensing Image Classification Using Interactive Adaptive Thresholding Method

Static Gesture Recognition with Restricted Boltzmann Machines

C E N T E R A T H O U S T O N S C H O O L of H E A L T H I N F O R M A T I O N S C I E N C E S. Image Operations I

Face Recognition Using Long Haar-like Filters

An Efficient Character Segmentation Based on VNP Algorithm

Face Recognition for Mobile Devices

Critique: Efficient Iris Recognition by Characterizing Key Local Variations

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation

Comparative Study of Hand Gesture Recognition Techniques

Implementation of a Face Recognition System for Interactive TV Control System

Component-based Face Recognition with 3D Morphable Models

Segmentation and Tracking of Partial Planar Templates

EE795: Computer Vision and Intelligent Systems

Computer and Machine Vision

Selection of Location, Frequency and Orientation Parameters of 2D Gabor Wavelets for Face Recognition

Available online at ScienceDirect. Procedia Computer Science 22 (2013 )

Chapter 3 Image Registration. Chapter 3 Image Registration

Video Processing for Judicial Applications

Figure 1 shows unstructured data when plotted on the co-ordinate axis

Computers and Mathematics with Applications. An embedded system for real-time facial expression recognition based on the extension theory

Reduced Image Noise on Shape Recognition Using Singular Value Decomposition for Pick and Place Robotic Systems

Non-rigid body Object Tracking using Fuzzy Neural System based on Multiple ROIs and Adaptive Motion Frame Method

Eye Tracking System to Detect Driver Drowsiness

A Supporting System for Beginners of Billiards Based on Predicting the Path of the Balls by Detecting the Cue Stick

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

Haresh D. Chande #, Zankhana H. Shah *

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

Face Tracking in Video

Motion Estimation for Video Coding Standards

Outline 7/2/201011/6/

COMPARATIVE ANALYSIS OF EYE DETECTION AND TRACKING ALGORITHMS FOR SURVEILLANCE

Face Quality Assessment System in Video Sequences

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Topic 4 Image Segmentation

Subject-Oriented Image Classification based on Face Detection and Recognition

Research Fellow, Korea Institute of Civil Engineering and Building Technology, Korea (*corresponding author) 2

Module 7 VIDEO CODING AND MOTION ESTIMATION

Facial Expression Recognition Based on Local Directional Pattern Using SVM Decision-level Fusion

Generic Face Alignment Using an Improved Active Shape Model

Face Recognition Technology Based On Image Processing Chen Xin, Yajuan Li, Zhimin Tian

CHAPTER 5 MOTION DETECTION AND ANALYSIS

Color Content Based Image Classification

ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries

Eye-blink Detection Using Gradient Orientations

Fast and Efficient Automated Iris Segmentation by Region Growing

Toward Part-based Document Image Decoding

An ICA based Approach for Complex Color Scene Text Binarization

Gesture Recognition using Temporal Templates with disparity information

Automatic Tracking of Moving Objects in Video for Surveillance Applications

AN EXAMINING FACE RECOGNITION BY LOCAL DIRECTIONAL NUMBER PATTERN (Image Processing)

Region Segmentation for Facial Image Compression

Human Motion Detection and Tracking for Video Surveillance

Occluded Facial Expression Tracking

A Study on Similarity Computations in Template Matching Technique for Identity Verification

A Novel Criterion Function in Feature Evaluation. Application to the Classification of Corks.

Corner Detection. Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology

Edge Detection and Template Matching Approaches for Human Ear Detection

A STUDY OF FEATURES EXTRACTION ALGORITHMS FOR HUMAN FACE RECOGNITION

Redundancy Resolution by Minimization of Joint Disturbance Torque for Independent Joint Controlled Kinematically Redundant Manipulators

World Journal of Engineering Research and Technology WJERT

An Algorithm for Real-time Stereo Vision Implementation of Head Pose and Gaze Direction Measurement

Automatic Shadow Removal by Illuminance in HSV Color Space

Meticulously Detailed Eye Model and Its Application to Analysis of Facial Image

Detecting Printed and Handwritten Partial Copies of Line Drawings Embedded in Complex Backgrounds

Conspicuous Character Patterns

NIH Public Access Author Manuscript Proc Int Conf Image Proc. Author manuscript; available in PMC 2013 May 03.

MediaTek Video Face Beautify

Motion analysis for broadcast tennis video considering mutual interaction of players

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

Finger Vein Biometric Approach for Personal Identification Using IRT Feature and Gabor Filter Implementation

Transcription:

Proceedings of the 9 IEEE International Conference on Systems, Man, and Cybernetics San Antonio, TX, USA - October 9 Algorithm Optimizations for Low- Eye Tracking Shinji Yamamoto and Vasily G.Moshnyaga Department of Electronics Engineering and Computer Science, Fukuoka University 8-19-1 Nanakuma, Jonan-ku, Fukuoka 814-18, Japan Abstract This paper investigates techniques for reducing computational complexity of tracking e of computer user. By empirical evaluation of a number of techniques we define a techlogy capable of monitoring e of computer user in real time with high accuracy and very low computational overhead. Keywords eye tracking, algorithm, complexity, optimization I. INTRODUCTION A. Motivation With the wide popularity of user centric applications, the role of smart and intelligent devices, capable of monitoring human e is increasing. In addition to traditional applications, such as human-computer interface, security, health care and commercial applications, eye tracking has been applied to new fields, such as of drowsy-driver detection [1], display energy management [2]. Up to date, all above mentioned applications, but the last one, have t been constrained by energy consumption. Therefore research efforts have been focused on delivering accurate eye-tracing in real-time. However, as demands to make the eye-tracking ubiquitous or embedded in portable devices grow, the needs to reduce energy consumption become very important. Although it is easy for a human to detect and track e, performing it on hardware requires complex algorithms and many computations. Because each computation burns energy, lowering the computation complexity is considered to be one of the simplest and effective ways to reduce energy consumption. This paper focuses on minimizing complexity of detecting and tracking e of PC user by a single camera. Based on application features [2], we derive optimizations, capable of reducing complexity of original eye-tracking algorithm by 84% while maintaining the same accuracy. Though we target a particular application, the techniques presented here are general and can be used in other applications. B. Related Research The techniques proposed so far for eye-tracking may be grouped into two categories. The first one makes use of infrared (IR) devices and exploits the reflective properties of pupils [3-4]. Since pupils become bright under IR-illumination, the eye tracking is accomplished by detecting and tracking pupils using the differential IR lighting scheme. The approach is simple and accurate but requires additional resources in the form of multiple IR sources and sensors. The work was sponsored by The Ministry of Education, Culture, Sports, Science and Techlogy of Japan under the Kwledge Cluster Initiative (The Second Stage) and Grant-in-Aid for Scientific Research (C) No.21563 The second approach attempts to track e with the aid of ordinary camera in the absence of any kind of IR devices. The techniques can be broadly classified into three groups: template matching, appearance-based and feature-based. The template-based techniques [5-6] are based on generic eye models and their recursive positioning, translation, rotation, and deformation to best fit the eye representation in the image. While these techniques can detect e accurately, they require good image contrast to converge correctly and are computationally expensive. The appearance-based techniques [7-8] detect e based on their photometric appearance. These techniques usually employ large amount of training data representing the e of different subjects under different face orientations and illumination conditions. These data are used to train a classifier such as neural networks (NN) or support vector machine (SVM) and detection is achieved via classification. Since NN treat each frame independently, many redundant data have to be processed. The feature-based techniques identify e by exploring their distinctive features, such as eye-corner points [9], color distribution of the e [1], edge and intensity of iris [11], intensity variation around the e [12], etc). Amid large variety of features available, the between-the-e point (BTE) [12,13] does t depend on illumination, face occlusion, eye closure and hence is considered as the most stable and robust. When the BTE point is found, the e can be easily located as two small darkest points on each side of the point. To detect the BTE point based on circle-frequency filter [12], however, is very computationally intensive. To accelerate the search, Kawato et al [13] proposed to use a less complex Six-Segment Rectangular (SSR) filter that reflects the bright-dark relations in between-the-e region of the image. However, even with this filter, the algorithm runs over 28Million operations, such as addition or subtraction per frame (64x48 pixels frame size). In the rest of the paper we address complexity reduction of this eye-tracking algorithm. C. Related Research The contribution of this paper is two-fold. First, we empirically analyze the complexity of eye detection and tracking algorithm [13] and present solutions capable of lowering it without affecting the accuracy of results. Second, we propose a modified algorithm that, as experiments show, achieves same accuracy as [13] while reducing the number of computations by 84%. 978-1-4244-2794-9/9/$25. 9 IEEE SMC 9 978-1-4244-2794-9/9/$25. 9 IEEE 18

(x,y) Next frame Compute Integral image Extract face candidates Next face candidate Locate the e Confirm by SVM Positive Save BTE, eye locations Tracking resolution ther candidates? Select SSR filter size Run SSR filter Face candidate? Save location, SSR size Run complete? ther sizes? Fig.1. Flowchart of original eye tracking algorithm 1 2 3 (x1,y1) (x2,y1) 4 5 6 (x1,y2) (x2,y2) Fig.2. Illustration of the SSR filter. The paper is organized as follows. In the next section we discuss original eye-tracking algorithm. Section 3 presents algorithmic optimizations and outlines the modified algorithm. Section 4 summarizes our findings and the future work. II. THE ORIGINAL EYE DETECTION ALGORITHM Before you begin to format your paper, first write and save the Fig. 1 shows the original face detection/tracing algorithm [13]. The system captures RGB video sequence and transforms the green component of each image frame i(x,y) into the integral image representation I(x,y) as follows: S(x,y)=S(x,y-1) +i(x,y) (1) I(x,y)=I(x-1,y)+S(x,y), (2) where S(x,y) is the cumulative row sum, S(x,-1)=, I(-1,y)=. The integral image is then scanned (pixel by pixel) by a sixsegment rectangle (Fig.2), computing the sums of pixel values in each segment. For instance, the sum of pixels within rectangle 5 is defined as: Sum(5)=I(x2,y2)+I(x1,y1)-I(x2,y1)- I(x1,y2) The SSR filter then compares the computed sums of the segments as follows: Sum(1) < Sum(2) and Sum(1) < Sum(4) (3) Sum(3) <Sum(2) and Sum(3)<Sum(6) (4) If the above criteria (3)(4) are satisfied, the SSR is considered to be a candidate for Between-The-E (BTE) pattern and two local minimum (i.e. dark) points each are extracted from the regions 1,4 and 3,6 of the SSR for left and right eye candidates. As the eye candidates are selected, the Table 1: Number of computations per frame (x1 3 ) Function Image size 3x24 64x48 Integral image computation 76.2 36.1 Face candidate extraction 5867.6 27869 Eye confirmation by SVM 82.9 147.3 BTE pattern is rmalized by scaling, rotation and histogram equalization and then fed to Support Vector Machine (SVM) for confirmation. Thus for one face candidate, at most 4 patterns are tested. The search is then repeated for 6 other SSR filter sizes (1x72, 8x48, 6x36, 4x24, 3x18 and x12) in order to extract e of smaller faces if any. For the next frame, the BTE location is predicted as x p =2x 1 -x 2, where x p is the predicted position, x 1 and x 2 are its positions in the most and the second most previous frames, respectively. Any incorrect prediction starts the full search. We implemented the algorithm in software and used specific instructions to count the number of computations performed. As results (Table 1) show, the face candidate extraction (see Fig.1, right side) accounts for most of computational load. Because the face/e detection is general ( restriction on the number of faces, face size, motion, and rotation), it implies the following: 1. Full SSR filter scan over the whole image frame. 2. Repeats the scan six times (for all filter sizes). Although this full search is necessary to find multiple faces in an image in general case, it might be redundant in particular applications. Taking into account features of target application is the only way to reduce complexity of the task. III. COMPLEXITY REDUCTION TECHNIQUES A. Assumptions In this study we focus on application [2] that uses eyetracking to monitor gaze of computer user and lower the power consumption of computer display whenever the user detracts his/her e from its screen. In contrast to a general case, we assume that 1. The target object is a single PC user. The user sits in front of PC at a relatively close distance (5-7cm). 2. The camera is located at the top of display. When the user looks at display it faces the camera frontally. 4. The user s motion is slow relatively to the frame rate. 5. The background is stable and constant. Based on the assumptions, we explore several optimizations. B. Filter Size Optimization First we studied the influence of the filter size optimization on the detection rate and complexity. If we consider images of a person, which were taken by same camera from various distances, the interval between his/her e changes with the distance, as shown in Fig.3. Namely, the closer the camera, the larger is the interval. When the distance is about 5-7cm, the interval is about is 45-58 pixels. From experiments, we found for this distance range, the BTE of 55 pixels ensures almost 1% detection rate. 19 SMC 9

% 5cm 1 1 8 6 4 1m pixels 16 1 8 4 6:5 5:4 4:3 3:2 2:1 1:2 2:3 3:4 4:5 5:6 6:5 SSR size ratio (Hight/Width) BTE range 25 3 35 4 45 5 55 6 65 Distance from the camera Fig 3: BTE variation with the user s distance from camera: user images captured from 5 cm and 1m (left); the dependency graph (right) Fig 4: The effect of SSR filter ratio on the results Next, we empirically estimated the most suitable SSR filter size ratio (R=Height/Width) for the selected BTE. We used 18 facial images (92x116 pixels, 256 gray levels) from FERET data base, re-sized when necessary to have the BTE interval in 45-58 pixel range. For each filter size we manually checked whether it covered two e, two eyebrows and cheekbone area, including se. The ratio of the positive cases to the total number of tasted cases represented the detection ratio. Fig.4 shows the results obtained for different values of R. Clearly, the filter ratio (R=2:3) ensures minimal complexity at the highest detection rate. The equations are an exception to the prescribed specifications of this template. You will need to determine whether or t your equation should be typed using either the Times New Roman or the Symbol font (please other font). To create multileveled equations, it may be necessary to treat the equation as a graphic and insert it into the text after your paper is styled. C. Enlarging the SSR displacement In the original algorithm (Fig.1), the SSR filter is scanned over the image in the raster-scan mode pixel by pixel, i.e. with the displacement (D) between two consequent SSR locations of 1 pixel. If we increase the distance (D), the detection rate almost does t change while the number of operations decreases significantly, as shown in Table 2. In this table, the detection rate shows the ratio of the number of times the e were detected in regions 1 and 3 of SSR to the total number of trials. We observe that the computational complexity decreases by a factor of 3 for D=2, and by a factor of 4.5 for D=3 without affecting the detection rate of the original (full scan) algorithm. D. Background Subtraction Background subtraction is a commonly used class of techniques for segmenting out objects of interest. The main 1 1 8 6 4 Table 2: Effect of the SSR displacement on the results SSR displacement in pixels Detection rate (%) Operations (x1 6 ) Reduction factor 1 1 36.13 1 2 1 11.89 3.4 3 1 7.89 4.57 4 94 7.33 4.93 % 1 1 8 6 4 5 1 15 25 3 35 4 45 5 55 6 Threshold Fig 5: Detection accuracy and complexity vs. threshold goal is given a frame sequence (f 1 f 2, f 3,,f n ) from a fixed camera, detect the foreground object with highest accuracy and minimum number of computations. Because in our target application, the object (i.e. the PC user) is t moving fast, the simplest solution to detect foreground is to find the difference between the current frame ft and the previous frame f t+1 that is larger than a threshold (T) [14], i.e. f t - f t-1 >T. Fig. 5 shows variation of the detection rate and complexity rmalized to the full search, empirically evaluated for different threshold values. The tests revealed that when T<, the search area expanded almost over the whole frame because most of pixels were treated as foreground. This large search leads to 1% detection rate but costs many redundant operations. As the threshold increases, the search area shrinks, because fewer pixels belong to foreground. This reduces computational load of the algorithm but lowers the detection rate. For example, at T=6, the algorithm has 26% of original complexity and 77% detection rate. The value of T=25 provides a tradeoff with 4% of complexity reduction and 96% detection rate in comparison to the full search. E. Background Subtraction Skin-color is a well kwn-technique to filter out the nfaces in image frames. To define skin we use the following criteria [15]:.55<R<.85, 1.15<R/G<1.19, 1.15<R/B<1.5,.6<(R+G+B)<1.8. Furthermore, to speed-up the face-area extraction, we use two filters. The first one limits the size of the head in reasonable range. The second one verifies that the face contains a minimum of 25% of skin colored pixels. Fig.6 shows the results of applying the optimization to the results of the background extraction. As one can see, at T=25, the optimization lowers complexity by 23% with the detection rate loss of 7%. F. Frame rate reduction Next we evaluated the influence of the frame rate on detection complexity. Regarding to the frame rate of camera SMC 9

% % 1 1 8 6 4 15 1 95 9 85 8 5 1 15 25 3 35 4 45 5 55 6 Threshold Accuracy Fig 6: Results of joint background extraction and skin-color based search reduction 1 2 3 4 5 6 7 8 9 1 15 25 3 fps Fig 7: Detection accuracy and complexity vs. frame rate 6 5 4 3 1 operations x1 6 Next frame BTE kwn? Set search area (S) around previous BTE Compute integral image of S Extract face candidates Next face candidate Locate the e Confirm by SVM Positive Save BTE, eye locations Set search area (S) by background extraction and skin-color segmentation Run SSR filter Face candidate? Save BTE location Run complete? ther candidates? Fig.8. The modified eye-tracking algorithm readings (3 fps), the movement of PC user is very slow. Hence, tracing the PC user by using the highest frame rate is usually redundant because the lower rate brings the same results with fewer computations. Fig. 7 shows the results obtained for 3min. long user detection at different frame rates (measured in frames per second). As we process less frames per second, both the total number of operations performed per second and the detection accuracy decrease though the number of computations per frame increases due to larger difference between the frames and frequent mis-prediction of eye locations. Clearly, the 1 fps rate ensures minimum computations at maximum accuracy. G. The Modified Algorithm Based on the proposed optimizations, we modified the original eye tracking algorithm, as shown in Fig.8. For initial frame or frames, in which the BTE point is unkwn, the algorithm runs background extraction and skin-color segmentation to reduce thesearch area (S) to face. Else, it searches only a small area (S) of ±8 pixels around the BTE point. For the chosen area (S), the algorithm computes the integral image and scans it by the SSR filter with D=3 to accelerate the BTE search. The eye-location and verification steps are done as in the original algorithm. In comparison to original formulation, our algorithm t only scans the SSR over a smaller image area, but scans it faster. Fig.9 illustrates the search area reduction of our algorithm: dashed line shows the area defined by background extraction; dotted line depicts the area obtained by skin-colorsegmentation; the plain (dark line) shows the area around the previous BTE point; white crosses show the computed locations of e. Fig.9. An illustration of the search area reduction Table 3: Results of evaluation on test sequences test Frames True/False Accuracy %) Or. Mod. Or Mod. CFR (%) 1 151/16 142/9 14/1 94 93 84 2 24/25 217/23 22/2 88 87 87 3 1/1 99/1 1/ 99 1 82 Av. 164/17 153/11 15/1 94 94 84 Additionally to algorithm modification, we propose to run the eye-tracing at low (1fps) frame rate. Fig.1 profiles the amount of computations performed per frame by the modified algorithm (1 fps) and the original one (3fps) in a test, which had the PC user looking at display (camera) and for 123 frames then t looking at the display (camera) at all. The peaks in profile of the original algorithm reflect eye-blinks and miss-predictions. By analyzing the performance on all 151 frames we found that the modified algorithm traced the human e by only 1% less correctly than the original one (94% accuracy), while taking only 16% of computations. Table 3 shows comparisons on 3 different tests. Here, columns marked by Or. and Mod. reflect the original and the modified algorithms, respectively; CRF is complexity reduction factor, calculated as ratio of the total number of 21 SMC 9

Millions computations 3 original modified 25 15 1 5 Fig.1. 4 comparison 6frame 8 per frame 1 1 14 Fig.1. comparison per frame computations taken by the modified algorithm to those taken by the original algorithm; Av. is the average. We see that proposed modifications decrease complexity of the original algorithm by 84% on average without affecting accuracy of eye-tracking. IV. CONCLUSION In this paper we have studied computational savings in eye/face monitoring due to the reduction of the scan area. Combining all the discussed techniques produces an eyetracking system with satisfactory accuracy and significantly better performance than the original one. In this study we have t tacked the computational complexity of the SVM-based decision making. Because the SVM searches a huge space, it is imperative to investigate ways to reduce its complexity of the search. A comparative analysis with other eye-tracing systems will also be conducted in the near future. REFERENCES [1] M.Kutila, Methods for machine vision-based driver monitoring applications, VTT Research, 6, Available from http://www.vtt.fi/publications/index.jsp [2] V.G.Moshnyaga, E.Morikawa, LCD display energy reduction by user monitoring, Proc. IEEE ICCD, 5, pp.94-97. [3] Y.Ebisawa, Improved video-based eye-gaze detection method, IEEE Trans. Instrum. Meas., vol.47,.2, pp.948-955, 1998 [4] C.Morimoto, M.Flickner, Real-time multiple face detection and using active illumination. Proc.IEEE Int. Conf. Automatic Face and Gesture Recognition,. [5] X.Liu, et al., Real-time eye detection and tracking for driver observation under various lighting conditions, IEEE Intelligent Vehicle Symp. 2 [6] K.M.Lam, H.Yuan, Locating and extracting the eye in human face images, Pattern Recognit., vol.27, pp.99-111, 1998 [7] W.Huang and R.Mariani, Face detection and precise eye location, Proc. Int. Conf. Pattern Recognition, [8] M.Reinders, et al, Locating facial features in image sequences using neural networks. Proc. IEEE Int. Conf. Automatic Face and Gesture Recognition, 1997 [9] G.C.Feng, P.C.Yuen, Variance projection function and its application to eye detection for human face recognition, Int. J. Computer Vision, vol.19, pp.899-96, 1998. [1] S.Amarnag, R.S.Kumaran, J.N.Gowdy, Real time eye tracking for human computer interface s, Proc. ICME 3, vol.3, pp.557-56. [11] Y.Tian, T.Kanade, J.F.Coin, Dual-state parametric eye-tracking, Proc. IEEE Int. Conf. Automatic Face and Gesture Recognition,. 22 SMC 9