On the Design and Evaluation of Robust Head Pose for Visual User Interfaces: Algorithms, Databases, and Comparisons

Size: px
Start display at page:

Download "On the Design and Evaluation of Robust Head Pose for Visual User Interfaces: Algorithms, Databases, and Comparisons"

Transcription

1 On the Design and Evaluation of Robust Head Pose for Visual User Interfaces: Algorithms, Databases, and Comparisons Sujitha Martin Shinko Y. Cheng Ashish Tawari Mohan Trivedi Erik Murphy-Chutorian ABSTRACT An important goal in automotive user interface research is to predict a user s reactions and behaviors in a driving environment. The behavior of both drivers and passengers can be studied by analyzing eye gaze, head, hand, and foot movement, upper body posture, etc. In this paper, we focus on estimating head pose, which has been shown to be a good predictor of driver intent and a good proxy for gaze estimation, and provide a valuable head pose database for future comparative studies. Most existing head pose estimation algorithms are still struggling under large spatial head turns. Our method, however, relies on using facial features that are visible even during large spatial head turns to estimate head pose. The method is evaluated on the LISA-P Head Pose database, which has head pose data from onroad daytime and nighttime drivers of varying age, race, and gender; ground truth for head pose is provided using a motion capture system. In special regards to eye gaze estimation for automotive user interface study, the automatic head pose estimation technique presented in this paper can replace previous eye gaze estimation methods that rely on manual data annotation or be used in conjunction with them when necessary. Categories and Subject Descriptors I.4.6 [Image Processing and Computer Vision]: Seg- Dr. Murphy-Chutorian is currently with Google Inc. Dr. Cheng is currently with HRL Laboratories LLC Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. AutomotiveUI 12, October 17-19, Portsmouth, NH, USA. Copyright (c) 2012 ACM /12/10... $15.00 mentation Edge and Feature Detection; I.5.1 [Pattern Recognition]: Models Structural General Terms Algorithms, Experimentation, Performance Keywords Database; Facial features; Head pose 1. INTRODUCTION Safety of the driver and those in the vicinity is highly dependent on the driver s awareness of the changing driving environment. The driver has to be attentive to elements outside of the vehicle such as other vehicles in the surround, traffic lights, and pedestrians, as well as elements inside of the vehicle such as navigation tools, communication modules, and driver assistance systems. In such a visually demanding environment, an efficient Visual User Interface (VUI) for Driver Assistance is crucial. An intelligent VUI can monitor the visual attention of the driver operating the vehicle and can provide assistance when the driver s attention is distracted from the primary task of safe driving. Monitoring a driver s focus of attention can be achieved by analyzing the driver s physical pose and motion; for example, the driver s upper body posture, head pose, foot movement, and eye gaze. In this paper we focus on head pose which plays a vital role in driver intent prediction [8] [4] and serves as a proxy for gaze estimation. While gaze can be estimated accurately with eye-localization, it is more practical to do so using head pose in a driving environment. A major reason for this is that eye-localization requires careful integration and calibration of multiple high-resolution cameras around the driver in order to accommodate for large spatial head turns. Additionally, eye-occluding objects can negatively affect gaze estimation when using eye-localization. As stated in the conclusion of the study of visual distraction on driving performance [29]:...the more severe distractions, that last for a 149

2 Figure 1: A block diagram of our proposed approach. It s a two step approach: extracting features from an image and using corresponding features points on a 3D generic-face model to estimate head pose longer duration or are directed further away from forward, are more likely to involve a head pose component. Given some of the limitations associated with eye-based measures (e.g., eyes sometimes occluded by eyewear), optimizing a headbased measure should become a goal. To this end, we introduce a database containing video sequences of various head poses from on-road driving environments. To the best of our knowledge, this is a unique database with continuous head pose measurements along with manually annotated facial landmarks in a video sequence. This database will help researchers in the automotive field to develop and evaluate methods related to head dynamics and analysis. We additionally present a hybrid approach (Fig. 1) of prominent facial feature detection and using corresponding points on a 3D generic face model to estimate head pose. Evaluations of this method is done using manual feature annotation and automatic feature detection. The remainder of the paper is organized as follows. Section 2 provides studies related to VUI. Section 3 describes the LISA-P Head Pose database. Section 4 provides details on the proposed hybrid head pose approach, followed by its evaluation on the LISA-P Head Pose database in Section 5. Finally, Section 6 concludes the discussion with future direction. 2. RELATED STUDIES In this section, we present studies related to head pose estimation applications and algorithms, and available databases for evaluation of head pose estimation algorithms in order to show the contributions we make with our work. In the automotive field, head pose estimation has been shown to be a good predictor of driver intent and a good proxy for gaze estimation. For the former case, some contributions include driver lane change intent prediction [19] and turn intent prediction [4]. Interestingly, the lane change intent work by McCall et al. was used by Doshi et al. [8] to determine relative usefulness of eye gaze and head dynamics for determining intent. While the lane change intent work uses block matching for head pose estimation, which is further detailed in [15], the turn intent predictive system uses a commercial product. This commercial motion-capture system uses retroreflective markers, which are placed on the subject s head, and a kinematics software to compute head pose. Even though we gain precision in head pose estimation by using markers, many researchers are inclined towards non- 150 intrusive measures in vehicular applications. One such application is a forward collision warning system where a driver s head orientation is determined using shape features and ellipsoidal face model. In another application, robust monocular head pose estimators were developed to infer a driver s focus of attention in a real car environment [22]. Although the method used in [22] works well in the range from 90 to +90 in yaw where 0 refers to the driver looking straight at the camera, it needs an initial estimate of the head position and orientation, which is used to generate a texture-mapped 3-D model of the head. HyHOPE [20] on the other hand has automatic initialization but tracking becomes less reliable as the head approaches ±90 in yaw. These head pose estimation algorithms and others help head gesture analyzers like OHMeGA [17], [18]. The OHMeGA analyzes head gestures as viewed from a frontal facing monocular camera using mostly optical flow and sometimes head pose to remove uncertainty in spatial head fixation locations. Just as head pose estimation plays an important role in driver intent prediction system, it also plays an important role as a proxy for gaze estimation. Many like Valenti et al. [26] emphasize the importance of head pose for gaze estimation in the presence of non-frontal faces and introduce hybrid schemes of combining head pose and eye location for gaze estimation. Other literatures, although not mentioned explicitly, compute eye-gaze estimation with respect to head pose; either manually or automatically. In one application, a framework using eye-gaze estimation along with road scene understanding (i.e. sign and pedestrian detection) has been used to give feedback to the driver when a sign was missing [9]. This system, however, uses an off-the shelf stereo-camera based eye gaze tracker, called facelab, that requires manual calibration to individual drivers. facelab is also used in [10] to develop a road scene monotony detector, which gives context awareness to fatigue detection tools. While some studies use automatic eye-gaze trackers, some studies rely on manual eye gaze annotation with respect to head pose. In [6], Curin et al. studied effects of dictation and short text editing while driving. Road attention evaluation was done using direction and fixation of eye gaze, as annotated manually. Similarly, Christiansen et. al [5] studied driver s reaction to warning systems by monitoring the driver s eyes. In regards to the head pose algorithm discussed in this work, other head pose algorithms using features points are of significant interest. Researches that use feature points to estimate head pose fall into two categories: geometric methods and flexible models. For those papers that use geometric methods, most require 5-6 features points in addition to: known camera parameters [13], face symmetry using lip corners and both eyes [11], parallel lines between eye corners and between lip corners [27], anatomical structure of the eye and person-specific parameters [3], etc. Unlike our method that requires any set of distinct features on the face, these methods require a particular set of features to be visible or not-occluded. The same is true for, many flexible face model oriented head pose [28]. In [14], Ishikawa et al. describes driver gaze tracking by way of using global head model, specifically an Active Appearance Model(AAM), to track the head. These and other head pose estimation algorithms are further explored in recent surveys [21], [12]. Many databases are available for the evaluation of head pose estimation algorithms from images to videos sequences,

3 Figure 2: Video sequence is taken from on-road driving in the LISA-P Head Pose database. This sequence shows the driver performing spatially large head turns. from far-field to near-field camera views. CMU PIE is one of many well known databases for containing facial images of different people under different illuminations, across multiple poses and with varying expressions [24]. Another interesting database is the IDIAP Head Pose Database [1] of video sequences showing one or more heads of participants engaged in natural activities (i.e. meetings, debates), with head pose measured using magnetic trackers called the Flock of Birds system. This magnetic tracker system is also used in the BU Face Tracking [16]database for ground truth head pose measurements. The BU Face Tracking database contains two sets of video sequences, one collected under uniform illumination conditions and one under varying illumination, with free head motions of several subjects. Unlike our database, none of the above mentioned databases have neither on-road drivers as subjects nor manually annotated facial features additionally provided. More information on other databases can be found in [21]. 3. shown in Fig. 2. The video sequences were collected at a frame rate of 30 frames per second, with a 640x480 pixel resolution. Since the on-road driving environment is uncontrolled, the distribution of head orientation is not spatially equal. Fig. 3 shows a definite bias towards the driver looking straight most of the time which is to be expected in on-road driving. Nonetheless the subjects do perform spatially large head turns as shown in Fig.2. LISA-P HEAD POSE DATABASE To the best of our knowledge, there exists no public database that provides a collection of video sequences of head movements from on-road driving, with continuous ground truth measurements and manually annotated facial features. Providing such a database will allow researchers to have a common ground to compare future head pose estimation algorithms and head behavior studies. The LISA-P Head Pose Database will soon be released to the public at cvrr.ucsd. edu/lisa/datasets.html. Number of frames per subject Yaw Rot. angle Pitch Rot. angle Average STD 2843 Min 9708 Max Figure 3: A normalized histogram of all head pose orientations collected from on-road driving of the LISA-P Head Pose database separated by the three independent axis of rotation. Table 1 gives a summary on the accumulated head pose data over all 14 video sequences. There are close to frames, and the range of yaw/pitch rotation angles achieved differ from driver to driver. Even though the large portion consists of looking straight, the drivers do perform spatially large head turns as indicated by the min/max statistics. 3.2 Table 1: A summary on the accumulated head pose data over all 14 video sequences of on-road driving 3.1 On-Road Driving All head pose data collected in the on-road driving environment is natural and spontaneous. There are 14 video sequences of drivers with varying age, ethnicity and gender, and the drives took place both in the daytime and the night time. The subjects have markers and cameras relevant to a motion capture system on the back of their head and a few feet behind them, respectively, to capture ground truth head pose. Additionally, a frontal facing monocular camera records their head movements. The camera is located on the dashboard, where the view of the driver is as Facial Feature Annotation In addition to providing a database with a continuous view of the driver s head and ground truth for head pose, the database contains annotations of 7 prominent facial features. As shown in Fig. 4, the features are eyes corners, nose tip and nose corners. The illustrated features are chosen for the purpose that they conform to rigid-body motion constraints and are least sensitive to facial expression changes. Current annotations of these feature points exist for two video sequences and are performed once every 5 or 10 frames; annotation frequency increases when there are interesting head movements. These chosen features have two more useful traits: one is that not all features are occluded due to large spatial head turns and the second is that these features have a higher probability of being uniquely detected in blurry video se- 151

4 Figure 4: Illustration of the 7 prominent features as annotated in the database: eye corners, nose tip, and nose corners. Annotations are shown using a real image (Left) from the dataset and it s corresponding points on a 3D generic-face model [2](Right) quence caused by fast head movements. Fig.5 shows a histogram of the number of features annotated in the dataset. This comes to show that most of the time, the chosen prominent features are visible or not-occluded in the image plane. 4. OUR APPROACH As shown in the block diagram of our proposed approach (Fig. 1), the framework consists of two steps. First, at least four prominent facial features are extracted from the detected face region in a monocular 2D image. In the next step, the detected facial features and their correspondence on a 3D generic-face model [2] are used to estimate the head orientation. Specifically, we use Pose from Orthography and Scaling (POS) algorithm [7] to find the rotation matrix and the translation vector, such that when applied to the 3D face model and projected onto a 2D plane, it should yield a representation of the 2D face image of interest. The details of each step are provided in the following sections. 4.1 ROI Detection Feature detection has been an ongoing effort for many decades. In literature, many algorithms exist to extract the features robustly and accurately. We use CLM [23], a model based technique for deformable object fitting. It utilizes a parametrized shape model and an ensemble of local feature detectors to jointly detect and track facial features. The number of landmarks used for typical CLM model fitting is 66 landmarks. The model fitting is such that some reference model is known a priori and affine transformation is used to align the input image with the reference model. Overall the process is such that it takes an image and an initial estimate of landmark positions on the image if available, and outputs the optimal landmark positions for the current image. 4.2 POS Algorithm to Head Pose The POS algorithm [7] approximates perspective projection with a scaled orthographic projection to describe the relationship between a 3D point face model and the projection of the model onto an image plane. Although the POSIT algorithm [7] (POS with ITerations) is designed to compute 152 Figure 5: A normalized histogram of annotated facial features points from on-road driving in the LISA-P Head Pose database. Feature acronyms: L=Left, R=Right, E=Eye, N=Nose, C=Corner, T=Tip. better scaled orthographic projection of the feature points on the 3D face model onto the image plane at every iteration, it requires knowledge of an intrinsic camera parameter - the focal length. Therefore, in our current implementation, we use the POS algorithm. The POS algorithm assumes the distribution of the features on the objects as well as the image of these points under perspective projection to be known. Approximating the perspective projection with a scaled orthographic projection, the algorithm efficiently estimates the rotation matrix by solving a linear system of equations. Details and geometric proofs regarding the POS algorithm can be found in [7]. From the obtained rotation matrix, pitch (ψ), yaw ( θ) and roll (φ ) rotation angles in the Euler coordinate system are computed as follows [25]. We assume that R11 R12 R13 R = Rz (φ)ry (θ)rx (ψ) = R21 R22 R23 R31 R32 R33 and compute θ = sin 1 (R31 ) ψ = tan 1 (R32 /R33 ) φ = tan 1 (R21 /R11 ) 5. RESULTS AND DISCUSSION Our approach is evaluated using the on-road driving video sequences of the LISA-P Head Pose database. The evaluation is conducted in two parts.the first involves using the manually annotated features points of two video sequence in the database for head pose estimation. The second part involves using CLM for fully automatic feature extraction. The manually annotated features from the LISA-P Head Pose database as available for two video sequences and automatic feature extraction using CLM on four subjects from the database are used to evaluate our approach. In Fig. 6, the red curve and the green curve show the result of using manual feature annotation and CLM feature extraction, respectively, to compute head pose estimation for one of the

5 Figure 6: Shown are ground truth head pose measurements (blue), head pose estimation using manually annotated facial features and the POS algorithm (red), and head pose estimation using CLM for feature extraction and the POS algorithm (green). The database for this evaluation is taken from on-road driving. subjects in the database. It shows that our approach follows the ground truth head pose well. The standard deviation (STD) errors of head pose estimation using manual feature annotation and automatic feature extraction are given in Table 2. The presented STD values were computed when the deviation of the head pose estimation was within 30 of the ground truth head pose. This resulted in discarding 5% and 22% of the frames when using manually annotated features and CLM features, respectively. The reported error can be attributed to the fact that the mean 3D face model differs from the actual 3D configuration of the tracked facial features. Furthermore, manual annotations are often not consistent. As for CLM, the deviations occurred mainly because CLM allows for any facial orientation in the image plane. If further restraints on the location and possible orientations of the face are used, we can expect improved performance. Approach Manual feature annotation + POS CLM feature extraction + POS Pitch STD 6.0 Yaw STD 5.0 Roll STD Table 2: In evaluating our approach, feature detection was done using manual annotation and CLM. Standard deviation error in the three independent axis of rotation is given. The database for this evaluation is taken from real-car driving. Taking head pose estimation using manual feature annotation as a baseline for the best possible case using POS algorithm, using CLM for feature extraction for head pose estimation shows promising results for automating our approach. 6. CONCLUDING REMARKS In this paper we have introduced a valuable database containing video sequence of various head poses in an on-road driving environment. Additionally, ground truth for head pose is provided with the help of motion capture system and manual facial feature annotation is provided for up to 7 prominent features. We have also introduced an approach to estimate head pose during large spatial head turns where some, but not all, facial features are self-occluded and during varying speed in head movements. Evaluation of our approach on the introduced database shows an average of 7 in standard deviation error. Going forward, the 3D generic-face model used for the POS algorithm should be learned on-line to better fit the user s face and thus produce better head pose estimation. Another future work is to find an optimal way to track minimal number of good features such that head pose can be estimated with high precision over large spatial deviations. 7. ACKNOWLEDGMENTS This research is supported in part by projects sponsored by the University of California Discovery program, the National Science Foundation and the Electronics Research Laboratory of Volkswagen of America. The authors would like to thank Amruta Trivedi and Aaron Briones for helping with annotation of the facial features in our database, and the members of the for their assistance. 153

6 8. REFERENCES [1] S. Ba and J.-M. Odobez. Evaluation of multiple cue head pose estimation algorithms in natural environements. In Multimedia and Expo, ICME IEEE International Conference on, pages , july [2] J. Busby. 3d head scan, August [3] J. Chen and Q. Ji. 3d gaze estimation with a single camera without ir illumination. In Pattern Recognition, ICPR th International Conference on, pages 1 4, dec [4] S. Cheng and M. Trivedi. Turn-intent analysis using body pose for intelligent driver assistance. Pervasive Computing, IEEE, 5(4):28 37, oct.-dec [5] L. H. Christiansen, N. Y. Frederiksen, A. Ranch, and M. B. Skov. Investigating the effects of an advance warning in-vehicle system on behavior and attention in controlled driving. In Proceedings of the 3rd Internation Conference on Automotive User Interface and Interactive Vehicular Applications, pages , [6] J. Curin, M. Labsky, T. Macek, and J. Kleindienst. Dictating and editing short texts while driving: Distraction and task completion. In Proceedings of the 3rd Internation Conference on Automotive User Interface and Interactive Vehicular Applications, AutomotiveUI 11, pages 13 20, [7] D. F. Dementhon and L. S. Davis. Model-based object pose in 25 lines of code. International Journal of Computer Vision, 15: , [8] A. Doshi and M. Trivedi. On the roles of eye gaze and head dynamics in predicting driver s intent to change lanes. Intelligent Transportation Systems, IEEE Transactions on, 10(3): , sept [9] L. Fletcher, L. Petersson, N. Barnes, D. Austin, and A. Zelinsky. A sign reading driver assistance system using eye gaze. In Robotics and Automation, ICRA Proceedings of the 2005 IEEE International Conference on, pages , april [10] L. Fletcher, L. Petersson, and A. Zelinsky. Road scene monotony detection in a fatigue management driver assistance system. In Intelligent Vehicles Symposium, Proceedings. IEEE, pages , june [11] A. Gee and R. Cipolla. Determining the gaze of faces in images. Image and Vision Computing, 12(10): , [12] D. Hansen and Q. Ji. In the eye of the beholder: A survey of models for eyes and gaze. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 32(3): , march [13] T. Horprasert, Y. Yacoob, and L. Davis. Computing 3-d head orientation from a monocular image sequence. In Automatic Face and Gesture Recognition, 1996., Proceedings of the Second International Conference on, pages , oct [14] T. Ishikawa, S. Baker, I. Matthews, and T. Kanade. Passive driver gaze tracking with active appearance models. In Proceedings of the 11th World Congress on Intelligent Transportation Systems, October [15] J. Jain and A. Jain. Displacement measurement and its application in interframe image coding. Communications, IEEE Transactions on, 29(12): , dec [16] M. La Cascia and S. Sclaroff. Fast, reliable head tracking under varying illumination. In Computer Vision and Pattern Recognition, IEEE Computer Society Conference on., volume 1, pages 2 vol. (xxiii ), [17] S. Martin, C. Tran, A. Tawari, J. Kwan, and M. M. Trivedi. Optical flow based head movement and gesture analyzer (ohmega). In Pattern Recognition (ICPR), 21st International Conference on, Nov [18] S. Martin, C. Tran, and M. M. Trivedi. Optical flow based head movement and gesture analysis in automotive environment. In IEEE International Conference on Intelligent Transportation Systems-ITSC, Sept [19] J. McCall, D. Wipf, M. Trivedi, and B. Rao. Lane change intent analysis using robust operators and sparse bayesian learning. Intelligent Transportation Systems, IEEE Transactions on, 8(3): , sept [20] E. Murphy-Chutorian and M. Trivedi. Hyhope: Hybrid head orientation and position estimation for vision-based driver head tracking. In Intelligent Vehicles Symposium, 2008 IEEE, pages , june [21] E. Murphy-Chutorian and M. Trivedi. Head pose estimation in computer vision: A survey. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 31(4): , april [22] E. Murphy-Chutorian and M. Trivedi. Head pose estimation and augmented reality tracking: An integrated system and evaluation for monitoring driver awareness. Intelligent Transportation Systems, IEEE Transactions on, 11(2): , june [23] J. Saragih, S. Lucey, and J. Cohn. Face alignment through subspace constrained mean-shifts. In Computer Vision, 2009 IEEE 12th International Conference on, pages , oct [24] T. Sim, S. Baker, and M. Bsat. The cmu pose, illumination, and expression (pie) database. In Automatic Face and Gesture Recognition, Proceedings. Fifth IEEE International Conference on, pages 46 51, may [25] G. Slabaugh. Computing euler angles from a rotation matrix. [26] R. Valenti, N. Sebe, and T. Gevers. Combining head pose and eye location information for gaze estimation. Image Processing, IEEE Transactions on, 21(2): , feb [27] J.-G. Wang and E. Sung. Em enhancement of 3d head pose estimated by point at infinity. Image and Vision Computing, 25(12): , The age of human computer interaction. [28] J. Wu and M. M. Trivedi. A two-stage head pose estimation framework and evaluation. Pattern Recognition, 41(3): , Part Special issue: Feature Generation and Machine Learning for Robust Multimodal Biometrics. [29] H. Zhang, M. Smith, and R. Dufour. A final report of safety vehicles using adaptive interface technology: Visual distraction. 154

Monitoring Head Dynamics for Driver Assistance Systems: A Multi-Perspective Approach

Monitoring Head Dynamics for Driver Assistance Systems: A Multi-Perspective Approach Proceedings of the 16th International IEEE Annual Conference on Intelligent Transportation Systems (ITSC 2013), The Hague, The Netherlands, October 6-9, 2013 WeC4.5 Monitoring Head Dynamics for Driver

More information

Real Time Face Tracking and Pose Estimation Using an Adaptive Correlation Filter for Human-Robot Interaction

Real Time Face Tracking and Pose Estimation Using an Adaptive Correlation Filter for Human-Robot Interaction Real Time Face Tracking and Pose Estimation Using an Adaptive Correlation Filter for Human-Robot Interaction Vo Duc My and Andreas Zell Abstract In this paper, we present a real time algorithm for mobile

More information

Analyzing the Relationship Between Head Pose and Gaze to Model Driver Visual Attention

Analyzing the Relationship Between Head Pose and Gaze to Model Driver Visual Attention Analyzing the Relationship Between Head Pose and Gaze to Model Driver Visual Attention Sumit Jha and Carlos Busso Multimodal Signal Processing (MSP) Laboratory Department of Electrical Engineering, The

More information

Gaze interaction (2): models and technologies

Gaze interaction (2): models and technologies Gaze interaction (2): models and technologies Corso di Interazione uomo-macchina II Prof. Giuseppe Boccignone Dipartimento di Scienze dell Informazione Università di Milano boccignone@dsi.unimi.it http://homes.dsi.unimi.it/~boccignone/l

More information

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,

More information

On Driver Gaze Estimation: Explorations and Fusion of Geometric and Data Driven Approaches

On Driver Gaze Estimation: Explorations and Fusion of Geometric and Data Driven Approaches 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC) Windsor Oceanico Hotel, Rio de Janeiro, Brazil, November 1-4, 2016 On Driver Gaze Estimation: Explorations and Fusion

More information

Occlusion Robust Multi-Camera Face Tracking

Occlusion Robust Multi-Camera Face Tracking Occlusion Robust Multi-Camera Face Tracking Josh Harguess, Changbo Hu, J. K. Aggarwal Computer & Vision Research Center / Department of ECE The University of Texas at Austin harguess@utexas.edu, changbo.hu@gmail.com,

More information

3D Active Appearance Model for Aligning Faces in 2D Images

3D Active Appearance Model for Aligning Faces in 2D Images 3D Active Appearance Model for Aligning Faces in 2D Images Chun-Wei Chen and Chieh-Chih Wang Abstract Perceiving human faces is one of the most important functions for human robot interaction. The active

More information

Information and Communication Engineering July 9th Survey of Techniques for Head Pose Estimation from Low Resolution Images

Information and Communication Engineering July 9th Survey of Techniques for Head Pose Estimation from Low Resolution Images Information and Communication Engineering July 9th 2010 Survey of Techniques for Head Pose Estimation from Low Resolution Images Sato Laboratory D1, 48-107409, Teera SIRITEERAKUL Abstract Head pose estimation

More information

Determining pose of a human face from a single monocular image

Determining pose of a human face from a single monocular image Determining pose of a human face from a single monocular image Jian-Gang Wang 1, Eric Sung 2, Ronda Venkateswarlu 1 1 Institute for Infocomm Research 21 Heng Mui Keng Terrace, Singapore 119613 2 Nanyang

More information

Full-Motion Recovery from Multiple Video Cameras Applied to Face Tracking and Recognition

Full-Motion Recovery from Multiple Video Cameras Applied to Face Tracking and Recognition Full-Motion Recovery from Multiple Video Cameras Applied to Face Tracking and Recognition Josh Harguess, Changbo Hu, J. K. Aggarwal Computer & Vision Research Center / Department of ECE The University

More information

Baseline Face Detection, Head Pose Estimation, and Coarse Direction Detection for Facial Data in the SHRP2 Naturalistic Driving Study

Baseline Face Detection, Head Pose Estimation, and Coarse Direction Detection for Facial Data in the SHRP2 Naturalistic Driving Study Baseline Face Detection, Head Pose Estimation, and Coarse Direction Detection for Facial Data in the SHRP2 Naturalistic Driving Study J. Paone, D. Bolme, R. Ferrell, Member, IEEE, D. Aykac, and T. Karnowski,

More information

Occluded Facial Expression Tracking

Occluded Facial Expression Tracking Occluded Facial Expression Tracking Hugo Mercier 1, Julien Peyras 2, and Patrice Dalle 1 1 Institut de Recherche en Informatique de Toulouse 118, route de Narbonne, F-31062 Toulouse Cedex 9 2 Dipartimento

More information

Abstract. 1 Introduction. 2 Motivation. Information and Communication Engineering October 29th 2010

Abstract. 1 Introduction. 2 Motivation. Information and Communication Engineering October 29th 2010 Information and Communication Engineering October 29th 2010 A Survey on Head Pose Estimation from Low Resolution Image Sato Laboratory M1, 48-106416, Isarun CHAMVEHA Abstract Recognizing the head pose

More information

Available online at ScienceDirect. Procedia Computer Science 50 (2015 )

Available online at   ScienceDirect. Procedia Computer Science 50 (2015 ) Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 50 (2015 ) 617 622 2nd International Symposium on Big Data and Cloud Computing (ISBCC 15) Monitoring Driver Head Postures

More information

Real Time Eye s off the Road Detection System for Driver Assistance

Real Time Eye s off the Road Detection System for Driver Assistance Real Time Eye s off the Road Detection System for Driver Assistance Bandi Ramagopal M. Tech Student, Department of ECE, Shri Sai Institute of Engineering and Technology, Anantapur (dt) A.P. Abstract Automated

More information

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation Face Tracking Amit K. Roy-Chowdhury and Yilei Xu Department of Electrical Engineering, University of California, Riverside, CA 92521, USA {amitrc,yxu}@ee.ucr.edu Synonyms Facial Motion Estimation Definition

More information

Real Time Eye s Off the Road Detection System for Driver Assistance

Real Time Eye s Off the Road Detection System for Driver Assistance Real Time Eye s Off the Road Detection System for Driver Assistance Manthri Lavanya & C.Ravi Teja 1 M.Tech,Assistant Professor, Department of ECE, Anantha Lakshmi Institute of Technology, Anantapur, Ap,

More information

Passive driver gaze tracking with active appearance models

Passive driver gaze tracking with active appearance models Carnegie Mellon University Research Showcase @ CMU Robotics Institute School of Computer Science 2004 Passive driver gaze tracking with active appearance models Takahiro Ishikawa Carnegie Mellon University

More information

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices Technical University of Cluj-Napoca Image Processing and Pattern Recognition Research Center www.cv.utcluj.ro Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving

More information

A Video Database for Head Pose Tracking Evaluation C O M M U N I C A T I O N I D I A P. Jean-Marc Odobez a. Silèye O. Ba a.

A Video Database for Head Pose Tracking Evaluation C O M M U N I C A T I O N I D I A P. Jean-Marc Odobez a. Silèye O. Ba a. C O M M U N I C A T I O N A Video Database for Head Pose Tracking Evaluation Silèye O. Ba a Jean-Marc Odobez a IDIAP Com 5-4 I D I A P September 25 a IDIAP Research Institute IDIAP Research Institute www.idiap.ch

More information

On Road Vehicle Detection using Shadows

On Road Vehicle Detection using Shadows On Road Vehicle Detection using Shadows Gilad Buchman Grasp Lab, Department of Computer and Information Science School of Engineering University of Pennsylvania, Philadelphia, PA buchmag@seas.upenn.edu

More information

A Robust Two Feature Points Based Depth Estimation Method 1)

A Robust Two Feature Points Based Depth Estimation Method 1) Vol.31, No.5 ACTA AUTOMATICA SINICA September, 2005 A Robust Two Feature Points Based Depth Estimation Method 1) ZHONG Zhi-Guang YI Jian-Qiang ZHAO Dong-Bin (Laboratory of Complex Systems and Intelligence

More information

AAM Based Facial Feature Tracking with Kinect

AAM Based Facial Feature Tracking with Kinect BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 15, No 3 Sofia 2015 Print ISSN: 1311-9702; Online ISSN: 1314-4081 DOI: 10.1515/cait-2015-0046 AAM Based Facial Feature Tracking

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Vision-based Lane Analysis: Exploration of Issues and Approaches for Embedded Realization

Vision-based Lane Analysis: Exploration of Issues and Approaches for Embedded Realization 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops Vision-based Lane Analysis: Exploration of Issues and Approaches for Embedded Realization R. K. Satzoda and Mohan M. Trivedi Computer

More information

Probabilistic Head Pose Tracking Evaluation in Single and Multiple Camera Setups

Probabilistic Head Pose Tracking Evaluation in Single and Multiple Camera Setups Probabilistic Head Pose Tracking Evaluation in Single and Multiple Camera Setups Sileye O. Ba 1 and Jean-Marc Odobez 1 IDIAP Research Institute, Martigny, Switzerland Abstract. This paper presents our

More information

Measurement of Pedestrian Groups Using Subtraction Stereo

Measurement of Pedestrian Groups Using Subtraction Stereo Measurement of Pedestrian Groups Using Subtraction Stereo Kenji Terabayashi, Yuki Hashimoto, and Kazunori Umeda Chuo University / CREST, JST, 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551, Japan terabayashi@mech.chuo-u.ac.jp

More information

3D TRACKING AND DYNAMIC ANALYSIS OF HUMAN HEAD MOVEMENTS AND ATTENTIONAL TARGETS. Erik Murphy-Chutorian and Mohan M. Trivedi

3D TRACKING AND DYNAMIC ANALYSIS OF HUMAN HEAD MOVEMENTS AND ATTENTIONAL TARGETS. Erik Murphy-Chutorian and Mohan M. Trivedi 3D TRACKING AND DYNAMIC ANALYSIS OF HUMAN HEAD MOVEMENTS AND ATTENTIONAL TARGETS Erik Murphy-Chutorian and Mohan M. Trivedi Systems for Human Interactivity, Visualization and Analysis Computer Vision and

More information

ADVANCES in NATURAL and APPLIED SCIENCES

ADVANCES in NATURAL and APPLIED SCIENCES ADVANCES in NATURAL and APPLIED SCIENCES ISSN: 1995-0772 Published BY AENSI Publication EISSN: 1998-1090 http://www.aensiweb.com/anas 2016 Special 10(6): pages Open Access Journal Estimation of Various

More information

Fully Automatic Endoscope Calibration for Intraoperative Use

Fully Automatic Endoscope Calibration for Intraoperative Use Fully Automatic Endoscope Calibration for Intraoperative Use Christian Wengert, Mireille Reeff, Philippe C. Cattin, Gábor Székely Computer Vision Laboratory, ETH Zurich, 8092 Zurich, Switzerland {wengert,

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Model Based Techniques for Tracking Human Head for Human Robot Interaction

Model Based Techniques for Tracking Human Head for Human Robot Interaction Model Based Techniques for Tracking Human Head for Human Robot Interaction Faisal Khan University of Wisconsin, Madison faisal@cs.wisc.edu Varghese Mathew University of Wisconsin, Madison vmathew@cs.wisc.edu

More information

Intuitive Human-Robot Interaction through Active 3D Gaze Tracking

Intuitive Human-Robot Interaction through Active 3D Gaze Tracking Intuitive Human-Robot Interaction through Active 3D Gaze Tracking Rowel Atienza and Alexander Zelinsky Research School of Information Sciences and Engineering The Australian National University Canberra,

More information

Postprint.

Postprint. http://www.diva-portal.org Postprint This is the accepted version of a paper presented at 14th International Conference of the Biometrics Special Interest Group, BIOSIG, Darmstadt, Germany, 9-11 September,

More information

Eye Detection by Haar wavelets and cascaded Support Vector Machine

Eye Detection by Haar wavelets and cascaded Support Vector Machine Eye Detection by Haar wavelets and cascaded Support Vector Machine Vishal Agrawal B.Tech 4th Year Guide: Simant Dubey / Amitabha Mukherjee Dept of Computer Science and Engineering IIT Kanpur - 208 016

More information

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation IJCSNS International Journal of Computer Science and Network Security, VOL.13 No.11, November 2013 1 Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial

More information

Recognition (Part 4) Introduction to Computer Vision CSE 152 Lecture 17

Recognition (Part 4) Introduction to Computer Vision CSE 152 Lecture 17 Recognition (Part 4) CSE 152 Lecture 17 Announcements Homework 5 is due June 9, 11:59 PM Reading: Chapter 15: Learning to Classify Chapter 16: Classifying Images Chapter 17: Detecting Objects in Images

More information

A Robust Method of Facial Feature Tracking for Moving Images

A Robust Method of Facial Feature Tracking for Moving Images A Robust Method of Facial Feature Tracking for Moving Images Yuka Nomura* Graduate School of Interdisciplinary Information Studies, The University of Tokyo Takayuki Itoh Graduate School of Humanitics and

More information

DEPTH AND GEOMETRY FROM A SINGLE 2D IMAGE USING TRIANGULATION

DEPTH AND GEOMETRY FROM A SINGLE 2D IMAGE USING TRIANGULATION 2012 IEEE International Conference on Multimedia and Expo Workshops DEPTH AND GEOMETRY FROM A SINGLE 2D IMAGE USING TRIANGULATION Yasir Salih and Aamir S. Malik, Senior Member IEEE Centre for Intelligent

More information

Multi-sensor Gaze-tracking

Multi-sensor Gaze-tracking Multi-sensor Gaze-tracking Joe Rice March 2017 Abstract In this paper, we propose a method using multiple gaze-tracking-capable sensors along with fuzzy data-fusion techniques to improve gaze-estimation.

More information

On Performance Evaluation Metrics for Lane Estimation

On Performance Evaluation Metrics for Lane Estimation On Performance Evaluation Metrics for Lane Estimation Ravi Kumar Satzoda and Mohan M. Trivedi Laboratory for Intelligent and Safe Automobiles, University of California San Diego, La Jolla, CA-92093 Email:

More information

Using temporal seeding to constrain the disparity search range in stereo matching

Using temporal seeding to constrain the disparity search range in stereo matching Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department

More information

I D I A P. Probabilistic Head Pose Tracking Evaluation in Single and Multiple Camera Setups Sileye O. Ba a R E S E A R C H R E P O R T

I D I A P. Probabilistic Head Pose Tracking Evaluation in Single and Multiple Camera Setups Sileye O. Ba a R E S E A R C H R E P O R T R E S E A R C H R E P O R T I D I A P Probabilistic Head Pose Tracking Evaluation in Single and Multiple Camera Setups Sileye O. Ba a Jean-Marc Odobez a IDIAP RR 07-21 march 2007 published in Classification

More information

Announcements. Recognition (Part 3) Model-Based Vision. A Rough Recognition Spectrum. Pose consistency. Recognition by Hypothesize and Test

Announcements. Recognition (Part 3) Model-Based Vision. A Rough Recognition Spectrum. Pose consistency. Recognition by Hypothesize and Test Announcements (Part 3) CSE 152 Lecture 16 Homework 3 is due today, 11:59 PM Homework 4 will be assigned today Due Sat, Jun 4, 11:59 PM Reading: Chapter 15: Learning to Classify Chapter 16: Classifying

More information

Human Robot Communication. Paul Fitzpatrick

Human Robot Communication. Paul Fitzpatrick Human Robot Communication Paul Fitzpatrick Human Robot Communication Motivation for communication Human-readable actions Reading human actions Conclusions Motivation What is communication for? Transferring

More information

Head Pose Estimation for Driver Assistance Systems: A Robust Algorithm and Experimental Evaluation

Head Pose Estimation for Driver Assistance Systems: A Robust Algorithm and Experimental Evaluation Proceedings of the 2007 IEEE Intelligent Transportation Systems Conference Seattle, WA, USA, Sept. 30 - Oct. 3, 2007 WeA1.4 Head Pose Estimation for Driver Assistance Systems: A Robust Algorithm and Experimental

More information

3D Head Pose and Facial Expression Tracking using a Single Camera

3D Head Pose and Facial Expression Tracking using a Single Camera Journal of Universal Computer Science, vol. 16, no. 6 (2010), 903-920 submitted: 15/9/09, accepted: 2/3/10, appeared: 28/3/10 J.UCS 3D Head Pose and Facial Expression Tracking using a Single Camera LucasD.Terissi,JuanC.Gómez

More information

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Tomokazu Satoy, Masayuki Kanbaray, Naokazu Yokoyay and Haruo Takemuraz ygraduate School of Information

More information

818 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 15, NO. 2, APRIL 2014

818 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 15, NO. 2, APRIL 2014 818 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 15, NO. 2, APRIL 2014 Continuous Head Movement Estimator for Driver Assistance: Issues, Algorithms, and On-Road Evaluations Ashish Tawari,

More information

Learning the Deep Features for Eye Detection in Uncontrolled Conditions

Learning the Deep Features for Eye Detection in Uncontrolled Conditions 2014 22nd International Conference on Pattern Recognition Learning the Deep Features for Eye Detection in Uncontrolled Conditions Yue Wu Dept. of ECSE, Rensselaer Polytechnic Institute Troy, NY, USA 12180

More information

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES Jie Shao a, Wuming Zhang a, Yaqiao Zhu b, Aojie Shen a a State Key Laboratory of Remote Sensing Science, Institute of Remote Sensing

More information

Head Pose Estimation with One Camera, in Uncalibrated Environments

Head Pose Estimation with One Camera, in Uncalibrated Environments Head Pose Estimation with One Camera, in Uncalibrated Environments Stylianos Asteriadis Image, Video and Multimedia Systems Lab National Technical University of Athens 9, Iroon Polytechniou str, Athens,

More information

Preceding vehicle detection and distance estimation. lane change, warning system.

Preceding vehicle detection and distance estimation. lane change, warning system. Preceding vehicle detection and distance estimation for lane change warning system U. Iqbal, M.S. Sarfraz Computer Vision Research Group (COMVis) Department of Electrical Engineering, COMSATS Institute

More information

Vehicle Occupant Posture Analysis Using Voxel Data

Vehicle Occupant Posture Analysis Using Voxel Data Ninth World Congress on Intelligent Transport Systems, Chicago, Illinois, October Vehicle Occupant Posture Analysis Using Voxel Data Ivana Mikic, Mohan Trivedi Computer Vision and Robotics Research Laboratory

More information

Monocular Vision Based Autonomous Navigation for Arbitrarily Shaped Urban Roads

Monocular Vision Based Autonomous Navigation for Arbitrarily Shaped Urban Roads Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August 14-15, 2014 Paper No. 127 Monocular Vision Based Autonomous Navigation for Arbitrarily

More information

Adaptive Skin Color Classifier for Face Outline Models

Adaptive Skin Color Classifier for Face Outline Models Adaptive Skin Color Classifier for Face Outline Models M. Wimmer, B. Radig, M. Beetz Informatik IX, Technische Universität München, Germany Boltzmannstr. 3, 87548 Garching, Germany [wimmerm, radig, beetz]@informatik.tu-muenchen.de

More information

Facial Processing Projects at the Intelligent Systems Lab

Facial Processing Projects at the Intelligent Systems Lab Facial Processing Projects at the Intelligent Systems Lab Qiang Ji Intelligent Systems Laboratory (ISL) Department of Electrical, Computer, and System Eng. Rensselaer Polytechnic Institute jiq@rpi.edu

More information

Creating a distortion characterisation dataset for visual band cameras using fiducial markers.

Creating a distortion characterisation dataset for visual band cameras using fiducial markers. Creating a distortion characterisation dataset for visual band cameras using fiducial markers. Robert Jermy Council for Scientific and Industrial Research Email: rjermy@csir.co.za Jason de Villiers Council

More information

Definition, Detection, and Evaluation of Meeting Events in Airport Surveillance Videos

Definition, Detection, and Evaluation of Meeting Events in Airport Surveillance Videos Definition, Detection, and Evaluation of Meeting Events in Airport Surveillance Videos Sung Chun Lee, Chang Huang, and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA sungchun@usc.edu,

More information

Chapter 9 Object Tracking an Overview

Chapter 9 Object Tracking an Overview Chapter 9 Object Tracking an Overview The output of the background subtraction algorithm, described in the previous chapter, is a classification (segmentation) of pixels into foreground pixels (those belonging

More information

Comparative Study of Coarse Head Pose Estimation

Comparative Study of Coarse Head Pose Estimation Comparative Study of Coarse Head Pose Estimation Lisa M. Brown and Ying-Li Tian IBM T.J. Watson Research Center Hawthorne, NY 10532 {lisabr,yltian}@us.ibm.com Abstract For many practical applications,

More information

Improving Vision-Based Distance Measurements using Reference Objects

Improving Vision-Based Distance Measurements using Reference Objects Improving Vision-Based Distance Measurements using Reference Objects Matthias Jüngel, Heinrich Mellmann, and Michael Spranger Humboldt-Universität zu Berlin, Künstliche Intelligenz Unter den Linden 6,

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

On Modeling Variations for Face Authentication

On Modeling Variations for Face Authentication On Modeling Variations for Face Authentication Xiaoming Liu Tsuhan Chen B.V.K. Vijaya Kumar Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA 15213 xiaoming@andrew.cmu.edu

More information

A Street Scene Surveillance System for Moving Object Detection, Tracking and Classification

A Street Scene Surveillance System for Moving Object Detection, Tracking and Classification A Street Scene Surveillance System for Moving Object Detection, Tracking and Classification Huei-Yung Lin * and Juang-Yu Wei Department of Electrical Engineering National Chung Cheng University Chia-Yi

More information

DISTANCE MAPS: A ROBUST ILLUMINATION PREPROCESSING FOR ACTIVE APPEARANCE MODELS

DISTANCE MAPS: A ROBUST ILLUMINATION PREPROCESSING FOR ACTIVE APPEARANCE MODELS DISTANCE MAPS: A ROBUST ILLUMINATION PREPROCESSING FOR ACTIVE APPEARANCE MODELS Sylvain Le Gallou*, Gaspard Breton*, Christophe Garcia*, Renaud Séguier** * France Telecom R&D - TECH/IRIS 4 rue du clos

More information

#65 MONITORING AND PREDICTING PEDESTRIAN BEHAVIOR AT TRAFFIC INTERSECTIONS

#65 MONITORING AND PREDICTING PEDESTRIAN BEHAVIOR AT TRAFFIC INTERSECTIONS #65 MONITORING AND PREDICTING PEDESTRIAN BEHAVIOR AT TRAFFIC INTERSECTIONS Final Research Report Luis E. Navarro-Serment, Ph.D. The Robotics Institute Carnegie Mellon University Disclaimer The contents

More information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information Proceedings of the World Congress on Electrical Engineering and Computer Systems and Science (EECSS 2015) Barcelona, Spain July 13-14, 2015 Paper No. 335 Efficient SLAM Scheme Based ICP Matching Algorithm

More information

A REAL-TIME FACIAL FEATURE BASED HEAD TRACKER

A REAL-TIME FACIAL FEATURE BASED HEAD TRACKER A REAL-TIME FACIAL FEATURE BASED HEAD TRACKER Jari Hannuksela, Janne Heikkilä and Matti Pietikäinen {jari.hannuksela, jth, mkp}@ee.oulu.fi Machine Vision Group, Infotech Oulu P.O. Box 45, FIN-914 University

More information

Head Pose Estimation by using Morphological Property of Disparity Map

Head Pose Estimation by using Morphological Property of Disparity Map Head Pose Estimation by using Morphological Property of Disparity Map Sewoong Jun*, Sung-Kee Park* and Moonkey Lee** *Intelligent Robotics Research Center, Korea Institute Science and Technology, Seoul,

More information

Image processing techniques for driver assistance. Razvan Itu June 2014, Technical University Cluj-Napoca

Image processing techniques for driver assistance. Razvan Itu June 2014, Technical University Cluj-Napoca Image processing techniques for driver assistance Razvan Itu June 2014, Technical University Cluj-Napoca Introduction Computer vision & image processing from wiki: any form of signal processing for which

More information

Time-to-Contact from Image Intensity

Time-to-Contact from Image Intensity Time-to-Contact from Image Intensity Yukitoshi Watanabe Fumihiko Sakaue Jun Sato Nagoya Institute of Technology Gokiso, Showa, Nagoya, 466-8555, Japan {yukitoshi@cv.,sakaue@,junsato@}nitech.ac.jp Abstract

More information

An Interactive Technique for Robot Control by Using Image Processing Method

An Interactive Technique for Robot Control by Using Image Processing Method An Interactive Technique for Robot Control by Using Image Processing Method Mr. Raskar D. S 1., Prof. Mrs. Belagali P. P 2 1, E&TC Dept. Dr. JJMCOE., Jaysingpur. Maharashtra., India. 2 Associate Prof.

More information

Generic Face Alignment Using an Improved Active Shape Model

Generic Face Alignment Using an Improved Active Shape Model Generic Face Alignment Using an Improved Active Shape Model Liting Wang, Xiaoqing Ding, Chi Fang Electronic Engineering Department, Tsinghua University, Beijing, China {wanglt, dxq, fangchi} @ocrserv.ee.tsinghua.edu.cn

More information

Using Geometric Blur for Point Correspondence

Using Geometric Blur for Point Correspondence 1 Using Geometric Blur for Point Correspondence Nisarg Vyas Electrical and Computer Engineering Department, Carnegie Mellon University, Pittsburgh, PA Abstract In computer vision applications, point correspondence

More information

Advanced Driver Assistance Systems: A Cost-Effective Implementation of the Forward Collision Warning Module

Advanced Driver Assistance Systems: A Cost-Effective Implementation of the Forward Collision Warning Module Advanced Driver Assistance Systems: A Cost-Effective Implementation of the Forward Collision Warning Module www.lnttechservices.com Table of Contents Abstract 03 Introduction 03 Solution Overview 03 Output

More information

THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA

THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA Hwadong Sun 1, Dong Yeop Kim 1 *, Joon Ho Kwon 2, Bong-Seok Kim 1, and Chang-Woo Park 1 1 Intelligent Robotics Research Center,

More information

Categorization by Learning and Combining Object Parts

Categorization by Learning and Combining Object Parts Categorization by Learning and Combining Object Parts Bernd Heisele yz Thomas Serre y Massimiliano Pontil x Thomas Vetter Λ Tomaso Poggio y y Center for Biological and Computational Learning, M.I.T., Cambridge,

More information

Tracking driver actions and guiding phone usage for safer driving. Hongyu Li Jan 25, 2018

Tracking driver actions and guiding phone usage for safer driving. Hongyu Li Jan 25, 2018 Tracking driver actions and guiding phone usage for safer driving Hongyu Li Jan 25, 2018 1 Smart devices risks and opportunities Phone in use 14% Other distractions 86% Distraction-Affected Fatalities

More information

Vehicle Dimensions Estimation Scheme Using AAM on Stereoscopic Video

Vehicle Dimensions Estimation Scheme Using AAM on Stereoscopic Video Workshop on Vehicle Retrieval in Surveillance (VRS) in conjunction with 2013 10th IEEE International Conference on Advanced Video and Signal Based Surveillance Vehicle Dimensions Estimation Scheme Using

More information

Pose estimation using a variety of techniques

Pose estimation using a variety of techniques Pose estimation using a variety of techniques Keegan Go Stanford University keegango@stanford.edu Abstract Vision is an integral part robotic systems a component that is needed for robots to interact robustly

More information

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Zhiyan Zhang 1, Wei Qian 1, Lei Pan 1 & Yanjun Li 1 1 University of Shanghai for Science and Technology, China

More information

Parametric Manifold of an Object under Different Viewing Directions

Parametric Manifold of an Object under Different Viewing Directions Parametric Manifold of an Object under Different Viewing Directions Xiaozheng Zhang 1,2, Yongsheng Gao 1,2, and Terry Caelli 3 1 Biosecurity Group, Queensland Research Laboratory, National ICT Australia

More information

HEAD TRACKING FOR 3D AUDIO USING THE NINTENDO WII REMOTE

HEAD TRACKING FOR 3D AUDIO USING THE NINTENDO WII REMOTE HEAD TRACKING FOR 3D AUDIO USING THE NINTENDO WII REMOTE Mauricio Ubilla School of Enginering Department of Computer Science Pontificia Universidad Catolica de Chile Domingo Mery School of Enginering Department

More information

A Simulation Study and Experimental Verification of Hand-Eye-Calibration using Monocular X-Ray

A Simulation Study and Experimental Verification of Hand-Eye-Calibration using Monocular X-Ray A Simulation Study and Experimental Verification of Hand-Eye-Calibration using Monocular X-Ray Petra Dorn, Peter Fischer,, Holger Mönnich, Philip Mewes, Muhammad Asim Khalil, Abhinav Gulhar, Andreas Maier

More information

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press,   ISSN ransactions on Information and Communications echnologies vol 6, 996 WI Press, www.witpress.com, ISSN 743-357 Obstacle detection using stereo without correspondence L. X. Zhou & W. K. Gu Institute of Information

More information

Flexible Calibration of a Portable Structured Light System through Surface Plane

Flexible Calibration of a Portable Structured Light System through Surface Plane Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured

More information

Generic 3D Face Pose Estimation using Facial Shapes

Generic 3D Face Pose Estimation using Facial Shapes Generic 3D Face Pose Estimation using Facial Shapes Jingu Heo CyLab Biometrics Center Carnegie Mellon University 5000 Forbes Ave, Pittsburgh, PA 15213 jheo@cmu.edu Marios Savvides CyLab Biometrics Center

More information

SHRP 2 Safety Research Symposium July 27, Site-Based Video System Design and Development: Research Plans and Issues

SHRP 2 Safety Research Symposium July 27, Site-Based Video System Design and Development: Research Plans and Issues SHRP 2 Safety Research Symposium July 27, 2007 Site-Based Video System Design and Development: Research Plans and Issues S09 Objectives Support SHRP2 program research questions: Establish crash surrogates

More information

Intensity-Depth Face Alignment Using Cascade Shape Regression

Intensity-Depth Face Alignment Using Cascade Shape Regression Intensity-Depth Face Alignment Using Cascade Shape Regression Yang Cao 1 and Bao-Liang Lu 1,2 1 Center for Brain-like Computing and Machine Intelligence Department of Computer Science and Engineering Shanghai

More information

Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images

Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images 1 Introduction - Steve Chuang and Eric Shan - Determining object orientation in images is a well-established topic

More information

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences Jean-François Lalonde, Srinivasa G. Narasimhan and Alexei A. Efros {jlalonde,srinivas,efros}@cs.cmu.edu CMU-RI-TR-8-32 July

More information

Three-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera

Three-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera Three-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera Kazuki Sakamoto, Alessandro Moro, Hiromitsu Fujii, Atsushi Yamashita, and Hajime Asama Abstract

More information

Tracking Requirements for ATC Tower Augmented Reality Environments

Tracking Requirements for ATC Tower Augmented Reality Environments Tracking Requirements for ATC Tower Augmented Reality Environments Magnus Axholt Department of Science and Technology (ITN) Linköping University, Sweden Supervisors: Anders Ynnerman Linköping University,

More information

Expression Detection in Video. Abstract Expression detection is useful as a non-invasive method of lie detection and

Expression Detection in Video. Abstract Expression detection is useful as a non-invasive method of lie detection and Wes Miller 5/11/2011 Comp Sci 534 Expression Detection in Video Abstract Expression detection is useful as a non-invasive method of lie detection and behavior prediction, as many facial expressions are

More information

Multi-Camera Calibration, Object Tracking and Query Generation

Multi-Camera Calibration, Object Tracking and Query Generation MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Multi-Camera Calibration, Object Tracking and Query Generation Porikli, F.; Divakaran, A. TR2003-100 August 2003 Abstract An automatic object

More information

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model TAE IN SEOL*, SUN-TAE CHUNG*, SUNHO KI**, SEONGWON CHO**, YUN-KWANG HONG*** *School of Electronic Engineering

More information

Visible and Long-Wave Infrared Image Fusion Schemes for Situational. Awareness

Visible and Long-Wave Infrared Image Fusion Schemes for Situational. Awareness Visible and Long-Wave Infrared Image Fusion Schemes for Situational Awareness Multi-Dimensional Digital Signal Processing Literature Survey Nathaniel Walker The University of Texas at Austin nathaniel.walker@baesystems.com

More information

Lane Markers Detection based on Consecutive Threshold Segmentation

Lane Markers Detection based on Consecutive Threshold Segmentation ISSN 1746-7659, England, UK Journal of Information and Computing Science Vol. 6, No. 3, 2011, pp. 207-212 Lane Markers Detection based on Consecutive Threshold Segmentation Huan Wang +, Mingwu Ren,Sulin

More information