Monitoring Head Dynamics for Driver Assistance Systems: A Multi-Perspective Approach

Size: px
Start display at page:

Download "Monitoring Head Dynamics for Driver Assistance Systems: A Multi-Perspective Approach"

Transcription

1 Proceedings of the 16th International IEEE Annual Conference on Intelligent Transportation Systems (ITSC 2013), The Hague, The Netherlands, October 6-9, 2013 WeC4.5 Monitoring Head Dynamics for Driver Assistance Systems: A Multi-Perspective Approach Sujitha Martin, Ashish Tawari and Mohan M. Trivedi Abstract A visually demanding driving environment, where elements surrounding a driver are constantly and rapidly changing, requires a driver to make spatially large head turns. Many existing state of the art vision based head pose algorithms, however, still have difficulties in continuously monitoring the head dynamics of a driver. This occurs because, from the perspective of a single camera, spatially large head turns induce self-occlusions of facial features, which are key elements in determining head pose. In this paper, we introduce a shape feature based multi-perspective framework for continuously monitoring the driver s head dynamics. The proposed approach utilizes a distributed camera setup to observe the driver over a wide range of head movements. Using head dynamics and a confidence measure based on symmetry of facial features, a particular perspective is chosen to provide the final head pose estimate. Our analysis on real world driving data shows promising results. I. INTRODUCTION In a 2011 American Community Survey (ACS) of 3- year estimates [1], it is reported that of the 138 million surveyed people, 86.2 % used a car, truck or van as means of transportation to work and on an average spent at least 25 minutes traveling to work daily. Meanwhile, the World Health Organization (WHO) predicted that by 2020, traffic related injuries would become the 3rd leading cause of global burden of diseases [2]. Even though faced with these alarming predictions, drivers will continue to drive to work and other recreational places, and be exposed to safety critical situations. However, injury-related collisions can be avoided or mitigated if the driver is notified when such situations arise. A comprehensive survey on automotive collisions demonstrated a driver was 31% less likely to cause an injuryrelated collision when he had one or more passengers who could alert him to unseen hazards [3]. Consequently, there is a great potential for intelligent driver assistance systems (IDAS) that can alert the driver about potential dangers or even guide them briefly through the critical situation. Accessing the source and level of danger in a driving situation, with a driver in the loop, is still a very challenging task and requires context-specific systems that monitor not only the outside environment but also the inside of the vehicle [4]. For example, if a vehicle is drifting from it s lane, looking inside the vehicle at the driver will show whether the driver is distracted, in which case an IDAS can alert the driver, or if he is intending to change lanes, in which case an IDAS can assist in changing lanes. Another application can be in the design of active heads-up display [5] where safety Laboratory of Intelligent and Safe Automobiles, UCSD, CA 92092, USA {scmartin, atawari, mtrivedi} at ucsd.edu critical information is efficiently presented to the driver, knowing his or her filed of view. In our present discussion, we focus on one of the integral components for monitoring driver awareness: analysis of head dynamics. It involves continuous estimation of head position and orientation/pose. These continuous estimations of head pose are necessary for applications like lane change intent prediction [6][7], which looks for particular patterns in head dynamics over time, and driver vigilance [8][9], which uses the driver s nodding frequency to monitor fatigue. In a recent publication describing the LISA-P head pose database from naturalistic on-road driving [10], Martin et al. showed using head pose, as taken from a motion capture system, that 95% of the time the driver s head pose was within 36 of forward facing in the yaw rotation angle. While it is true that a driver spends most of the time looking at the road directly in front, situations critical to driver safety usually occur when the driver makes spatially large head turns away from forward facing. Therefore, there is a need for evaluating over databases more concentrated on large head movements during naturalistic on-road driving, in order to better evaluate any algorithm for performance, robustness and reliability of continuous head tracking. To the best of our knowledge, there exists only one other study [11] that presents evaluations especially over the instances where driver s make spatially large head turns during naturalistic driving. In this paper, we introduce a vision based multiperspective framework for head pose estimation, which targets the issues presented by spatially large head turns. In the proposed framework, we show how to robustly combine the different perspectives while treating them separately. For evaluations, we utilize naturalistic on-road driving data. It is important to note that we choose segments of the naturalistic driving data around specific maneuvers which demand large head movement thereby presenting realistic requirement for continuous driver monitoring system. We present comparisons between single- and multi-perspective using mean-squared errors in yaw, pitch and roll rotation angles, and failure rate, percentage of time when system s output is unreliable. II. RELATED STUDIES In this section, we present some representative literature on vision based head pose estimation algorithms and show that these literatures explicitly or implicitly state their need for visibility of facial features to estimate head pose. This will emphasize that during spatially large head movements /$ IEEE 2286

2 TABLE I SUMMARY OF ISSUES WITH RESPECT TO MONITORING THE HEAD DYNAMICS DURING NATURALISTIC ON ROAD DRIVING FACED AND ADDRESSED (IF AT ALL) BY SINGLE AND MULTIPLE PERSPECTIVE SYSTEMS. Issues in monitoring the head dynamics of a driver Single-perspective Multi-perspective Wide changes in head pose in the yaw rotation angle Even with a wide field of view camera, the limitation in perspective from a single camera makes it susceptible Varying lighting conditions Occlusion of the driver s face (i.e. due to hand movements on the wheel, adjusting the sun shade) Any fixed placement of a single camera, depending on time of day and direction of car motion, will be affected by lighting conditions (e.g. washed out face) Highly susceptible Less susceptible with careful placement of multiple cameras Lower probability for all perspectives being severely affected Lower probability of all perspectives being severely affected. over a period of time, the lack of continuous visibility of facial features from the perspective of a single camera can be overcome with multiple camera perspectives. A recent survey by Muphy-Chutoria and Trivedi [12] gives a more thorough comparison on a variety of head pose estimation algorithms. Facial feature and geometric based algorithms is an intuitive way of estimating head pose which explicitly state the visibility requirement of facial features. Many of them require at least 5 to 6 specific features to be visible in the image plane of the camera and use geometric constraints such as face symmetry using lip corners and both eyes [13], anatomical structure of the eye and person-specific parameters [14], parallel lines between eye corners and between lip corners [15]. Yet another method, while not requiring specific facial features, requires at least 4 features in general positions on the face and uses POS [16] on corresponding points between image and mean face model to estimate head pose [10]. On the other hand, algorithms like HyHOPE [17], which uses a texture based 3D-model to represent the driver s head, implicitly require the visibility of facial features. The approach is to obtain an initial 3D model of the face, and then use rotation, translation and projection of the model to match views from video frames to estimate head pose. As claimed in [17], such algorithms are also unreliable during spatially large head movements due to self-occlusion of a majority of the face. The consequence of non-visible facial features from a single perspective camera during spatially large head movements has been suppressed because many contributions in literature have evaluated over datasets where a majority of the time, the driver is close to forward facing. In this paper, we explicitly address this concern and show results of improved performance in multi-perspective approach when compared against single-perspective. III. CHALLENGES IN SINGLE PERSPECTIVES Monitoring the head dynamics of a driver requires continuous estimation of head position and orientation. From a computer vision perspective, however, intrinsic properties of head dynamics present a challenge to the robustness of many existing algorithms. As discussed in Section II, many existing Fig. 1. Illustration of disadvantages present in single perspective systems. Each row of images shows images are from a particular camera location and each column of images are time-synchronized. Single perspective systems are susceptible to spatially large head movements (column 1), hand movements (column 2) and lighting conditions (column 3). state of the art head pose estimation algorithms rely on a portion of the face to be visible in the image plane to estimate head pose. This means that even during spatially large head movements, algorithms require the visibility of facial features to continuously track the state of the head. With a single perspective of the driver s head, however, large spatial head movements induce self-occlusions of facial features. The first column of Fig. 1 illustrates this issue. In Fig. 1, each row of images is taken from a different camera perspective and each column of images is time-synchronized. Clearly, the availability of multiple perspectives decreases the severity of self-occlusions at any instant in time which translates to an increase in the robustness of continuous head tracking. Occlusions of facial features can also occur due to external objects (e.g. hand movements near the face region, sunglasses). Depending on camera perspective, hand movements on steering wheel during vehicle maneuvers, to adjust sunshade, to point, etc. can cause occlusion. The second column in Fig. 1 shows an example of occlusion by hand due to the driver adjusting the sunshade. Another source of challenge for vision based driver head monitoring systems in naturalistic driving conditions is illumination changes. Effects of lighting conditions are also highly dependent on the camera location. In Fig. 1, the /$ IEEE 2287

3 Fig. 2. Time-synchronized images over a period of time from distributed camera setup around the driver. In the image series, the driver is moving his head from almost frontal facing to the left. last column illustrates the effects of lighting conditions. Therefore, multiple perspective approach with suitable camera placements, can mitigate the adverse effect of any one camera perspective being unreliable to track the head. Outlines of these issues and how, if at all, they are addressed under single-perspective systems and multiperspective systems are given in Table I. In our current implementation, our main focus is on improving head tracking accuracies for continuous driver monitoring system during spatially large head movements. IV. MULTI-PERSPECTIVE FRAMEWORK In the multi-perspective framework, each perspective is kept independent of one another and can be processed in parallel. To estimate head dynamics, we use a shape based head pose method to find local features, such as eye corners, nose corners and nose tip, and determine pose from the relative configuration of these features. A camera selection procedure then selects the appropriate perspective based on head dynamics and a confidence measure based on symmetry of facial features. A. Head pose estimation The head pose estimation algorithm used in the multiperspective framework is based on facial features and their respective geometry. At least 4 facial features points in general positions in the image plane and their corresponding points in a 3D generic mean face model, are required in this geometric approach. With these corresponding points, Pose from Orthography and Scaling (POS) [16] is used to estimate head pose. In our current implementation, manual annotation plus POS is used as a reference to evaluate Constrained Local Models (CLM) [18] plus POS in the multi-view selective framework. 1) CLM: In recent times, model-based techniques have been extensively used in nonrigid deformable object fitting. One such technique, CLM, is used for facial landmarks detection and tracking in a video data. CLM represents objects, in our case faces, using local appearance descriptions centered around landmarks of interest, and a parameterized shape model to capture plausible deformation of the the landmark locations. An ensemble of local detectors predicts the probable locations of the landmarks, called a response map. The true response map of the these detectors are approximated by various models such as Gaussian density with diagonal covariance [19], full covariance [20], Gaussian mixture model [21] etc. to make optimization efficient and numerically stable. In [18], the response map is represented non-parametrically using kernel density estimate (KDE) approach and the landmarks locations are optimized via subspace constrained meanshifts while enforcing their joint motion via shape model. CLM is generally used to fit faces of various orientations, positions and sizes in the image plane. However, when streaming images from a fixed camera looking at the driver s face while driving, there are constraints in orientation, position and size of the face that can be used to further improve facial feature tracking. For each perspective, we estimate the probable face location and face size. When the tracked features are not within the expected region or if the size of the face is beyond what is expected from that perspective, the tracked facial features are discarded and the system is re-initialized. Similarly, if the head pose computed from the detected facial features is not realizable in a driving scenario, it is also rejected. The rejected frames are accounted in one of the metrics (failure rate) to assess the performance of the system. 2) POS: Given a 3D model of an object, POS finds the position and orientation of the camera coordinate with respect to the object reference frame, such that when the 3D object is orthographically projected onto a plane parallel to the camera s image plane and scaled, it matches the image in the camera with minimal error. Given M i and p i = (x i, y i ) are known points in the 3D model and image plane, respectively, the system of equation to solve is as follows: M 0 M i αi = x 0 x i M 0 M i αj = y 0 y i /$ IEEE 2288

4 The notation M 0 M i represents the vector from the reference point M 0 to M i and alpha is the earlier mentioned scaling factor that occurs due perspective projection. The vectors i and j form the first two rows of the rotation matrix and the vector k = i j, a cross product. Note, however, that while k is perpendicular to i and j, vectors i and j are not necessarily perpendicular because of noisy 2D-3D correspondence. Therefore, the rotation matrix is projected into the SO3 space by normalizing the magnitude of the eigenvalues. A. Testbed and Dataset Data is collected from naturalistic, on-road driving using the LISA-A testbed [11]. There are three cameras observing the driver. One camera was placed on left most edge of the front windshield (near A-pillar) to get a better view of the driver s face when the driver turns his head leftward. Another camera was placed on the front windshield, close to the dashboard, to get the best view when driver is mostly near forward facing. The last camera was placed near the rearview mirror, for when the driver turns his head rightward. Most importantly, data from all three cameras were captured in a time-synchronized manner. Using this testbed, four drivers were asked to drive naturally on local streets and freeways near UCSD. Approximately 30 minutes of data was collected with each driver in sunny weather conditions, which allowed for varying lighting conditions. In addition, driving in an urban setting, the drivers passed through many stops signs and made multiple turns, and driving on the freeway allowed for multiple lane change occurrences, resulting in a dataset with wide spatial changes in head pose. TABLE II A LIST OF EVENTS CONSIDERED FOR EVALUATION, AND ITS RESPECTIVE COUNT AND NUMBER OF FRAMES. Fig. 3. A block diagram of the multi-view selective framework for head dynamics analysis. Inputs are images from multiple cameras. Head dynamics and a confidence based measure are used to select which perspective to use for head dynamics analysis. B. Perspective Selection algorithm In our current implementation, we use three cameras: one is placed to look straight at the driver, one looks at the driver from left and the last looks at the driver from the right, as shown in Fig. 2. When the estimated head pose, respective of camera s reference frame, in the current frame is outside of a defined range in the yaw rotation angle, the direction of the head movement is used to select the camera perspective for the next frame. If tracking becomes unreliable, the system is re-initialized with a camera based on the symmetry of the detected facial features. Higher symmetry score means detected face is closer to frontal pose in the given perspective. The block diagram in Fig. 3 illustrates this process and Fig. 4 illustrates this process as applied to data from naturalistic on-road driving test data. V. RESULTS AND DISCUSSION The proposed multi-perspective framework is evaluated on naturalistic on-road driving test data. We evaluate on events that induce spatially wide changes in the head pose of the driver. By evaluating on these select events, we show the usefulness of a multi-perspective in continuously tracking the head. Events No. of events Total no. of frames Right turns Left turns Stop sign Right Lane change Left Lane Change Merge All From the collected data, we select events during the drive that induced large head movements. These events are when the driver is making right/left turns, right/left lane changes, stops at stop signs, and freeway merges. Table II shows the events considered, the number of respective events analyzed during the total 120 minute drive containing all drivers, and the total number of frames accumulated for each event. The evaluations reported in the following section will be on these selected events. It is important to note that each event can induce more than one set of spatially wide head movements. Ground truth for evaluations is available in the form of manual facial feature annotation. The features annotated are eye corners, nose corners and nose tip. These features are chosen particularly because they are distinct features on the face and are less susceptible to deformation. The annotations are available once every 5 frames for all three camera perspectives for the events considered. These manual features annotations and their corresponding location on a 3D generic face model are used with POS to estimate head pose. In the following evaluation, we use this estimated head pose as a reference to compute the error in estimating head pose from automatic facial feature tracking. During annotations of the facial features, there were instances when less than /$ IEEE 2289

5 Fig. 4. Illustration of the multi-perspective framework. The graph at the bottom shows a segment taken from a subject s naturalist on-road driving experiment. The horizontal axis represents time and the vertical axis represents head rotations in the yaw rotation angle relative to a particular camera s reference frame. The blue asterisks represent the left camera, the red circles represent the center camera and the magenta crosses represent the right camera. The plot shows the relative head pose from the selected camera. Images from each perspective for three instances are shown while red boxes are drawn to show chosen camera perspective at respective instances. TABLE III E VALUATION OF THE MULTI - VIEW SELECTIVE FRAMEWORK. features were visible. Since we need at least 4 features in general positions to estimate head pose, we did not annotate frames where less than 4 features were visible and accounted for it in one of the metrics (failure rate). Multi-view Single-view B. On-road Performance Evaluation Evaluations are done in two parts. First, we test how well the head is tracked with the multi-perspective framework and compare with that which comes from a single perspective. In our evaluation, the single perspective refers to the output from the front facing camera and the multi-perspective includes all three camera perspectives described in the previous section. We determine that head tracking has failed when for the chosen camera perspective the estimated head pose is unavailable or if the estimated head pose is more than 30 from ground truth in the yaw rotation angle. Normalizing this count with the total number of frames over all events gives us the failure rate. As shown in Table III, the failure rate with multi-perspective is only 9.4% whereas it is 29% with the single-perspective. These percentages are important for many applications, such as lane change intent prediction, where large percentage of loss in head tracking can critically handicap a system. Therefore, smaller percentage of loss in head tracking is highly prefreable in safety critical situations like the driving environment /$ IEEE Mean Squared Erorr Pitch Yaw Roll Failure rate 9.4% 29.0% In the second part of the evaluation, we compute the mean square error (MSE) of the estimated head pose for multiperspective and for single-perspective. We see from Table III that the MSE is comparable between multi-perspective and single-perspective. VI. CONCLUDING REMARKS In a driving environment, the driver is prone to make large spatial head movements during maneuvers such as lane changes, right/left turns. At these crucial moments, it is important to continuously and reliably track the head of the driver. Since a single perspective of the driver is insufficient during spatially large head movements, we proposed a multi-perspective framework. In this framework, we use the estimation of head pose and head dynamics from the previous frame to select which perspective should be used in the current frame for head pose estimation; symmetry of facial 2290

6 features is used for initialization purposes. We evaluated this system on events during naturalistic on-road driving that induces large spatial head turns and showed that in the multiperspective framework the head tracking was reliable over 90% of the time whereas it was reliable less than 72% of the time with the single-perspective approach. In future studies, the multi-perspective framework can be furthered improved by considering other cues for fusing camera perspective. One possibility is in using the confidence of the detected facial features. Furthermore, it will be interesting to experiment whether having more/less than the current number of cameras will improve the tracking ability as well as decrease the MSE significantly. As for the vision based head pose algorithm, we used generic face model which artificially introduces error. In the future, we will pursue automatic adoption to an individual driver s face. ACKNOWLEDGMENT The authors would like to thank the support from UC Discovery and industry partners especially Audi and VW Electronic Research Laboratory. We also thank our UCSD LISA colleagues who helped in a variety of important ways in our research studies. [14] J. Chen and Q. Ji, 3d gaze estimation with a single camera without ir illumination, in IEEE International Conference on Pattern Recognition (ICPR), dec. 2008, pp [15] J.-G. Wang and E. Sung, Em enhancement of 3d head pose estimated by point at infinity, Image and Vision Computing, vol. 25, no. 12, pp , 2007, the age of human computer interaction. [16] D. F. Dementhon and L. S. Davis, Model-based object pose in 25 lines of code, International Journal of Computer Vision, vol. 15, pp , [17] E. Murphy-Chutorian and M. Trivedi, Head pose estimation and augmented reality tracking: An integrated system and evaluation for monitoring driver awareness, IEEE Transactions on Intelligent Transportation Systems, vol. 11, no. 2, pp , june [18] J. Saragih, S. Lucey, and J. Cohn, Face alignment through subspace constrained mean-shifts, in IEEE 12th International Conference on Computer Vision, oct , pp [19] T. Cootes and C.J.Taylor, Active shape models - smart snakes, in In British Machine Vision Conference, 1992, pp [20] Y. Wang, S. Lucey, and J. Cohn, Enforcing convexity for improved alignment with constrained local models, in IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), June 2008, pp [21] L. Gu and T. Kanade, A generative shape regularization model for robust face alignment, in Proceedings of the 10th European Conference on Computer Vision: Part I, ser. ECCV 08, 2008, pp REFERENCES [1] (2012, October) Means of transportation to work by selected characteristics, american community survey 3-year estimates. [Online]. Available: [2] M. Peden, R. Scurfield, D. Sleet, D. Mohan, A. Hyder, E. Jarawan, and C. Mathers, World report on road traffic injury prevention, World Health Organization, Tech. Rep., [Online]. Available: [3] E. Murphy-Chutorian and M. Trivedi, Hyhope: Hybrid head orientation and position estimation for vision-based driver head tracking, in IEEE Intelligent Vehicles Symposium, june 2008, pp [4] C. Tran and M. Trivedi, Driver assistance for keeping hands on the wheel and eyes on the road, in IEEE International Conference on Vehicular Electronics and Safety (ICVES), nov. 2009, pp [5] A. Doshi, S. Y. Cheng, and M. M. Trivedi, A novel active heads-up display for driver assistance, IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, vol. 39, no. 1, pp , [6] A. Doshi, B. Morris, and M. Trivedi, On-road prediction of driver s intent with multimodal sensory cues, IEEE Pervasive Computing, vol. 10, no. 3, pp , july-september [7] A. Doshi and M. Trivedi, On the roles of eye gaze and head dynamics in predicting driver s intent to change lanes, IEEE Transactions on Intelligent Transportation Systems, vol. 10, no. 3, pp , sept [8] L. Bergasa, J. Nuevo, M. Sotelo, R. Barea, and M. Lopez, Realtime system for monitoring driver vigilance, IEEE Transactions on Intelligent Transportation Systems, vol. 7, no. 1, pp , march [9] Z. Zhu and Q. Ji, Real time and non-intrusive driver fatigue monitoring, in IEEE International Conference on Intelligent Transportation Systems, oct. 2004, pp [10] S. Martin, A. Tawari, E.-M. Chutorian, S. Y. Cheng, and M. M. Trivedi, On the design and evaluation of robust head pose for visual user interfaces: Algorithms, databases, and comparisons, in 4th ACM SIGCHI International Conference on Automotive User Interfaces and Interactive Vehicular Applications (AUTO-UI), [11] A. Tawari and M. M. Trivedi, Head dynamic analysis: A multi-view framework, in Workshop on Social Behaviour Analysis, International Conference on Image Analysis and Processing, [12] E. Murphy-Chutorian and M. Trivedi, Head pose estimation in computer vision: A survey, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no. 4, pp , april [13] A. Gee and R. Cipolla, Determining the gaze of faces in images, Image and Vision Computing, vol. 12, no. 10, pp , /$ IEEE 2291

On the Design and Evaluation of Robust Head Pose for Visual User Interfaces: Algorithms, Databases, and Comparisons

On the Design and Evaluation of Robust Head Pose for Visual User Interfaces: Algorithms, Databases, and Comparisons On the Design and Evaluation of Robust Head Pose for Visual User Interfaces: Algorithms, Databases, and Comparisons Sujitha Martin scmartin@ucsd.edu Shinko Y. Cheng sycheng@hrl.com Ashish Tawari atawari@ucsd.edu

More information

On Driver Gaze Estimation: Explorations and Fusion of Geometric and Data Driven Approaches

On Driver Gaze Estimation: Explorations and Fusion of Geometric and Data Driven Approaches 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC) Windsor Oceanico Hotel, Rio de Janeiro, Brazil, November 1-4, 2016 On Driver Gaze Estimation: Explorations and Fusion

More information

Analyzing the Relationship Between Head Pose and Gaze to Model Driver Visual Attention

Analyzing the Relationship Between Head Pose and Gaze to Model Driver Visual Attention Analyzing the Relationship Between Head Pose and Gaze to Model Driver Visual Attention Sumit Jha and Carlos Busso Multimodal Signal Processing (MSP) Laboratory Department of Electrical Engineering, The

More information

818 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 15, NO. 2, APRIL 2014

818 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 15, NO. 2, APRIL 2014 818 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 15, NO. 2, APRIL 2014 Continuous Head Movement Estimator for Driver Assistance: Issues, Algorithms, and On-Road Evaluations Ashish Tawari,

More information

Occlusion Robust Multi-Camera Face Tracking

Occlusion Robust Multi-Camera Face Tracking Occlusion Robust Multi-Camera Face Tracking Josh Harguess, Changbo Hu, J. K. Aggarwal Computer & Vision Research Center / Department of ECE The University of Texas at Austin harguess@utexas.edu, changbo.hu@gmail.com,

More information

Full-Motion Recovery from Multiple Video Cameras Applied to Face Tracking and Recognition

Full-Motion Recovery from Multiple Video Cameras Applied to Face Tracking and Recognition Full-Motion Recovery from Multiple Video Cameras Applied to Face Tracking and Recognition Josh Harguess, Changbo Hu, J. K. Aggarwal Computer & Vision Research Center / Department of ECE The University

More information

Towards Automated Drive Analysis: A Multimodal Synergistic Approach

Towards Automated Drive Analysis: A Multimodal Synergistic Approach Proceedings of the 6th International IEEE Annual Conference on Intelligent Transportation Systems (ITSC 23), The Hague, The Netherlands, October 6-9, 23 WeB.2 Towards Automated Drive Analysis: A Multimodal

More information

Eye Detection by Haar wavelets and cascaded Support Vector Machine

Eye Detection by Haar wavelets and cascaded Support Vector Machine Eye Detection by Haar wavelets and cascaded Support Vector Machine Vishal Agrawal B.Tech 4th Year Guide: Simant Dubey / Amitabha Mukherjee Dept of Computer Science and Engineering IIT Kanpur - 208 016

More information

Generic Face Alignment Using an Improved Active Shape Model

Generic Face Alignment Using an Improved Active Shape Model Generic Face Alignment Using an Improved Active Shape Model Liting Wang, Xiaoqing Ding, Chi Fang Electronic Engineering Department, Tsinghua University, Beijing, China {wanglt, dxq, fangchi} @ocrserv.ee.tsinghua.edu.cn

More information

Available online at ScienceDirect. Procedia Computer Science 50 (2015 )

Available online at   ScienceDirect. Procedia Computer Science 50 (2015 ) Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 50 (2015 ) 617 622 2nd International Symposium on Big Data and Cloud Computing (ISBCC 15) Monitoring Driver Head Postures

More information

Snap-DAS: A Vision-based Driver Assistance System on a Snapdragon TM Embedded Platform

Snap-DAS: A Vision-based Driver Assistance System on a Snapdragon TM Embedded Platform Snap-DAS: A Vision-based Driver Assistance System on a Snapdragon TM Embedded Platform Ravi Kumar Satzoda, Sean Lee, Frankie Lu, and Mohan M. Trivedi Abstract In the recent years, mobile computing platforms

More information

Gaze interaction (2): models and technologies

Gaze interaction (2): models and technologies Gaze interaction (2): models and technologies Corso di Interazione uomo-macchina II Prof. Giuseppe Boccignone Dipartimento di Scienze dell Informazione Università di Milano boccignone@dsi.unimi.it http://homes.dsi.unimi.it/~boccignone/l

More information

Vision-based Lane Analysis: Exploration of Issues and Approaches for Embedded Realization

Vision-based Lane Analysis: Exploration of Issues and Approaches for Embedded Realization 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops Vision-based Lane Analysis: Exploration of Issues and Approaches for Embedded Realization R. K. Satzoda and Mohan M. Trivedi Computer

More information

Real Time Face Tracking and Pose Estimation Using an Adaptive Correlation Filter for Human-Robot Interaction

Real Time Face Tracking and Pose Estimation Using an Adaptive Correlation Filter for Human-Robot Interaction Real Time Face Tracking and Pose Estimation Using an Adaptive Correlation Filter for Human-Robot Interaction Vo Duc My and Andreas Zell Abstract In this paper, we present a real time algorithm for mobile

More information

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,

More information

3D Active Appearance Model for Aligning Faces in 2D Images

3D Active Appearance Model for Aligning Faces in 2D Images 3D Active Appearance Model for Aligning Faces in 2D Images Chun-Wei Chen and Chieh-Chih Wang Abstract Perceiving human faces is one of the most important functions for human robot interaction. The active

More information

Intensity-Depth Face Alignment Using Cascade Shape Regression

Intensity-Depth Face Alignment Using Cascade Shape Regression Intensity-Depth Face Alignment Using Cascade Shape Regression Yang Cao 1 and Bao-Liang Lu 1,2 1 Center for Brain-like Computing and Machine Intelligence Department of Computer Science and Engineering Shanghai

More information

Face Alignment Robust to Occlusion

Face Alignment Robust to Occlusion Face Alignment Robust to Occlusion Anonymous Abstract In this paper we present a new approach to robustly align facial features to a face image even when the face is partially occluded Previous methods

More information

Real Time Eye s Off the Road Detection System for Driver Assistance

Real Time Eye s Off the Road Detection System for Driver Assistance Real Time Eye s Off the Road Detection System for Driver Assistance Manthri Lavanya & C.Ravi Teja 1 M.Tech,Assistant Professor, Department of ECE, Anantha Lakshmi Institute of Technology, Anantapur, Ap,

More information

Baseline Face Detection, Head Pose Estimation, and Coarse Direction Detection for Facial Data in the SHRP2 Naturalistic Driving Study

Baseline Face Detection, Head Pose Estimation, and Coarse Direction Detection for Facial Data in the SHRP2 Naturalistic Driving Study Baseline Face Detection, Head Pose Estimation, and Coarse Direction Detection for Facial Data in the SHRP2 Naturalistic Driving Study J. Paone, D. Bolme, R. Ferrell, Member, IEEE, D. Aykac, and T. Karnowski,

More information

Real Time Eye s off the Road Detection System for Driver Assistance

Real Time Eye s off the Road Detection System for Driver Assistance Real Time Eye s off the Road Detection System for Driver Assistance Bandi Ramagopal M. Tech Student, Department of ECE, Shri Sai Institute of Engineering and Technology, Anantapur (dt) A.P. Abstract Automated

More information

Abstract. 1 Introduction. 2 Motivation. Information and Communication Engineering October 29th 2010

Abstract. 1 Introduction. 2 Motivation. Information and Communication Engineering October 29th 2010 Information and Communication Engineering October 29th 2010 A Survey on Head Pose Estimation from Low Resolution Image Sato Laboratory M1, 48-106416, Isarun CHAMVEHA Abstract Recognizing the head pose

More information

AAM Based Facial Feature Tracking with Kinect

AAM Based Facial Feature Tracking with Kinect BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 15, No 3 Sofia 2015 Print ISSN: 1311-9702; Online ISSN: 1314-4081 DOI: 10.1515/cait-2015-0046 AAM Based Facial Feature Tracking

More information

Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images

Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images 1 Introduction - Steve Chuang and Eric Shan - Determining object orientation in images is a well-established topic

More information

Chapter 9 Object Tracking an Overview

Chapter 9 Object Tracking an Overview Chapter 9 Object Tracking an Overview The output of the background subtraction algorithm, described in the previous chapter, is a classification (segmentation) of pixels into foreground pixels (those belonging

More information

The Template Update Problem

The Template Update Problem The Template Update Problem Iain Matthews, Takahiro Ishikawa, and Simon Baker The Robotics Institute Carnegie Mellon University Abstract Template tracking dates back to the 1981 Lucas-Kanade algorithm.

More information

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices Technical University of Cluj-Napoca Image Processing and Pattern Recognition Research Center www.cv.utcluj.ro Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving

More information

Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing

Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material Introduction In this supplementary material, Section 2 details the 3D annotation for CAD models and real

More information

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Literature Survey Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Omer Shakil Abstract This literature survey compares various methods

More information

Vehicle Occupant Posture Analysis Using Voxel Data

Vehicle Occupant Posture Analysis Using Voxel Data Ninth World Congress on Intelligent Transport Systems, Chicago, Illinois, October Vehicle Occupant Posture Analysis Using Voxel Data Ivana Mikic, Mohan Trivedi Computer Vision and Robotics Research Laboratory

More information

Canny Edge Based Self-localization of a RoboCup Middle-sized League Robot

Canny Edge Based Self-localization of a RoboCup Middle-sized League Robot Canny Edge Based Self-localization of a RoboCup Middle-sized League Robot Yoichi Nakaguro Sirindhorn International Institute of Technology, Thammasat University P.O. Box 22, Thammasat-Rangsit Post Office,

More information

On Road Vehicle Detection using Shadows

On Road Vehicle Detection using Shadows On Road Vehicle Detection using Shadows Gilad Buchman Grasp Lab, Department of Computer and Information Science School of Engineering University of Pennsylvania, Philadelphia, PA buchmag@seas.upenn.edu

More information

Laserscanner Based Cooperative Pre-Data-Fusion

Laserscanner Based Cooperative Pre-Data-Fusion Laserscanner Based Cooperative Pre-Data-Fusion 63 Laserscanner Based Cooperative Pre-Data-Fusion F. Ahlers, Ch. Stimming, Ibeo Automobile Sensor GmbH Abstract The Cooperative Pre-Data-Fusion is a novel

More information

A Robust Method of Facial Feature Tracking for Moving Images

A Robust Method of Facial Feature Tracking for Moving Images A Robust Method of Facial Feature Tracking for Moving Images Yuka Nomura* Graduate School of Interdisciplinary Information Studies, The University of Tokyo Takayuki Itoh Graduate School of Humanitics and

More information

Vehicle Dimensions Estimation Scheme Using AAM on Stereoscopic Video

Vehicle Dimensions Estimation Scheme Using AAM on Stereoscopic Video Workshop on Vehicle Retrieval in Surveillance (VRS) in conjunction with 2013 10th IEEE International Conference on Advanced Video and Signal Based Surveillance Vehicle Dimensions Estimation Scheme Using

More information

Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material

Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material Chi Li, M. Zeeshan Zia 2, Quoc-Huy Tran 2, Xiang Yu 2, Gregory D. Hager, and Manmohan Chandraker 2 Johns

More information

Sparse Shape Registration for Occluded Facial Feature Localization

Sparse Shape Registration for Occluded Facial Feature Localization Shape Registration for Occluded Facial Feature Localization Fei Yang, Junzhou Huang and Dimitris Metaxas Abstract This paper proposes a sparsity driven shape registration method for occluded facial feature

More information

Multi-Camera Occlusion and Sudden-Appearance-Change Detection Using Hidden Markovian Chains

Multi-Camera Occlusion and Sudden-Appearance-Change Detection Using Hidden Markovian Chains 1 Multi-Camera Occlusion and Sudden-Appearance-Change Detection Using Hidden Markovian Chains Xudong Ma Pattern Technology Lab LLC, U.S.A. Email: xma@ieee.org arxiv:1610.09520v1 [cs.cv] 29 Oct 2016 Abstract

More information

Real-Time Detection of Road Markings for Driving Assistance Applications

Real-Time Detection of Road Markings for Driving Assistance Applications Real-Time Detection of Road Markings for Driving Assistance Applications Ioana Maria Chira, Ancuta Chibulcutean Students, Faculty of Automation and Computer Science Technical University of Cluj-Napoca

More information

A 3D experimental framework for exploring drivers body activity using infrared depth sensors

A 3D experimental framework for exploring drivers body activity using infrared depth sensors A 3D experimental framework for exploring drivers body activity using infrared depth sensors Alexandra Kondyli Department of Civil and Coastal Engineering University of Florida Gainesville, Florida 32611

More information

FACE RECOGNITION USING INDEPENDENT COMPONENT

FACE RECOGNITION USING INDEPENDENT COMPONENT Chapter 5 FACE RECOGNITION USING INDEPENDENT COMPONENT ANALYSIS OF GABORJET (GABORJET-ICA) 5.1 INTRODUCTION PCA is probably the most widely used subspace projection technique for face recognition. A major

More information

Driver s Behavior Assessment by On-board/Off-board Video Context Analysis

Driver s Behavior Assessment by On-board/Off-board Video Context Analysis Driver s Behavior Assessment by On-board/Off-board Video Context Analysis Lorenzo Ciardelli 1,, Andrea Beoldo 2, Francesco Pasini 1, and Carlo Regazzoni 1 1 Department of Biophysical and Electronic Engineering,

More information

CONTRIBUTION TO THE INVESTIGATION OF STOPPING SIGHT DISTANCE IN THREE-DIMENSIONAL SPACE

CONTRIBUTION TO THE INVESTIGATION OF STOPPING SIGHT DISTANCE IN THREE-DIMENSIONAL SPACE National Technical University of Athens School of Civil Engineering Department of Transportation Planning and Engineering Doctoral Dissertation CONTRIBUTION TO THE INVESTIGATION OF STOPPING SIGHT DISTANCE

More information

Face detection in a video sequence - a temporal approach

Face detection in a video sequence - a temporal approach Face detection in a video sequence - a temporal approach K. Mikolajczyk R. Choudhury C. Schmid INRIA Rhône-Alpes GRAVIR-CNRS, 655 av. de l Europe, 38330 Montbonnot, France {Krystian.Mikolajczyk,Ragini.Choudhury,Cordelia.Schmid}@inrialpes.fr

More information

HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION

HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION Dipankar Das Department of Information and Communication Engineering, University of Rajshahi, Rajshahi-6205, Bangladesh ABSTRACT Real-time

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H.

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H. Nonrigid Surface Modelling and Fast Recovery Zhu Jianke Supervisor: Prof. Michael R. Lyu Committee: Prof. Leo J. Jia and Prof. K. H. Wong Department of Computer Science and Engineering May 11, 2007 1 2

More information

In Between 3D Active Appearance Models and 3D Morphable Models

In Between 3D Active Appearance Models and 3D Morphable Models In Between 3D Active Appearance Models and 3D Morphable Models Jingu Heo and Marios Savvides Biometrics Lab, CyLab Carnegie Mellon University Pittsburgh, PA 15213 jheo@cmu.edu, msavvid@ri.cmu.edu Abstract

More information

Published in: IEEE 17th International Conference on Intelligent Transportation Systems (ITSC), 2014

Published in: IEEE 17th International Conference on Intelligent Transportation Systems (ITSC), 2014 Aalborg Universitet Attention estimation by simultaneous analysis of viewer and view Tawari, Ashish; Møgelmose, Andreas; Martin, Sujitha; Moeslund, Thomas B.; Trivedi, Mohan M. Published in: IEEE 17th

More information

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation Face Tracking Amit K. Roy-Chowdhury and Yilei Xu Department of Electrical Engineering, University of California, Riverside, CA 92521, USA {amitrc,yxu}@ee.ucr.edu Synonyms Facial Motion Estimation Definition

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Contexts and 3D Scenes

Contexts and 3D Scenes Contexts and 3D Scenes Computer Vision Jia-Bin Huang, Virginia Tech Many slides from D. Hoiem Administrative stuffs Final project presentation Nov 30 th 3:30 PM 4:45 PM Grading Three senior graders (30%)

More information

Model Based Techniques for Tracking Human Head for Human Robot Interaction

Model Based Techniques for Tracking Human Head for Human Robot Interaction Model Based Techniques for Tracking Human Head for Human Robot Interaction Faisal Khan University of Wisconsin, Madison faisal@cs.wisc.edu Varghese Mathew University of Wisconsin, Madison vmathew@cs.wisc.edu

More information

Performance Evaluation Metrics and Statistics for Positional Tracker Evaluation

Performance Evaluation Metrics and Statistics for Positional Tracker Evaluation Performance Evaluation Metrics and Statistics for Positional Tracker Evaluation Chris J. Needham and Roger D. Boyle School of Computing, The University of Leeds, Leeds, LS2 9JT, UK {chrisn,roger}@comp.leeds.ac.uk

More information

DEPTH AND GEOMETRY FROM A SINGLE 2D IMAGE USING TRIANGULATION

DEPTH AND GEOMETRY FROM A SINGLE 2D IMAGE USING TRIANGULATION 2012 IEEE International Conference on Multimedia and Expo Workshops DEPTH AND GEOMETRY FROM A SINGLE 2D IMAGE USING TRIANGULATION Yasir Salih and Aamir S. Malik, Senior Member IEEE Centre for Intelligent

More information

Occluded Facial Expression Tracking

Occluded Facial Expression Tracking Occluded Facial Expression Tracking Hugo Mercier 1, Julien Peyras 2, and Patrice Dalle 1 1 Institut de Recherche en Informatique de Toulouse 118, route de Narbonne, F-31062 Toulouse Cedex 9 2 Dipartimento

More information

Light Field Occlusion Removal

Light Field Occlusion Removal Light Field Occlusion Removal Shannon Kao Stanford University kaos@stanford.edu Figure 1: Occlusion removal pipeline. The input image (left) is part of a focal stack representing a light field. Each image

More information

Motion Estimation for Video Coding Standards

Motion Estimation for Video Coding Standards Motion Estimation for Video Coding Standards Prof. Ja-Ling Wu Department of Computer Science and Information Engineering National Taiwan University Introduction of Motion Estimation The goal of video compression

More information

Expression Detection in Video. Abstract Expression detection is useful as a non-invasive method of lie detection and

Expression Detection in Video. Abstract Expression detection is useful as a non-invasive method of lie detection and Wes Miller 5/11/2011 Comp Sci 534 Expression Detection in Video Abstract Expression detection is useful as a non-invasive method of lie detection and behavior prediction, as many facial expressions are

More information

Factorization Method Using Interpolated Feature Tracking via Projective Geometry

Factorization Method Using Interpolated Feature Tracking via Projective Geometry Factorization Method Using Interpolated Feature Tracking via Projective Geometry Hideo Saito, Shigeharu Kamijima Department of Information and Computer Science, Keio University Yokohama-City, 223-8522,

More information

Factorization with Missing and Noisy Data

Factorization with Missing and Noisy Data Factorization with Missing and Noisy Data Carme Julià, Angel Sappa, Felipe Lumbreras, Joan Serrat, and Antonio López Computer Vision Center and Computer Science Department, Universitat Autònoma de Barcelona,

More information

WeVe: When Smart Wearables Meet Intelligent Vehicles

WeVe: When Smart Wearables Meet Intelligent Vehicles WeVe: When Smart Wearables Meet Intelligent Vehicles Jiajia Liu School of Cyber Engineering, Xidian University, Xi an, China Smart wearables and intelligent vehicles constitute indispensable parts of Internet

More information

THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA

THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA Hwadong Sun 1, Dong Yeop Kim 1 *, Joon Ho Kwon 2, Bong-Seok Kim 1, and Chang-Woo Park 1 1 Intelligent Robotics Research Center,

More information

Model Based Perspective Inversion

Model Based Perspective Inversion Model Based Perspective Inversion A. D. Worrall, K. D. Baker & G. D. Sullivan Intelligent Systems Group, Department of Computer Science, University of Reading, RG6 2AX, UK. Anthony.Worrall@reading.ac.uk

More information

Review of distracted driving factors. George Yannis, Associate Professor, NTUA

Review of distracted driving factors. George Yannis, Associate Professor, NTUA Review of distracted driving factors George Yannis, Associate Professor, NTUA www.nrso.ntua.gr/geyannis Research Objectives Presentation Structure Research Objectives To provide a comprehensive picture

More information

Definition, Detection, and Evaluation of Meeting Events in Airport Surveillance Videos

Definition, Detection, and Evaluation of Meeting Events in Airport Surveillance Videos Definition, Detection, and Evaluation of Meeting Events in Airport Surveillance Videos Sung Chun Lee, Chang Huang, and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA sungchun@usc.edu,

More information

Map Guided Lane Detection Alexander Döbert 1,2, Andre Linarth 1,2, Eva Kollorz 2

Map Guided Lane Detection Alexander Döbert 1,2, Andre Linarth 1,2, Eva Kollorz 2 Map Guided Lane Detection Alexander Döbert 1,2, Andre Linarth 1,2, Eva Kollorz 2 1 Elektrobit Automotive GmbH, Am Wolfsmantel 46, 91058 Erlangen, Germany {AndreGuilherme.Linarth, Alexander.Doebert}@elektrobit.com

More information

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level

More information

Evaluation of Moving Object Tracking Techniques for Video Surveillance Applications

Evaluation of Moving Object Tracking Techniques for Video Surveillance Applications International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347 5161 2015INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Research Article Evaluation

More information

Face Alignment Under Various Poses and Expressions

Face Alignment Under Various Poses and Expressions Face Alignment Under Various Poses and Expressions Shengjun Xin and Haizhou Ai Computer Science and Technology Department, Tsinghua University, Beijing 100084, China ahz@mail.tsinghua.edu.cn Abstract.

More information

Adaptive Skin Color Classifier for Face Outline Models

Adaptive Skin Color Classifier for Face Outline Models Adaptive Skin Color Classifier for Face Outline Models M. Wimmer, B. Radig, M. Beetz Informatik IX, Technische Universität München, Germany Boltzmannstr. 3, 87548 Garching, Germany [wimmerm, radig, beetz]@informatik.tu-muenchen.de

More information

Final Exam Study Guide

Final Exam Study Guide Final Exam Study Guide Exam Window: 28th April, 12:00am EST to 30th April, 11:59pm EST Description As indicated in class the goal of the exam is to encourage you to review the material from the course.

More information

Information and Communication Engineering July 9th Survey of Techniques for Head Pose Estimation from Low Resolution Images

Information and Communication Engineering July 9th Survey of Techniques for Head Pose Estimation from Low Resolution Images Information and Communication Engineering July 9th 2010 Survey of Techniques for Head Pose Estimation from Low Resolution Images Sato Laboratory D1, 48-107409, Teera SIRITEERAKUL Abstract Head pose estimation

More information

Spatio-Temporal Stereo Disparity Integration

Spatio-Temporal Stereo Disparity Integration Spatio-Temporal Stereo Disparity Integration Sandino Morales and Reinhard Klette The.enpeda.. Project, The University of Auckland Tamaki Innovation Campus, Auckland, New Zealand pmor085@aucklanduni.ac.nz

More information

A REAL-TIME FACIAL FEATURE BASED HEAD TRACKER

A REAL-TIME FACIAL FEATURE BASED HEAD TRACKER A REAL-TIME FACIAL FEATURE BASED HEAD TRACKER Jari Hannuksela, Janne Heikkilä and Matti Pietikäinen {jari.hannuksela, jth, mkp}@ee.oulu.fi Machine Vision Group, Infotech Oulu P.O. Box 45, FIN-914 University

More information

Digital Makeup Face Generation

Digital Makeup Face Generation Digital Makeup Face Generation Wut Yee Oo Mechanical Engineering Stanford University wutyee@stanford.edu Abstract Make up applications offer photoshop tools to get users inputs in generating a make up

More information

Real-time Detection of Illegally Parked Vehicles Using 1-D Transformation

Real-time Detection of Illegally Parked Vehicles Using 1-D Transformation Real-time Detection of Illegally Parked Vehicles Using 1-D Transformation Jong Taek Lee, M. S. Ryoo, Matthew Riley, and J. K. Aggarwal Computer & Vision Research Center Dept. of Electrical & Computer Engineering,

More information

DSRC Field Trials Whitepaper

DSRC Field Trials Whitepaper DSRC Field Trials Whitepaper August 19, 2017 www.cohdawireless.com 1 Overview Cohda has performed more than 300 Dedicated Short Range Communications (DSRC) field trials, comparing DSRC radios from multiple

More information

Optimizing Monocular Cues for Depth Estimation from Indoor Images

Optimizing Monocular Cues for Depth Estimation from Indoor Images Optimizing Monocular Cues for Depth Estimation from Indoor Images Aditya Venkatraman 1, Sheetal Mahadik 2 1, 2 Department of Electronics and Telecommunication, ST Francis Institute of Technology, Mumbai,

More information

Cohda Wireless White Paper DSRC Field Trials

Cohda Wireless White Paper DSRC Field Trials Cohda Wireless White Paper DSRC Field Trials Copyright Cohda Wireless Pty Ltd ABN 84 107 936 309 Cohda Wireless Pty Ltd 82-84 Melbourne Street North Adelaide, SA 5006 Australia P +61 8 8364 4719 F +61

More information

Facial Expression Analysis

Facial Expression Analysis Facial Expression Analysis Jeff Cohn Fernando De la Torre Human Sensing Laboratory Tutorial Looking @ People June 2012 Facial Expression Analysis F. De la Torre/J. Cohn Looking @ People (CVPR-12) 1 Outline

More information

From Gaze to Focus of Attention

From Gaze to Focus of Attention From Gaze to Focus of Attention Rainer Stiefelhagen, Michael Finke, Jie Yang, Alex Waibel stiefel@ira.uka.de, finkem@cs.cmu.edu, yang+@cs.cmu.edu, ahw@cs.cmu.edu Interactive Systems Laboratories University

More information

Tracking driver actions and guiding phone usage for safer driving. Hongyu Li Jan 25, 2018

Tracking driver actions and guiding phone usage for safer driving. Hongyu Li Jan 25, 2018 Tracking driver actions and guiding phone usage for safer driving Hongyu Li Jan 25, 2018 1 Smart devices risks and opportunities Phone in use 14% Other distractions 86% Distraction-Affected Fatalities

More information

Stable Vision-Aided Navigation for Large-Area Augmented Reality

Stable Vision-Aided Navigation for Large-Area Augmented Reality Stable Vision-Aided Navigation for Large-Area Augmented Reality Taragay Oskiper, Han-Pang Chiu, Zhiwei Zhu Supun Samarasekera, Rakesh Teddy Kumar Vision and Robotics Laboratory SRI-International Sarnoff,

More information

Saliency Extraction for Gaze-Contingent Displays

Saliency Extraction for Gaze-Contingent Displays In: Workshop on Organic Computing, P. Dadam, M. Reichert (eds.), Proceedings of the 34th GI-Jahrestagung, Vol. 2, 646 650, Ulm, September 2004. Saliency Extraction for Gaze-Contingent Displays Martin Böhme,

More information

Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement

Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement Daegeon Kim Sung Chun Lee Institute for Robotics and Intelligent Systems University of Southern

More information

A Survey of Light Source Detection Methods

A Survey of Light Source Detection Methods A Survey of Light Source Detection Methods Nathan Funk University of Alberta Mini-Project for CMPUT 603 November 30, 2003 Abstract This paper provides an overview of the most prominent techniques for light

More information

Using temporal seeding to constrain the disparity search range in stereo matching

Using temporal seeding to constrain the disparity search range in stereo matching Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department

More information

A Robust Facial Feature Point Tracker using Graphical Models

A Robust Facial Feature Point Tracker using Graphical Models A Robust Facial Feature Point Tracker using Graphical Models Serhan Coşar, Müjdat Çetin, Aytül Erçil Sabancı University Faculty of Engineering and Natural Sciences Orhanlı- Tuzla, 34956 İstanbul, TURKEY

More information

Fast Road Sign Detection Using Hough Transform for Assisted Driving of Road Vehicles

Fast Road Sign Detection Using Hough Transform for Assisted Driving of Road Vehicles Fast Road Sign Detection Using Hough Transform for Assisted Driving of Road Vehicles Miguel Ángel García-Garrido, Miguel Ángel Sotelo, and Ernesto Martín-Gorostiza Department of Electronics, University

More information

Social Behavior Prediction Through Reality Mining

Social Behavior Prediction Through Reality Mining Social Behavior Prediction Through Reality Mining Charlie Dagli, William Campbell, Clifford Weinstein Human Language Technology Group MIT Lincoln Laboratory This work was sponsored by the DDR&E / RRTO

More information

DISTANCE MAPS: A ROBUST ILLUMINATION PREPROCESSING FOR ACTIVE APPEARANCE MODELS

DISTANCE MAPS: A ROBUST ILLUMINATION PREPROCESSING FOR ACTIVE APPEARANCE MODELS DISTANCE MAPS: A ROBUST ILLUMINATION PREPROCESSING FOR ACTIVE APPEARANCE MODELS Sylvain Le Gallou*, Gaspard Breton*, Christophe Garcia*, Renaud Séguier** * France Telecom R&D - TECH/IRIS 4 rue du clos

More information

Embedded ECG Based Real Time Monitoring and Control of Driver Drowsiness Condition

Embedded ECG Based Real Time Monitoring and Control of Driver Drowsiness Condition International Journal of Science, Technology and Society 2015; 3(4): 146-150 Published online June 15, 2015 (http://www.sciencepublishinggroup.com/j/ijsts) doi: 10.11648/j.ijsts.20150304.17 ISSN: 2330-7412

More information

A Street Scene Surveillance System for Moving Object Detection, Tracking and Classification

A Street Scene Surveillance System for Moving Object Detection, Tracking and Classification A Street Scene Surveillance System for Moving Object Detection, Tracking and Classification Huei-Yung Lin * and Juang-Yu Wei Department of Electrical Engineering National Chung Cheng University Chia-Yi

More information

Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information

Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information Mustafa Berkay Yilmaz, Hakan Erdogan, Mustafa Unel Sabanci University, Faculty of Engineering and Natural

More information

Automatic Shadow Removal by Illuminance in HSV Color Space

Automatic Shadow Removal by Illuminance in HSV Color Space Computer Science and Information Technology 3(3): 70-75, 2015 DOI: 10.13189/csit.2015.030303 http://www.hrpub.org Automatic Shadow Removal by Illuminance in HSV Color Space Wenbo Huang 1, KyoungYeon Kim

More information

Contexts and 3D Scenes

Contexts and 3D Scenes Contexts and 3D Scenes Computer Vision Jia-Bin Huang, Virginia Tech Many slides from D. Hoiem Administrative stuffs Final project presentation Dec 1 st 3:30 PM 4:45 PM Goodwin Hall Atrium Grading Three

More information

Disguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network. Nathan Sun CIS601

Disguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network. Nathan Sun CIS601 Disguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network Nathan Sun CIS601 Introduction Face ID is complicated by alterations to an individual s appearance Beard,

More information

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems Abstract In this paper we present a method for mirror shape recovery and partial calibration for non-central catadioptric

More information

Time-to-Contact from Image Intensity

Time-to-Contact from Image Intensity Time-to-Contact from Image Intensity Yukitoshi Watanabe Fumihiko Sakaue Jun Sato Nagoya Institute of Technology Gokiso, Showa, Nagoya, 466-8555, Japan {yukitoshi@cv.,sakaue@,junsato@}nitech.ac.jp Abstract

More information