Human Robot Communication. Paul Fitzpatrick
|
|
- Chrystal Lloyd
- 5 years ago
- Views:
Transcription
1 Human Robot Communication Paul Fitzpatrick
2 Human Robot Communication Motivation for communication Human-readable actions Reading human actions Conclusions
3 Motivation What is communication for? Transferring information Coordinating behavior What is it built from? Commonality Perception of action Protocols
4 Communication protocols Computer computer protocols TCP/IP, HTTP, FTP, SMTP,
5 Communication protocols Human human protocols Initiating conversation, turn-taking, interrupting, directing attention, Human computer protocols Shell interaction, drag-and-drop, dialog boxes,
6 Communication protocols Human human protocols Initiating conversation, turn-taking, interrupting, directing attention, Human robot protocols Human computer protocols Shell interaction, drag-and-drop, dialog boxes,
7 Requirements on robot ENGAGED ACQUIRED Pointing (53,92,12) Fixating (47,98,37) Saying /o ver[200] \/there[325] Human-oriented perception Person detection, tracking Pose estimation Identity recognition Expression classification Speech/prosody recognition Objects of human interest Human-readable action Clear locus of attention Express engagement Express confusion, surprise Speech/prosody generation
8 Example: attention protocol Expressing attention Influencing other s attention Reading other s attention
9 Foveate gaze Motivation for communication Human-readable actions Reading human actions Conclusions
10 Human gaze reflects attention (Taken from C. Graham, Vision and Visual Perception )
11 Types of eye movement Ballistic saccade to new target Right eye Left eye Vergence angle Smooth pursuit and vergence co-operate to track object (Based on Kandel & Schwartz, Principles of Neural Science )
12 Engineering gaze Kismet
13 Collaborative effort Cynthia Breazeal Brian Scassellati And others Will describe components I m responsible for
14 Engineering gaze
15 Engineering gaze Cyclopean camera Stereo pair
16 Tip-toeing around 3D Field of view New field of view Object of interest Wide View camera Narrow view camera Rotate camera
17 Example
18 Influences on attention
19 Influences on attention Built in biases
20 Influences on attention Built in biases Behavioral state
21 Influences on attention Built in biases Behavioral state Persistence slipped recovered
22 Directing attention
23 Head pose estimation Motivation for communication Human-readable actions Reading human actions Conclusions
24 Head pose estimation (rigid) Yaw * Pitch * Roll * Translation in X, Y, Z * Nomenclature varies
25 Head pose literature Horprasert, Yacoob, Davis 97 McKenna, Gong 98 Wang, Brandstein 98 Basu, Essa, Pentland 96 Harville, Darrell, et al 99
26 Head pose: Anthropometrics Horprasert, Yacoob, Davis McKenna, Gong Wang, Brandstein Basu, Essa, Pentland Harville, Darrell, et al
27 Head pose: Eigenpose Horprasert, Yacoob, Davis McKenna, Gong Wang, Brandstein Basu, Essa, Pentland Harville, Darrell, et al
28 Head pose: Contours Horprasert, Yacoob, Davis McKenna, Gong Wang, Brandstein Basu, Essa, Pentland Harville, Darrell, et al
29 Head pose: mesh model Horprasert, Yacoob, Davis McKenna, Gong Wang, Brandstein Basu, Essa, Pentland Harville, Darrell, et al
30 Head pose: Integration Horprasert, Yacoob, Davis McKenna, Gong Wang, Brandstein Basu, Essa, Pentland Harville, Darrell, et al
31 My approach Integrate changes in pose (after Harville et al) Use mesh model (after Basu et al) Need automatic initialization Head detection, tracking, segmentation Reference orientation Head shape parameters Initialization drives design
32 Head tracking, segmentation Segment by color histogram, grouped motion Match against ellipse model (M. Pilu et al)
33 Mutual gaze as reference point
34 Mutual gaze as reference point
35 Tracking pose changes Choose coordinates to suit tracking 4 of 6 degrees of freedom measurable from monocular image Independent of shape parameters X translation Y translation Translation in depth In-plane rotation
36 Remaining coordinates 2 degrees of freedom remaining Choose as surface coordinate on head Specify where image plane is tangent to head Isolates effect of errors in parameters Tangent region shifts when head rotates in depth
37 Surface coordinates Establish surface coordinate system with mesh
38 Initializing a surface mesh
39 Example
40 Typical results Ground truth due to Sclaroff et al.
41 Merits No need for any manual initialization Capable of running for long periods Tracking accuracy is insensitive to model User independent Real-time
42 Problems Greater accuracy possible with manual initialization Deals poorly with certain classes of head movement (e.g. 360 rotation) Can t initialize without occasional mutual regard
43 Motivation for communication Human-readable actions Reading human actions Conclusions
44 Other protocols
45 Other protocols Protocol for negotiating interpersonal distance Person backs off Person draws closer Too close withdrawal response Comfortable interaction distance Too far calling behavior Beyond sensor range
46 Other protocols Protocol for negotiating interpersonal distance Protocol for controlling the presentation of objects Comfortable interaction speed Too fast, Too close threat response Too fast irritation response
47 Other protocols Protocol for negotiating interpersonal distance Protocol for controlling the presentation of objects Protocol for conversational turn-taking Protocol for introducing vocabulary Protocol for communicating processes Protocols make good modules
48 Cameras Eye, neck, jaw motors QNX Ear, eyebrow, eyelid, lip motors sockets, CORBA Motor ctrl Tracker Attent. system Dist. to target Eye finder Motion filter dual-port RAM Face Control L Percept & Motor Speakers NT speech synthesis affect recognition CORBA Skin filter Recog. pose Color filter Track pose audio speech comms Track head Emotion Drives & Behavior Microphone Linux Speech Linux recognition Speech recognition CORBA
49 Wide Camera Wide Camera 2 Left Foveal Camera Right Foveal Camera Wide Frame Grabber Left Frame Grabber Left Frame Grabber Right Frame Grabber Skin Detector Color Detector Motion Detector Face Detector Distance to target Eye finder Foveal Disparity W W W W Attention Tracked target Salient target Wide Tracker Behaviors Motivations Fixed Action Pattern Affective Postural Shifts w/ gaze comp. Ballistic movement q, dq, d 2 p p q p q, dq, d 2 f f q f Saccade w/ neck VOR comp. q, dq, d 2 s s q s Q, Q, Q Arbitor... Locus of attention Smooth Pursuit & Vergence w/ neck comp. q, dq, d 2 v v q v Disparity Eye-Head-Neck Control Motion Control Daemon Eye-Neck Motors
50 Other protocols What about robot robot protocol? Basically computer computer But physical states may be hard to model Borrow human robot protocol for these
51 Current, future work Protocols for reference Know how to point to an object How to point to an attribute? Or an action? Until a better answer comes along: Communicate task/game that depends on attribute/action Pull out number of classes, positive and negative examples for supervised learning
52 FIN
Yaw Pitch Roll. Head pose estimation without manual initialization. DISTRIBUTION STATEMENT A Approved for Public Release Distribution Unlimited
Head pose estimation without manual initialization Paul Fitzpatrick Al Lab, MIT, Cambridge, USA paulfitz@ai.mit.edu. Abstract. This paper describes a fully automatic system for recovering the rigid components
More informationAbstract. 1 Introduction. 2 Motivation. Information and Communication Engineering October 29th 2010
Information and Communication Engineering October 29th 2010 A Survey on Head Pose Estimation from Low Resolution Image Sato Laboratory M1, 48-106416, Isarun CHAMVEHA Abstract Recognizing the head pose
More informationGaze Control for a Two-Eyed Robot Head
Gaze Control for a Two-Eyed Robot Head Fenglei Du Michael Brady David Murray Robotics Research Group Engineering Science Department Oxford University 19 Parks Road Oxford 0X1 3PJ Abstract As a first step
More informationTracking Head Yaw by Interpolation of Template Responses
Tracking Head Yaw by Interpolation of Template Responses Mario Romero and Aaron Bobick College of Computing Georgia Institute of Technology Atlanta, Georgia 3318 Email: {mromero, afb}@cc.gatech.edu Abstract
More informationReal-time Talking Head Driven by Voice and its Application to Communication and Entertainment
ISCA Archive Real-time Talking Head Driven by Voice and its Application to Communication and Entertainment Shigeo MORISHIMA Seikei University ABSTRACT Recently computer can make cyberspace to walk through
More informationReal Time Face Tracking and Pose Estimation Using an Adaptive Correlation Filter for Human-Robot Interaction
Real Time Face Tracking and Pose Estimation Using an Adaptive Correlation Filter for Human-Robot Interaction Vo Duc My and Andreas Zell Abstract In this paper, we present a real time algorithm for mobile
More informationMulti-Modal Human- Computer Interaction
Multi-Modal Human- Computer Interaction Attila Fazekas University of Debrecen, Hungary Road Map Multi-modal interactions and systems (main categories, examples, benefits) Face detection, facial gestures
More informationGaze interaction (2): models and technologies
Gaze interaction (2): models and technologies Corso di Interazione uomo-macchina II Prof. Giuseppe Boccignone Dipartimento di Scienze dell Informazione Università di Milano boccignone@dsi.unimi.it http://homes.dsi.unimi.it/~boccignone/l
More informationDesign of an anthropomorphic robot head
Design of an anthropomorphic robot head K Undbekken, D Wilson, P Mowforth, S Solbakken and B Eiklid The Turing Institute Glasgow, UK * Abstract This paper describes the design of a robot head featuring
More informationTask analysis based on observing hands and objects by vision
Task analysis based on observing hands and objects by vision Yoshihiro SATO Keni Bernardin Hiroshi KIMURA Katsushi IKEUCHI Univ. of Electro-Communications Univ. of Karlsruhe Univ. of Tokyo Abstract In
More informationEugene Borovikov. Human Head Pose Estimation by Facial Features Location. UMCP Human Head Pose Estimation by Facial Features Location
Human Head Pose Estimation by Facial Features Location Abstract: Eugene Borovikov University of Maryland Institute for Computer Studies, College Park, MD 20742 4/21/1998 We describe a method for estimating
More informationFoveated Vision and Object Recognition on Humanoid Robots
Foveated Vision and Object Recognition on Humanoid Robots Aleš Ude Japan Science and Technology Agency, ICORP Computational Brain Project Jožef Stefan Institute, Dept. of Automatics, Biocybernetics and
More informationNon-Differentiable Image Manifolds
The Multiscale Structure of Non-Differentiable Image Manifolds Michael Wakin Electrical l Engineering i Colorado School of Mines Joint work with Richard Baraniuk, Hyeokho Choi, David Donoho Models for
More informationTracking. Hao Guan( 管皓 ) School of Computer Science Fudan University
Tracking Hao Guan( 管皓 ) School of Computer Science Fudan University 2014-09-29 Multimedia Video Audio Use your eyes Video Tracking Use your ears Audio Tracking Tracking Video Tracking Definition Given
More informationBasic distinctions. Definitions. Epstein (1965) familiar size experiment. Distance, depth, and 3D shape cues. Distance, depth, and 3D shape cues
Distance, depth, and 3D shape cues Pictorial depth cues: familiar size, relative size, brightness, occlusion, shading and shadows, aerial/ atmospheric perspective, linear perspective, height within image,
More informationAn attentional system for a humanoid robot exploiting space variant vision
An attentional system for a humanoid robot exploiting space variant vision Giorgio Metta MIT- AI-Lab, Humanoid Robotics Group 2, Technology Square Cambridge MA - USA Abstract This paper describes an implementation
More informationDetermining pose of a human face from a single monocular image
Determining pose of a human face from a single monocular image Jian-Gang Wang 1, Eric Sung 2, Ronda Venkateswarlu 1 1 Institute for Infocomm Research 21 Heng Mui Keng Terrace, Singapore 119613 2 Nanyang
More informationBinocular cues to depth PSY 310 Greg Francis. Lecture 21. Depth perception
Binocular cues to depth PSY 310 Greg Francis Lecture 21 How to find the hidden word. Depth perception You can see depth in static images with just one eye (monocular) Pictorial cues However, motion and
More informationKeywords Binary Linked Object, Binary silhouette, Fingertip Detection, Hand Gesture Recognition, k-nn algorithm.
Volume 7, Issue 5, May 2017 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Hand Gestures Recognition
More informationA Modular Software Framework for Eye-Hand Coordination in Humanoid Robots
A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots Jurgen Leitner, Simon Harding, Alexander Forster and Peter Corke Presentation: Hana Fusman Introduction/ Overview The goal of their
More information3D Active Appearance Model for Aligning Faces in 2D Images
3D Active Appearance Model for Aligning Faces in 2D Images Chun-Wei Chen and Chieh-Chih Wang Abstract Perceiving human faces is one of the most important functions for human robot interaction. The active
More informationComplex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors
Complex Sensors: Cameras, Visual Sensing The Robotics Primer (Ch. 9) Bring your laptop and robot everyday DO NOT unplug the network cables from the desktop computers or the walls Tuesday s Quiz is on Visual
More informationIBM Research Report. Automatic Neutral Face Detection Using Location and Shape Features
RC 22259 (W0111-073) November 27, 2001 Computer Science IBM Research Report Automatic Neutral Face Detection Using Location and Shape Features Ying-Li Tian, Rudolf M. Bolle IBM Research Division Thomas
More informationProc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp , Kobe, Japan, September 1992
Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp.957-962, Kobe, Japan, September 1992 Tracking a Moving Object by an Active Vision System: PANTHER-VZ Jun Miura, Hideharu Kawarabayashi,
More informationHead Pose Estimation with One Camera, in Uncalibrated Environments
Head Pose Estimation with One Camera, in Uncalibrated Environments Stylianos Asteriadis Image, Video and Multimedia Systems Lab National Technical University of Athens 9, Iroon Polytechniou str, Athens,
More informationAutomatic Enhancement of Correspondence Detection in an Object Tracking System
Automatic Enhancement of Correspondence Detection in an Object Tracking System Denis Schulze 1, Sven Wachsmuth 1 and Katharina J. Rohlfing 2 1- University of Bielefeld - Applied Informatics Universitätsstr.
More informationGesture Recognition Technique:A Review
Gesture Recognition Technique:A Review Nishi Shah 1, Jignesh Patel 2 1 Student, Indus University, Ahmedabad 2 Assistant Professor,Indus University,Ahmadabad Abstract Gesture Recognition means identification
More informationECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination
ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall 2008 October 29, 2008 Notes: Midterm Examination This is a closed book and closed notes examination. Please be precise and to the point.
More informationGesture Recognition: Hand Pose Estimation. Adrian Spurr Ubiquitous Computing Seminar FS
Gesture Recognition: Hand Pose Estimation Adrian Spurr Ubiquitous Computing Seminar FS2014 27.05.2014 1 What is hand pose estimation? Input Computer-usable form 2 Augmented Reality Gaming Robot Control
More informationAutomatic Detecting Neutral Face for Face Authentication and Facial Expression Analysis
From: AAAI Technical Report SS-03-08. Compilation copyright 2003, AAAI (www.aaai.org). All rights reserved. Automatic Detecting Neutral Face for Face Authentication and Facial Expression Analysis Ying-li
More informationMatthias Zobel, Joachim Denzler, Heinrich Niemann Binocular 3-D Object Tracking with Varying Focal Lengths
Matthias Zobel, Joachim Denzler, Heinrich Niemann Binocular 3-D Object Tracking with Varying Focal Lengths appeared in: Proceedings of the IASTED International Conference on Signal Processing, Pattern
More informationUsing temporal seeding to constrain the disparity search range in stereo matching
Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department
More informationImage processing techniques for driver assistance. Razvan Itu June 2014, Technical University Cluj-Napoca
Image processing techniques for driver assistance Razvan Itu June 2014, Technical University Cluj-Napoca Introduction Computer vision & image processing from wiki: any form of signal processing for which
More informationIntuitive Human-Robot Interaction through Active 3D Gaze Tracking
Intuitive Human-Robot Interaction through Active 3D Gaze Tracking Rowel Atienza and Alexander Zelinsky Research School of Information Sciences and Engineering The Australian National University Canberra,
More informationAn Interactive Technique for Robot Control by Using Image Processing Method
An Interactive Technique for Robot Control by Using Image Processing Method Mr. Raskar D. S 1., Prof. Mrs. Belagali P. P 2 1, E&TC Dept. Dr. JJMCOE., Jaysingpur. Maharashtra., India. 2 Associate Prof.
More informationHead Pose Estimation by using Morphological Property of Disparity Map
Head Pose Estimation by using Morphological Property of Disparity Map Sewoong Jun*, Sung-Kee Park* and Moonkey Lee** *Intelligent Robotics Research Center, Korea Institute Science and Technology, Seoul,
More informationThree-Dimensional Hand Pointing Recognition using Two Cameras by Interpolation and Integration of Classification Scores
Three-Dimensional Hand Pointing Recognition using Two Cameras by Interpolation and Integration of Classification Scores Dai Fujita and Takashi Komuro Graduate School of Science and Engineering, Saitama
More information3D Hand Pose Estimation by Finding Appearance-Based Matches in a Large Database of Training Views
3D Hand Pose Estimation by Finding Appearance-Based Matches in a Large Database of Training Views Vassilis Athitsos and Stan Sclaroff Computer Science Department Boston University 111 Cummington Street
More informationAdaptive View-Based Appearance Models
Adaptive View-Based Appearance Models Louis-Philippe Morency Ali Rahimi Trevor Darrell MIT Artificial Intelligence Laboratory Cambridge, MA 0239 Abstract We present a method for online rigid object tracking
More informationComparative Study of Coarse Head Pose Estimation
Comparative Study of Coarse Head Pose Estimation Lisa M. Brown and Ying-Li Tian IBM T.J. Watson Research Center Hawthorne, NY 10532 {lisabr,yltian}@us.ibm.com Abstract For many practical applications,
More information3D FIXATIONS IN REAL AND VIRTUAL SCENARIOS. Thies Pfeiffer, A.I. & V.R. Group, Bielefeld University
3D FIXATIONS IN REAL AND VIRTUAL SCENARIOS Thies Pfeiffer, A.I. & V.R. Group, Bielefeld University Background multimodal human computer interaction situated natural communication (gaze, gesture, speech)
More informationAnalyzing the Relationship Between Head Pose and Gaze to Model Driver Visual Attention
Analyzing the Relationship Between Head Pose and Gaze to Model Driver Visual Attention Sumit Jha and Carlos Busso Multimodal Signal Processing (MSP) Laboratory Department of Electrical Engineering, The
More informationA REAL-TIME FACIAL FEATURE BASED HEAD TRACKER
A REAL-TIME FACIAL FEATURE BASED HEAD TRACKER Jari Hannuksela, Janne Heikkilä and Matti Pietikäinen {jari.hannuksela, jth, mkp}@ee.oulu.fi Machine Vision Group, Infotech Oulu P.O. Box 45, FIN-914 University
More informationRobust Tracking of People by a Mobile Robotic Agent
Robust Tracking of People by a Mobile Robotic Agent Rawesak Tanawongsuwan, Alexander Stoytchev, Irfan Essa College of Computing, GVU Center, Georgia Institute of Technology Atlanta, GA 30332-0280 USA ftee
More informationCamera Actions. Shots, Angles & Movements
Camera Actions Shots, Angles & Movements Camera Shots (Shot Sizes) Camera Shots There are several different Camera Shots used in TV and Film Production. Each Camera Shot is given a name and description
More informationA Bio-Inspired Sensory-Motor Neural Model for a Neuro-Robotic Manipulation Platform
NEUROBOTICS Meeting Genova, September 22, 2005 A Bio-Inspired Sensory-Motor Neural Model for a Neuro-Robotic Manipulation Platform Gioel Asuni, Giancarlo Teti, Cecilia Laschi, Eugenio Guglielmelli and
More informationInformation and Communication Engineering July 9th Survey of Techniques for Head Pose Estimation from Low Resolution Images
Information and Communication Engineering July 9th 2010 Survey of Techniques for Head Pose Estimation from Low Resolution Images Sato Laboratory D1, 48-107409, Teera SIRITEERAKUL Abstract Head pose estimation
More informationMonocular Head Pose Estimation using Generalized Adaptive View-based Appearance Model
Monocular Head Pose Estimation using Generalized Adaptive View-based Appearance Model Louis-Philippe Morency a, Jacob Whitehill b, Javier Movellan b a USC Institute for Creative Technologies Marina del
More informationA System for Tracking and Recognizing Multiple People with Multiple Cameras
Georgia Institute of Technology, Graphics, Visualization and Usability Center, Technical Report # GIT-GVU-98-25, August 1998. A System for Tracking and Recognizing Multiple People with Multiple Cameras
More informationReal-time Driver Affect Analysis and Tele-viewing System i
Appeared in Intelligent Vehicles Symposium, Proceedings. IEEE, June 9-11, 2003, 372-377 Real-time Driver Affect Analysis and Tele-viewing System i Joel C. McCall, Satya P. Mallick, and Mohan M. Trivedi
More informationOn the Design and Evaluation of Robust Head Pose for Visual User Interfaces: Algorithms, Databases, and Comparisons
On the Design and Evaluation of Robust Head Pose for Visual User Interfaces: Algorithms, Databases, and Comparisons Sujitha Martin scmartin@ucsd.edu Shinko Y. Cheng sycheng@hrl.com Ashish Tawari atawari@ucsd.edu
More informationHead Pose Estimation Using Stereo Vision For Human-Robot Interaction
Head Pose Estimation Using Stereo Vision For Human-Robot Interaction Edgar Seemann Kai Nickel Rainer Stiefelhagen Interactive Systems Labs Universität Karlsruhe (TH) Germany Abstract In this paper we present
More informationThe Isometric Self-Organizing Map for 3D Hand Pose Estimation
The Isometric Self-Organizing Map for 3D Hand Pose Estimation Haiying Guan, Rogerio S. Feris, and Matthew Turk Computer Science Department University of California, Santa Barbara {haiying,rferis,mturk}@cs.ucsb.edu
More informationHead Tracking via Robust Registration in Texture Map Images
Boston University OpenBU Computer Science http://open.bu.edu CAS: Computer Science: Technical Reports 1997-11-24 Head Tracking via Robust Registration in Texture Map Images La Cascia, Marco Boston University
More information3D-Tracking of Head and Hands for Pointing Gesture Recognition in a Human-Robot Interaction Scenario
3D-Tracking of Head and Hands for Pointing Gesture Recognition in a Human-Robot Interaction Scenario Kai Nickel Edgar Seemann Rainer Stiefelhagen Interactive Systems Labs Universitt Karlsruhe (TH) Germany
More informationFingertips Tracking based on Gradient Vector
Int. J. Advance Soft Compu. Appl, Vol. 7, No. 3, November 2015 ISSN 2074-8523 Fingertips Tracking based on Gradient Vector Ahmad Yahya Dawod 1, Md Jan Nordin 1, and Junaidi Abdullah 2 1 Pattern Recognition
More informationDepth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth
Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze
More informationPose Estimation using 3D View-Based Eigenspaces
Pose Estimation using 3D View-Based Eigenspaces Louis-Philippe Morency Patrik Sundberg Trevor Darrell MIT Artificial Intelligence Laboratory Cambridge, MA 02139 Abstract In this paper we present a method
More informationGaze Tracking. Introduction :
Introduction : Gaze Tracking In 1879 in Paris, Louis Émile Javal observed that reading does not involve a smooth sweeping of the eyes along the text, as previously assumed, but a series of short stops
More informationHuman Upper Body Pose Estimation in Static Images
1. Research Team Human Upper Body Pose Estimation in Static Images Project Leader: Graduate Students: Prof. Isaac Cohen, Computer Science Mun Wai Lee 2. Statement of Project Goals This goal of this project
More informationModel Based Techniques for Tracking Human Head for Human Robot Interaction
Model Based Techniques for Tracking Human Head for Human Robot Interaction Faisal Khan University of Wisconsin, Madison faisal@cs.wisc.edu Varghese Mathew University of Wisconsin, Madison vmathew@cs.wisc.edu
More informationStereoscopic Systems Part 1
Stereoscopic Systems Part 1 Terminology: Stereoscopic vs. 3D 3D Animation refers to computer animation created with programs (like Maya) that manipulate objects in a 3D space, though the rendered image
More informationThink-Pair-Share. What visual or physiological cues help us to perceive 3D shape and depth?
Think-Pair-Share What visual or physiological cues help us to perceive 3D shape and depth? [Figure from Prados & Faugeras 2006] Shading Focus/defocus Images from same point of view, different camera parameters
More informationLast update: May 4, Vision. CMSC 421: Chapter 24. CMSC 421: Chapter 24 1
Last update: May 4, 200 Vision CMSC 42: Chapter 24 CMSC 42: Chapter 24 Outline Perception generally Image formation Early vision 2D D Object recognition CMSC 42: Chapter 24 2 Perception generally Stimulus
More informationSynthesizing Realistic Facial Expressions from Photographs
Synthesizing Realistic Facial Expressions from Photographs 1998 F. Pighin, J Hecker, D. Lischinskiy, R. Szeliskiz and D. H. Salesin University of Washington, The Hebrew University Microsoft Research 1
More informationVS 117 LABORATORY III: FIXATION DISPARITY INTRODUCTION
VS 117 LABORATORY III: FIXATION DISPARITY INTRODUCTION Under binocular viewing, subjects generally do not gaze directly at the visual target. Fixation disparity is defined as the difference between the
More informationLUMS Mine Detector Project
LUMS Mine Detector Project Using visual information to control a robot (Hutchinson et al. 1996). Vision may or may not be used in the feedback loop. Visual (image based) features such as points, lines
More informationEvaluation of Expression Recognition Techniques
Evaluation of Expression Recognition Techniques Ira Cohen 1, Nicu Sebe 2,3, Yafei Sun 3, Michael S. Lew 3, Thomas S. Huang 1 1 Beckman Institute, University of Illinois at Urbana-Champaign, USA 2 Faculty
More informationSteering without Representation using Active Fixation
Steering without Representation using Active Fixation D W Murray, I D Reid and A J Davison Department of Engineering Science, University of Oxford, Parks Road, Oxford OX1 3PJ, UK Abstract. This paper demonstrates
More informationHuman Face Classification using Genetic Algorithm
Human Face Classification using Genetic Algorithm Tania Akter Setu Dept. of Computer Science and Engineering Jatiya Kabi Kazi Nazrul Islam University Trishal, Mymenshing, Bangladesh Dr. Md. Mijanur Rahman
More informationCrazyTalk 5 Basic Face Fitting and Techniques
Category CrazyTalk 5 Basic Face Fitting and Techniques Date 2008.01.10 CrazyTalk_Tutorial_Model Version CrazyTalk 5.0 Author Abstract James C. Martin Fitting an image for use in CrazyTalk is quick and
More informationNew approaches to pattern recognition and automated learning
Z Y X New approaches to pattern recognition and automated learning Technology Forum 2015 Johannes Zuegner STEMMER IMAGING GmbH, Puchheim, Germany OUTLINE Introduction Description of the task What does
More informationAnimation of 3D surfaces.
Animation of 3D surfaces Motivations When character animation is controlled by skeleton set of hierarchical joints joints oriented by rotations the character shape still needs to be visible: visible =
More informationOcclusion Robust Multi-Camera Face Tracking
Occlusion Robust Multi-Camera Face Tracking Josh Harguess, Changbo Hu, J. K. Aggarwal Computer & Vision Research Center / Department of ECE The University of Texas at Austin harguess@utexas.edu, changbo.hu@gmail.com,
More informationVisual Perception for Robots
Visual Perception for Robots Sven Behnke Computer Science Institute VI Autonomous Intelligent Systems Our Cognitive Robots Complete systems for example scenarios Equipped with rich sensors Flying robot
More informationFrom Gaze to Focus of Attention
From Gaze to Focus of Attention Rainer Stiefelhagen, Michael Finke, Jie Yang, Alex Waibel stiefel@ira.uka.de, finkem@cs.cmu.edu, yang+@cs.cmu.edu, ahw@cs.cmu.edu Interactive Systems Laboratories University
More informationRobust Full-Motion Recovery of Head. by Dynamic Templates and Re-registration Techniques
Robust Full-Motion Recovery of Head by Dynamic Templates and Re-registration Techniques Jing Xiao, Tsuyoshi Moriyama, Takeo Kanade Robotics Institute, Carnegie Mellon University {jxiao, tmoriyam, tk}@cs.cmu.edu
More information3D Transformations. CS 4620 Lecture 10. Cornell CS4620 Fall 2014 Lecture Steve Marschner (with previous instructors James/Bala)
3D Transformations CS 4620 Lecture 10 1 Translation 2 Scaling 3 Rotation about z axis 4 Rotation about x axis 5 Rotation about y axis 6 Properties of Matrices Translations: linear part is the identity
More informationAnimation of 3D surfaces
Animation of 3D surfaces 2013-14 Motivations When character animation is controlled by skeleton set of hierarchical joints joints oriented by rotations the character shape still needs to be visible: visible
More informationReliable Face Recognition for Intelligent CCTV
Reliable Face Recognition for Intelligent CCTV Ting Shan, Brian C. Lovell, Shaokang Chen and Abbas Bigdeli National ICT Australia (NICTA) and Intelligent Real-Time Imaging and Sensing Group EMI, School
More information3D TRACKING AND DYNAMIC ANALYSIS OF HUMAN HEAD MOVEMENTS AND ATTENTIONAL TARGETS. Erik Murphy-Chutorian and Mohan M. Trivedi
3D TRACKING AND DYNAMIC ANALYSIS OF HUMAN HEAD MOVEMENTS AND ATTENTIONAL TARGETS Erik Murphy-Chutorian and Mohan M. Trivedi Systems for Human Interactivity, Visualization and Analysis Computer Vision and
More informationFace Alignment Across Large Poses: A 3D Solution
Face Alignment Across Large Poses: A 3D Solution Outline Face Alignment Related Works 3D Morphable Model Projected Normalized Coordinate Code Network Structure 3D Image Rotation Performance on Datasets
More informationComputer and Machine Vision
Computer and Machine Vision Lecture Week 4 Part-2 February 5, 2014 Sam Siewert Outline of Week 4 Practical Methods for Dealing with Camera Streams, Frame by Frame and De-coding/Re-encoding for Analysis
More informationMachine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices
Technical University of Cluj-Napoca Image Processing and Pattern Recognition Research Center www.cv.utcluj.ro Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving
More informationPractice Exam Sample Solutions
CS 675 Computer Vision Instructor: Marc Pomplun Practice Exam Sample Solutions Note that in the actual exam, no calculators, no books, and no notes allowed. Question 1: out of points Question 2: out of
More informationFACE ANALYSIS AND SYNTHESIS FOR INTERACTIVE ENTERTAINMENT
FACE ANALYSIS AND SYNTHESIS FOR INTERACTIVE ENTERTAINMENT Shoichiro IWASAWA*I, Tatsuo YOTSUKURA*2, Shigeo MORISHIMA*2 */ Telecommunication Advancement Organization *2Facu!ty of Engineering, Seikei University
More information3D Transformations. CS 4620 Lecture Kavita Bala w/ prior instructor Steve Marschner. Cornell CS4620 Fall 2015 Lecture 11
3D Transformations CS 4620 Lecture 11 1 Announcements A2 due tomorrow Demos on Monday Please sign up for a slot Post on piazza 2 Translation 3 Scaling 4 Rotation about z axis 5 Rotation about x axis 6
More information3D Face and Hand Tracking for American Sign Language Recognition
3D Face and Hand Tracking for American Sign Language Recognition NSF-ITR (2004-2008) D. Metaxas, A. Elgammal, V. Pavlovic (Rutgers Univ.) C. Neidle (Boston Univ.) C. Vogler (Gallaudet) The need for automated
More informationExperimental Humanities II. Eye-Tracking Methodology
Experimental Humanities II Eye-Tracking Methodology Course outline 22.3. Introduction to Eye-Tracking + Lab 1 (DEMO & stimuli creation) 29.3. Setup and Calibration for Quality Data Collection + Lab 2 (Calibration
More informationWaleed Pervaiz CSE 352
Waleed Pervaiz CSE 352 Computer Vision is the technology that enables machines to see and obtain information from digital images. It is seen as an integral part of AI in fields such as pattern recognition
More informationFacial Action Detection from Dual-View Static Face Images
Facial Action Detection from Dual-View Static Face Images Maja Pantic and Leon Rothkrantz Delft University of Technology Electrical Engineering, Mathematics and Computer Science Mekelweg 4, 2628 CD Delft,
More informationADVANCES in NATURAL and APPLIED SCIENCES
ADVANCES in NATURAL and APPLIED SCIENCES ISSN: 1995-0772 Published BY AENSI Publication EISSN: 1998-1090 http://www.aensiweb.com/anas 2016 Special 10(6): pages Open Access Journal Estimation of Various
More informationZürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary
Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza ETH Master Course: 151-0854-00L Autonomous Mobile Robots Summary 2 Lecture Overview Mobile Robot Control Scheme knowledge, data base mission
More informationPERSON FOLLOWING AND MOBILE ROBOT VIA ARM-POSTURE DRIVING USING COLOR AND STEREOVISION. Bogdan Kwolek
PERSON FOLLOWING AND MOBILE ROBOT VIA ARM-POSTURE DRIVING USING COLOR AND STEREOVISION Bogdan Kwolek Rzeszów University of Technology, W. Pola 2, 35-959 Rzeszów, Poland Abstract: This paper describes an
More informationComputer Animation Visualization. Lecture 5. Facial animation
Computer Animation Visualization Lecture 5 Facial animation Taku Komura Facial Animation The face is deformable Need to decide how all the vertices on the surface shall move Manually create them Muscle-based
More informationFacial Expression Recognition using a Dynamic Model and Motion Energy
M.I.T Media Laboratory Perceptual Computing Section Technical Report No. Appears: International Conference on Computer Vision 9, Cambridge, MA, June -, 99 Facial Expression Recognition using a Dynamic
More informationActions and Graphs in Blender - Week 8
Actions and Graphs in Blender - Week 8 Sculpt Tool Sculpting tools in Blender are very easy to use and they will help you create interesting effects and model characters when working with animation and
More informationA Real-Time Hand Gesture Recognition for Dynamic Applications
e-issn 2455 1392 Volume 2 Issue 2, February 2016 pp. 41-45 Scientific Journal Impact Factor : 3.468 http://www.ijcter.com A Real-Time Hand Gesture Recognition for Dynamic Applications Aishwarya Mandlik
More informationMonocular Visual Odometry
Elective in Robotics coordinator: Prof. Giuseppe Oriolo Monocular Visual Odometry (slides prepared by Luca Ricci) Monocular vs. Stereo: eamples from Nature Predator Predators eyes face forward. The field
More informationColorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.
Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Stereo Vision 2 Inferring 3D from 2D Model based pose estimation single (calibrated) camera > Can
More information