Multi-Modal Human- Computer Interaction

Similar documents
Disguised Face Identification Based Gabor Feature and SVM Classifier

Gender Classification Technique Based on Facial Features using Neural Network

Mouse Pointer Tracking with Eyes

Automatic Detecting Neutral Face for Face Authentication and Facial Expression Analysis

A Real Time Facial Expression Classification System Using Local Binary Patterns

Eye Detection by Haar wavelets and cascaded Support Vector Machine

Exploring Bag of Words Architectures in the Facial Expression Domain

Skin and Face Detection

Face Alignment Under Various Poses and Expressions

Head Pose Estimation by using Morphological Property of Disparity Map

ADVANCES in NATURAL and APPLIED SCIENCES

Human Face Classification using Genetic Algorithm

Face detection and recognition. Many slides adapted from K. Grauman and D. Lowe

Fully Automatic Facial Action Recognition in Spontaneous Behavior

Emotional States Control for On-line Game Avatars

The Analysis of Faces in Brains and Machines

HUMAN S FACIAL PARTS EXTRACTION TO RECOGNIZE FACIAL EXPRESSION

IBM Research Report. Automatic Neutral Face Detection Using Location and Shape Features

Learning the Deep Features for Eye Detection in Uncontrolled Conditions

A Hybrid Face Detection System using combination of Appearance-based and Feature-based methods

FACIAL EXPRESSION USING 3D ANIMATION

Short Paper Boosting Sex Identification Performance

Detection of a Single Hand Shape in the Foreground of Still Images

Texture Features in Facial Image Analysis

Angle Based Facial Expression Recognition

Multiple Kernel Learning for Emotion Recognition in the Wild

Face detection, validation and tracking. Océane Esposito, Grazina Laurinaviciute, Alexandre Majetniak

Evaluation of Gabor-Wavelet-Based Facial Action Unit Recognition in Image Sequences of Increasing Complexity

Gender Classification

FACE DETECTION BY HAAR CASCADE CLASSIFIER WITH SIMPLE AND COMPLEX BACKGROUNDS IMAGES USING OPENCV IMPLEMENTATION

Classifier Case Study: Viola-Jones Face Detector

LUMS Mine Detector Project

3D Facial Action Units Recognition for Emotional Expression

FACIAL EXPRESSION USING 3D ANIMATION TECHNIQUE

Lecture 14: Computer Vision

RGBD Face Detection with Kinect Sensor. ZhongJie Bi

Real-time facial feature point extraction

DA Progress report 2 Multi-view facial expression. classification Nikolas Hesse

Topics for thesis. Automatic Speech-based Emotion Recognition

Criminal Identification System Using Face Detection and Recognition

Face detection and recognition. Detection Recognition Sally

Generic Face Alignment Using an Improved Active Shape Model

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN

An Interactive Technique for Robot Control by Using Image Processing Method

A HYBRID APPROACH BASED ON PCA AND LBP FOR FACIAL EXPRESSION ANALYSIS

Facial Feature Extraction Based On FPD and GLCM Algorithms

Facial Expression Recognition Using Non-negative Matrix Factorization

Facial expression recognition is a key element in human communication.

A Simple Approach to Facial Expression Recognition

Last week. Multi-Frame Structure from Motion: Multi-View Stereo. Unknown camera viewpoints

Image Processing Pipeline for Facial Expression Recognition under Variable Lighting

Modelling Human Perception of Facial Expressions by Discrete Choice Models

Real Time Face Tracking and Pose Estimation Using an Adaptive Correlation Filter for Human-Robot Interaction

The CSU Face Identification Evaluation System: Its Purpose, Features and Structure

Face/Flesh Detection and Face Recognition

An Integration of Face detection and Tracking for Video As Well As Images

Facial expression recognition based on two-step feature histogram optimization Ling Gana, Sisi Sib

Discriminative classifiers for image recognition

Face Detection and Alignment. Prof. Xin Yang HUST

Machine Learning for Signal Processing Detecting faces (& other objects) in images

An Image Region Selection with Local Binary Pattern based for Face Recognition

Hybrid Approach of Facial Expression Recognition

Computer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki

Pattern Recognition Letters

Human Face Detection and Tracking using Skin Color Combined with Optical Flow Based Segmentation

Detection of Mouth Movements and Its Applications to Cross-Modal Analysis of Planning Meetings

Robust Facial Expression Classification Using Shape and Appearance Features

Describable Visual Attributes for Face Verification and Image Search

Classification of Face Images for Gender, Age, Facial Expression, and Identity 1

LOCAL FEATURE EXTRACTION METHODS FOR FACIAL EXPRESSION RECOGNITION

Learning to Recognize Faces in Realistic Conditions

Detecting Faces in Images. Detecting Faces in Images. Finding faces in an image. Finding faces in an image. Finding faces in an image

On Modeling Variations for Face Authentication

Waleed Pervaiz CSE 352

Visual Perception for Robots

Introduction to Computer Vision

Machine Learning Approach for Smile Detection in Real Time Images

Object Detection System

Recognition problems. Face Recognition and Detection. Readings. What is recognition?

Facial Expression Analysis

Eye Detection Based-on Color and Shape Features

Face Detection CUDA Accelerating

Facial Emotion Recognition using Eye

Operating an Application Using Hand Gesture Recognition System

Classification of Upper and Lower Face Action Units and Facial Expressions using Hybrid Tracking System and Probabilistic Neural Networks

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

The BJUT-3D Large-Scale Chinese Face Database

A Survey of Various Face Detection Methods

Automatic Facial Expression Recognition based on the Salient Facial Patches

Global fitting of a facial model to facial features for model based video coding

Human Robot Communication. Paul Fitzpatrick

Face Cyclographs for Recognition

Boosting Sex Identification Performance

MORPH-II: Feature Vector Documentation

Color-based Face Detection using Combination of Modified Local Binary Patterns and embedded Hidden Markov Models

Discovery Engineering

Facial Expression Recognition Using Expression- Specific Local Binary Patterns and Layer Denoising Mechanism

Facial Expression Recognition Using Improved Artificial Bee Colony Algorithm

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Viola Jones Face Detection. Shahid Nabi Hiader Raiz Muhammad Murtaz

Transcription:

Multi-Modal Human- Computer Interaction Attila Fazekas University of Debrecen, Hungary

Road Map Multi-modal interactions and systems (main categories, examples, benefits) Face detection, facial gestures recognition, etc. Body gesture detection Examples 2

Human-Centered View There are two views on multi-modal interaction: The first focuses on the human side: perception and control. There the word modality refers to human input and output channels. The second view focuses on synergistic using two or more computer input or output modalities. 3

The Modalities We can divide the modalities in seven groups Internal chemical (blood oxygen, etc.) External chemical (taste, etc.) Somatic senses (touch, etc.) Muscle sense (stretch, etc.) Sense of balance Hearing Vision 4

Definition of the Multimodality "Multi-modality is the capacity of the system to communicate with a user along different types of communication channels." Both multimedia and multi-modal systems use multiple communication channels. But a multi-modal system strives for meaning. 5

Two Types of Multi-modal Systems The goal is to use the computer as a tool. The computer as a dialogue partner. 6

History Bolt's Put-That-There system. In this system the user could move objects on screen by pointing and speaking. CUBRICON is a system that uses mouse pointing and speech. Oviatt presented a multi-modal system for dynamic interactive maps. 7

Benefits Efficiency follows from using each modality for the task that it is best suited for. Redundancy increases the likelihood that communication proceeds smoothly because there are many simultaneous references to the same issue. Perceptability increas when the tasks are facilitated in spatial context. 8

Benefits Naturalness follows from the free choice of modalities and may result in a human-computer communication that is close to human-human communication. Accuracy increases when another modality can indicate an object more accurately than the main modality. 9

Applications Mobile telecommunication Hands-free devices to computers Using in a car Interactive information panel 10

Face Analysis 11 Face carries a lot of important information in communication Monitoring the face is fundamental in HCI First step: face detection (localization) Using the localized face can be performed: Tracking face and facial features 2D face tracking, gaze estimation, head-shake detection Face classification gender, age, facial expressions, race Feature extraction skin/eye/hair color mustache/beard detection

Face Detection/Tracking Viola and Jones detector is used Face detection is reduced to image classification problem Given a set of feature types: 12 Training: positive (faces) and negative (random images) examples those features are selected which fits on the positive set (finding position and their extent) the selected features are collected into a cascade file Face detection the different scale of the input image is scanned through fitting the set of features: if all are fitting => there is a face

Face Classification Can be used for: gender, age, facial expression, race detection Preprocessing for feature extraction Cutting the face LBP transform Gabor transform Classification (using the model file) SVM AdaBoost Databases: Cohn-Kanade, FERET Currently we have the following classifiers trained (LBP+SVM) Gender (male vs. female) Facial expressions (happy, sad, surprised, angry) Age estimation (10-29, 30-49, 50-) Race (asian, hispanic, black, white) 13

Facial Feature Color Extraction Determining the color of various facial features: skin, hair and eyes. The full color range of the segmented face image will be reduced to color categories based on human cognition principles The segmentation steps of the HI plane in case of hair colors: (a) color marker points, (b) convex hull, (c) distance transform and (d) the segmented plane: 14

Facial Feature Color Extraction The colors of the facial features are determined in two steps: First, the skin, eyes and hair are segmented in the image using only structural information Then, within the segmented regions the huge number of colors in real color images is substituted by a smaller color set, which is used to determine the color of a given feature. 15

Head-pose Tracking, Gaze Estimation Head pose estimation Detecting the location of the facial features using individual feature detectors (Viola and Jones detectors): eyes, mouth, nose tip Based on the position of the facial features, the POSIT algorithm is used to estimate the three rotation and the translation vector of the head pose Further work: Pupil tracking Combining head pose with the position of the pupil for gaze estimation. 16

Body Gestures on Stereo Image Goal: estimate depth based on an image pair Parallel stereo configuration C l, C r - cams f focal length b baseline p l (x l, y l ) p r (x r, y r ) - ratio:pixels/metre Disparity: d x l x r Depth estimation: D bf d 17

Computing Disparity Feature based approach detect feature points (corners) on the both images compute correlation between feature points on the different image define stereo pairs extend values (region growing, interpolation) 18

Computing Disparity Problems: Too few matching pairs are there on the torso. Wrong matching on the similar areas. 19

Computing Disparity Decrease compute time: Compute disparity on the inportant part (ellipse shape) Computate disparity in the each intersection of 4. row and 4. column within ellipse, and extend disparity. time: ~35ms 20

Computing Disparity 21 Window based approach Compute correlation between all the points pair of images at the same line. Common correlation measure is Sum of Squared Difference ), (, 2 ), ( ), ( ), ( v u W y x y d x R y x L v SSD u ), (, )), ( ), ( ( )), ( ), ( ( ), ( v u W y x v d u R y d x R v u L y x L v u S - SSD is sensitiv to lighting condition - Our measure:

Computing Disparity Quality is depends on the size of window. Test: Image size: 320x240 disparity range: 15-45 Window size: 11x11 time: ~900ms 22

Chess Player Turk-2 23

Thank you for your attention! 24