Task analysis based on observing hands and objects by vision

Similar documents
Grasp Recognition using a 3D Articulated Model and Infrared Images

3D Digitization of a Hand-held Object with a Wearable Vision Sensor

Recognizing Buildings in Urban Scene of Distant View ABSTRACT

AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Symbolic Representation of Trajectories for Skill Generation

A High Speed Face Measurement System

Gesture Recognition using Temporal Templates with disparity information

Short Survey on Static Hand Gesture Recognition

An Interactive Technique for Robot Control by Using Image Processing Method

Photometric Stereo with Auto-Radiometric Calibration

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

Shape Modeling of A String And Recognition Using Distance Sensor

Artificial Finger Skin having Ridges and Distributed Tactile Sensors used for Grasp Force Control

Mapping textures on 3D geometric model using reflectance image

Simple and Robust Tracking of Hands and Objects for Video-based Multimedia Production

Structured light 3D reconstruction

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information

Recognition and Performance of Human Task with 9-eye Stereo Vision and Data Gloves

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

An Adaptive Eigenshape Model

Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System

Incremental Observable-Area Modeling for Cooperative Tracking

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

Intelligent Robots for Handling of Flexible Objects. IRFO Vision System

Robotized Assembly of a Wire Harness in Car Production Line

Computer Science Faculty, Bandar Lampung University, Bandar Lampung, Indonesia

Introducing Robotics Vision System to a Manufacturing Robotics Course

Assist System for Carrying a Long Object with a Human - Analysis of a Human Cooperative Behavior in the Vertical Direction -

Gesture Recognition Technique:A Review

Similarity Image Retrieval System Using Hierarchical Classification

Dynamic Time Warping for Binocular Hand Tracking and Reconstruction

CLASSIFICATION FOR ROADSIDE OBJECTS BASED ON SIMULATED LASER SCANNING

Detecting Printed and Handwritten Partial Copies of Line Drawings Embedded in Complex Backgrounds

Measurement of Pedestrian Groups Using Subtraction Stereo

Conspicuous Character Patterns

Segmentation and Tracking of Partial Planar Templates

A Novel Criterion Function in Feature Evaluation. Application to the Classification of Corks.

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera

Motion Planning for Dynamic Knotting of a Flexible Rope with a High-speed Robot Arm

Player Viewpoint Video Synthesis Using Multiple Cameras

Stereo and Epipolar geometry

The Novel Approach for 3D Face Recognition Using Simple Preprocessing Method

Real-time Image-based Reconstruction of Pipes Using Omnidirectional Cameras

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

6-dof Eye-vergence visual servoing by 1-step GA pose tracking

Determination of 3-D Image Viewpoint Using Modified Nearest Feature Line Method in Its Eigenspace Domain

A deformable model driven method for handling clothes

Hand Gesture Recognition using Depth Data

Machine Vision and Gesture Recognition Machine Vision Term Paper

HAND-GESTURE BASED FILM RESTORATION

Vehicle Ego-localization by Matching In-vehicle Camera Images to an Aerial Image

ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination

DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER

Stable Grasp and Manipulation in 3D Space with 2-Soft-Fingered Robot Hand

Fingertips Tracking based on Gradient Vector

CIRCULAR MOIRÉ PATTERNS IN 3D COMPUTER VISION APPLICATIONS

Tracking of Human Body using Multiple Predictors

Tracking system. Danica Kragic. Object Recognition & Model Based Tracking

Development of Vision System on Humanoid Robot HRP-2

DETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS

CS4670: Computer Vision

AAM Based Facial Feature Tracking with Kinect

Robotics Programming Laboratory

A Two-stage Scheme for Dynamic Hand Gesture Recognition

A Robust Two Feature Points Based Depth Estimation Method 1)

Multi-view Surface Inspection Using a Rotating Table

Toward Part-based Document Image Decoding

Visual Attention Control by Sensor Space Segmentation for a Small Quadruped Robot based on Information Criterion

Robust Lip Contour Extraction using Separability of Multi-Dimensional Distributions

Face Cyclographs for Recognition

3D object recognition used by team robotto

Visual Hulls from Single Uncalibrated Snapshots Using Two Planar Mirrors

Multi-view reconstruction for projector camera systems based on bundle adjustment

BIN PICKING APPLICATIONS AND TECHNOLOGIES

Ball tracking with velocity based on Monte-Carlo localization

High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond

HOG-Based Person Following and Autonomous Returning Using Generated Map by Mobile Robot Equipped with Camera and Laser Range Finder

Face Detection Using Color Based Segmentation and Morphological Processing A Case Study

Miniature faking. In close-up photo, the depth of field is limited.

Detection of Small-Waving Hand by Distributed Camera System

AN EFFICIENT BINARY CORNER DETECTOR. P. Saeedi, P. Lawrence and D. Lowe

Person identification from spatio-temporal 3D gait

Synchronized Ego-Motion Recovery of Two Face-to-Face Cameras

Algorithm research of 3D point cloud registration based on iterative closest point 1

A Robust Gesture Recognition Using Depth Data

Mouse Simulation Using Two Coloured Tapes

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

Estimation of common groundplane based on co-motion statistics

Expanding gait identification methods from straight to curved trajectories

Visuelle Perzeption für Mensch- Maschine Schnittstellen

Constraint-Based Task Programming with CAD Semantics: From Intuitive Specification to Real-Time Control

Object Detection in a Fixed Frame

3D model-based human face modeling

2. TARGET PHOTOS FOR ANALYSIS

Augmenting Reality, Naturally:

Robot localization method based on visual features and their geometric relationship

A Study of Medical Image Analysis System

Dynamic Human Shape Description and Characterization

L1 - Introduction. Contents. Introduction of CAD/CAM system Components of CAD/CAM systems Basic concepts of graphics programming

Transcription:

Task analysis based on observing hands and objects by vision Yoshihiro SATO Keni Bernardin Hiroshi KIMURA Katsushi IKEUCHI Univ. of Electro-Communications Univ. of Karlsruhe Univ. of Tokyo Abstract In order to transmit, share and store human knowledge, it is important for a robot to be able to acquire task models and skills by observing human actions. Since vision plays an important role for observation, we propose a technique for measuring the position and posture of objects and hands in 3-dimensional space at high speed and with high precision by vision. Next, we show a framework for analysis and description of a task by using an object s functions as elements. 1 Introduction It is important for a robot to be able to acquire task models and skills by observing human actions in order to transmit, share and store human knowledge. The strategy for understanding a human s action is the subject which should be tackled in order to solve processes such as studying and gaining experience from a human. Although research on teaching from observation is advanced on various points[1][2][3], the technique is still far from being applicable to real world problems, because of insufficient capabilities in observing and describing the real world. Since the aim of this research is the acquisition of human handworking skills, we tackle the problem of measuring the shape of a hand or an object s posture in 3-dimensional space with sufficient accuracy to do task analysis by vision. Furthermore, we tackle the problem of automatically generating a task description from the acquired vision information as an operation sequence on an abstraction level which can be understood by a human. A human determines the optimal shape of his hand based on the task[4]. The used grasp form is useful for estimating the purpose of the task or of a specific state of the task. Moreover, the kinds of tasks a robot can execute can be increased by generating the robot s grasp operation automatically from observation of the human s grasp operation[5]. Many techniques have been proposed for recognizing the form of the hand, such as classification by silhouette, using devices such as sensor gloves, the eigen-space method by using depth images, etc. However, it is not desirable to use equipment which has physical restrictions like a sensor glove in a scene where a human and a robot do cooperative work. Recognition from arbitrary 3-dimensional positions is required, without being limited to specific viewpoints. Moreover, in conventional research, only the hand region was observed using a low resolution range sensor, and it was not intended to make a wide range observation of the task[6]. In this paper, in order to solve these problems, we perform hand shape recognition with the eigen space method by using a depth image from a high resolution range sensor, in which the hand region was clipped and normalized. The principal subject of task analysis is to observe the state of objects, such as a tool, parts, etc. which are operated by a human. In order to analyze various operations, it is necessary to recognize objects with free curved surfaces, and to acquire their position and posture in 3-dimensional space continuously. The active search method, based on color histograms, is one of the techniques which look for a specific object in an image[7]. Using color histogram search makes the technique robust to little changes in form, and allows to find objects at high speed. However, it is difficult to look for a 3-dimensional object whose silhouette changes with its posture, and it is necessary to use a previously created image database of all silhouettes to obtain a solution. In this paper, we propose a technique for high-speed search by generating the color histogram of arbitrary postures using a color model and a geometric model of the object. Furthermore, the object position and its posture are estimated by the ICP method using a depth image[9]. Task analysis expresses the state of the environment such as a human s action and tools from observation of the human s demonstration. The problem is how to find a good modelling framework, that is, how to design a description language for a recognition strategy. Since handwork is a task that relates to the mutual interaction of humans and objects, it is difficult to generally design the modelling framework. At this stage, various trials are still being performed. In this

Base line 125mm Angle of convergence 7degrees Vertical angle of view 53.4degrees Input Images Neck 43mm distance color Robot body 51mm Observation area Desk error SSAD1 SSAD2 (a) t (b) Figure 1: Observation environment Figure 2: Input images: (a)input image array (b)side view of the distance image paper, we concentrate on the problem of segmenting events from vision information. Information at a low degree of abstraction is an advantageous element for the flexibility of a model. The functional part of an object is treated as attention point in space, and the speed as attention point in time. We show examples of event recognition by evaluating distance and posture between the functional parts of objects. 2 Vision Process 2.1 Observation Environment Here, we show the observation environment, composition, and input images. A side-view of the observation environment is shown in the Fig.1. Handwork performed on the desk set at a robot s front is observed by the range sensor located on the robot s head. The observable range is between about 5mm - 1 mm from the camera center. Since the resolution of the depth direction is 8 bits, precision is about 2mm. The size of an input image is 28 pixels. The input image sets are depth, color, SSAD, error, and sequence in time. The range sensor has a zoom function of 1.5 times to 2. times, obtained by using the electronic zoom function of the separate cameras(fig.2(a)). The depth image obtained is the so-called 2.5-dimensional figure.(fig.2(b)) The framerate is 3frame/sec. 2.2 Hand Shape Recognition In this section, we discuss the problem of recognizing the shape of a human s hand using the eigen space method on the distance and color images acquired by a high resolution range sensor. Hand Shape Recognition using the Eigen Space Method. Although the eigen space method is known to be used to recognize object postures, estimate their position, etc. we use it to classify hand shapes and estimate hand positions. Generally the problem is the need to use normalized data as the input of an eigen space method. The process flow of hand shape recognition is shown in Fig.3. First, skin color is extracted from a color image and the region of the human s hand is clipped. The position of a 2- dimensional image of the hand region is specified by template matching based on normalized correlation to the hand region of the last scene. The image of theobtainedhandregionispassedinanareafilter, and noise is removed. The center of gravity of the resulting image of the hand region is acquired as the position of the hand at that time. The hand region image is normalized, including the depth direction, using the size in real space already obtained from the range sensor. The depth value is treated as a vector and recognition performed using the eigen space method as follows: 1. The eigen vector of the covariance matrix of the vector group from study images is calculated. 2. The eigen space is constituted on the basis of the eigen vector with the largest eigen value. 3. Observation images and study images are projected and expressed as points in eigen space. 4. The distance between a pair of corresponding points is regarded as an evaluation measure for class resemblance. Consequently, the hand shape in the current scene is acquired from the distance values for each class. For more robust estimation, the evaluation is done continuously in time. The hand shape is obtained from the minimum value of the distance values to the second power from a number of scenes in the past.

Input Images Skin Color Detection Hand-Color Model Color Difference L* t f b* Input Images Color Model Creation via Visibility Image Plane a* SSAD Threshold Last Scene Pose Skin Color Images Normalized Correlation Masked Images Projected Images Template Model Sequence Search Last Scene Image Hand Image Normalize Model Histgram Result of Histgram Backprojection Region of Hand -Real Size -Center of Gravity Input Histgram Region of Object Histgram Backprojection Project Depth Data into Eigen-Space Hand Models Last Scene Pose Pose Estimation by 3DTM Iteration Current Scene Pose Recognition Result Output 6DOF Pose Eigen-Space :Input Image :Hand Models Figure 3: Hand-tracking sequence Figure 4: Object-tracking sequence 2.3 Object Tracking In this section, we discuss a method for 3- dimensional object tracking combining object region search in the 2-dimensional image by color and recognition of the 3-dimensional position and posture of the object by the ICP method. Tracking of the object region by color classification. Generally, searching the object region in a 2-dimensional image by color offers the advantage that one can obtain the edge of an object region clearly and at high speed, and the drawback that recognition of a 3-dimensional object is difficult. On the other hand, the ICP method can estimate the posture of an object with complicated form, but it needs exact range data. However, with the stereo vision system based on area matching used in this research, no clear range data of the edges of objects is obtained. In this paper, we argue that in combin- ing the two techniques, they compensate each other s faults. The process flow of object region tracking by color is shown in Fig.4. Color histograms are generated from an observation image and the silhouette of an object model, respectively, and the object region is extracted by comparison. The vote box I[r][g][b], a 3-dimensional array which divides into Q c onst pieces each of the 3-dimensions of color space for storing a color histogram, is prepared. The color histogram of an observation image is generated by scanning the pixels one by one, and voting a corresponding color to one corresponding vote box. Next, the color histogram of an object model is generated. Object information is given beforehand as a geometric model and a color model. The objects, such as a tool, are usually colored by two or more colors. Since there are differences in the views of an object depending on its posture, the color

of the mesh of the last observed scene is used to draw a model image. Like the color histogram of the observation image, the 3-dimensional vote box for the object model histogram is prepared, a projection image of the object s posture is scanned pixel by pixel, and the color histogram of the model is generated. Next, the degree of similarity of the two histogram is checked. The ratio R[r][g][b] of the color histogram of an object model and an observation image is calculated, and normalized so that its size is in the range -1. R [r][g][b] = min( M [r][g][b], 1) (1) I [r][g][b] If an observation image is replaced with the value of this ratio, a region with a color close to the object s color will take a value nearer to 1, and the object region will be extracted. Labeling of the object region is carried out, and region with values below a threshold are removed. The depth image is masked by the pixels which remained and an output in form of 3-dimensional point group data is obtained. Estimation of object s posture by the ICP method. The 6DOF position and posture are obtained by fitting the 3-dimensional point group data and the point of an object geometry model s mesh. In this paper, the ICP method known as 3-d template matching is used. Since this method uses M- estimators for optimization of a posture, it handles well the noise in the observation image and the incorrect correspondence between the object model and the observation image. The estimation of the object posture can be performed based on the following four procedures. 1. Visible points of the object model are evaluated from the normal direction of the meshes of the object model and the viewpoint. 2. The closest correspondence between visible points of an object model and the dots of the object region in the observation image is found. 3. An initial position and posture are estimated using the least-squares method. 4. An optimization is performed by repeated correction so that the distance between corresponding points becomes small, using a statistical M- estimator technique. In the repetition operation of the ICP method, it happens that the object position and posture do not converge, but oscillate. To remedy this, a kalman filter is applied. Moreover, an object speed can also be obtained from the state value estimated by the kalman filter, and used for the task analysis. This is shown in the following chapter. Pinch-1 Pinch-2 Grip Receive Figure 5: Hand classes f(3) f(2) f Driver Block-A Block-B f(2) f(1) 3 Task Analysis Figure 6: Objects functions In this section, we explain task analysis on an example in part assembly. The analysis is based on the segmentation of events by paying attention to object speed and position. 3.1 Primitives for the task analysis We defined four hand shape classes that will always be well visible at the time of work. The Eigen space was generated using study images of 25 pose variations of each hand shape. Here, the eigen space of a right hand and left hand was prepared, and both hands have been recognized. The object for observation is shown in Fig.6. The point which should be noted is defined as a functional part on the object s geometric form. The object s functional part such as a hole or an axis is expressed by a coordinate system on the object. It was manually determined a priori. The event to recognize is defined as the state of an operation, expressed as relations between functional parts of two objects, and the kind of operation is described as a predicate. For example, the event in which the right hand holds the object Obj s functional part 1 is described: Right-Hand Grasped Obj.f [1]. The objects and operations which are used as observation primitives in this paper are arranged in Table3.1. An example of the contents of a task is shown in Fig.7. There is the task of attaching Block-A and Block-B, and of them fixing using a f(1) f(1) f(2)

(1) Initial Scene (2) Right-Hand Grasped Block-A (3) Left-Hand Grasped Block-B Block-A & Block-B Fixed (5) Right-Hand Released Block-A (6) Right-Hand Grasped Driver (7) Block-AB & Driver Fixed (8) Right-Hand Released Driver Figure 7: Observation results 165 Z[mm] 16 155 15 145 1 135 125 11 1 Hand-Right (8) (6) 1 Y[mm] (5) (2)(3) (7) (5)(6)(7) (8) Hand-Left -3 - -1 1 3 X[mm] (3) 165 Z[mm] 16 155 Driver 15 145 1 (7) 135 Block-A Block-B (5)(6)(7)(8) (8) 125 11 (1)(2)(3)(5)(6) (1)(2)(3) (5)(6)(7)(8) (1)(2)(3) 1-3 - -1 1 3 1 X[mm] Y[mm] Figure 8: Trajectory of hands Figure 9: Trajectory of objects Screwdriver. The trajectories of hands and objects obtained from vision are shown in Fig.8 and Fig.9. Table3-1 Observation elements object Right-Hand operation pinch-1 Left-Hand pinch-2 Block-A grip Block-B receive Driver release fix 3.2 Event detection by speed of an object s functional part If we assume that only man s hand and the object move, observation will focus on the thing which is moving. Since people have a tendency to perform operations involving object contact slowly, carefully, speed can be used as feature to detect an event. The model controlling attention for event detection is defined by making the speed of a functional part of an object into a parameter. States of attention are disregard, caution, and gaze. The allowed change of states is disregard caution gaze disregard. The gaze state is defined as a trigger for event detection by geometric evaluation. When an object s functional part moves from a still state to a state of operation, the state it changed from disregard to caution. Event detection is triggered when shifting from operation to stillness, i.e. the state is changed from caution to gaze. If the system stays in the gaze state for a while, it will shift to the disregard state. 3.3 Event detection by position of object s functional parts The event detection by spatial evaluation is performed as follows: The distances between each functional part of objects whose positions were expressed within the world coordinate system are calculated. The position of the center-of-gravity and the finger-

[mm/sec] (1) (2) (3) (5) (6) (7) (8) 6 Hand-Right 6 Hand-Left 6 Block-A 6 Block-B 6 Driver. 2. 4. 6. 8. 1. 12. 14. 16. 18. 2. [sec] (a)object Speed Hand-Right Shape Class (b) Hand-Shape Recognition Results Hand-Left Shape Class Pinchi-1 Pinch-2 Grip Receive Figure 1: Analysis of observation results tips of the human s hand in a world coordinate system are calculated. The functional part in the gaze state and other functional parts are scanned one by one. The relation between two functional parts is evaluated based on distance and posture, and it is judged as follows whether an event happened or not: When the distance between functional parts is less than a threshold value, contact is recognized. Furthermore, when the posture of the coordinate system of two functional parts is in agreement, it is recognized that axis alignment was performed. The method of grasping operation of the human s hand on the object is described by the shape of the human hand, such as Pitch, Grip, etc. 3.4 The task analysis experiment The result of task analysis based on the speed of objects and hand shape is shown in Fig.1. The detected events were able to be described as follows with the time they occured at marked in parentheses. (1)Initial scene, (2)Right-Hand Grasped Block- A.f[1], (3)Left-Hand Grasped Block-B.f[1], Block- A.f[2] Block-B.f[2] Fixed, (5)Right-Hand Released Block-A.f[1], (6)Right-Hand Grasped Driver.f[1], (7)Driver.f[2] Block-A.f[3] Fixed, (8)Right-Hand Released Driver.f[1]. 4 Summary This paper showed a method for tracking the position and posture of objects by vision in real space and real time using a high-speed range sensor of high resolution, and also showed a method for recognizing hand shapes. Moreover, we showed a framework for task analysis and showed an experimental result for the example of assembly work on objects. Acknowledgments This work is supported in part by the Japan Science and Technology Corporation (JST) under the Ikeuchi CREST project, and in part by the Grant-in-Aid for Scientific Research on Priority Areas (C) 1322425 of the Ministry of Education, Culture, Sports, Science and Technology. References [1] K.Ikeuchi,T.Suehiro: Toward an Assembly Plan from Observation Part I: Task Recognition With Polyhedral Objects, IEEE Trans. Robotics and Automation,Vol.1,No.3,pp.368-384,1994. [2] Y.Kuniyoshi,M.Inaba,H.Inoue: Learning by Watching: Extracting Reusable Task Knowledge from Visual Observation of Human Performance,IEEE Trans. Robotics and Automation,Vol.1,No.6,1994. [3] R.Zollener,O.Rogalla,J.M.Zollner,R.Dillman: Dynamic Grasp Recognition within the Framework of Programming by Demonstration,Int. Conf. of Roman,1. [4] M.R.Cutkosky,R.D.How: Human grasp choice and robotic grasp analysis,dextrous Robot Hands,SpringerVerlag,pp.5-31,199. [5] S.B.Kang,K.Ikeuchi: Grasp Recognition and Manipulative Motion Characterization from Human Hand Motion Sequences,Proc. of ICRA94,pp.1759-1764,1994. [6] H.Kimura,T.Horiuchi: Task-Model Based Human Robot Cooperation Using Vision, Proc. of IROS1999,pp.71-76,1999. [7] M.J.Swain,D.H.Ballard: Color Indexing,Int. Journal of Computer Vision,vol7,pp11-32,1991. [8] T.Kanade,H.Kano,S.Kimura,E.Kawamura,A.Yoshida,K.Oda: Development of a Video-Rate Stereo Machine, Journal of the Robotics Society of Japan, Vol.15,No.2,pp.99-15,1997. [9] K.Kawamura,K.Hasegawa,Y.Someya,Y.Sato,K.Ikeuchi: Robust Localization for 3D Object Recognition Using Local EGI and 3D Template Matching with M- Estimators,Proc. of ICRA,Vol.2,1848-1855,.