Human Motion Reconstruction by Direct Control of Marker Trajectories

Similar documents
Robotics-based Synthesis of Human Motion

Journal of Physiology - Paris

PREDICTING REACHING POSTURES USING A KINEMATICALLY CONSTRAINED SHOULDER MODEL

SYNTHESIS OF WHOLE-BODY BEHAVIORS THROUGH HIERARCHICAL CONTROL OF BEHAVIORAL PRIMITIVES

Modeling of Humanoid Systems Using Deductive Approach

A Whole-Body Control Framework for Humanoids Operating in Human Environments

Last Time? Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation

A 12-DOF Analytic Inverse Kinematics Solver for Human Motion Control

Last Time? Animation, Motion Capture, & Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation

Singularity Handling on Puma in Operational Space Formulation

Research Subject. Dynamics Computation and Behavior Capture of Human Figures (Nakamura Group)

Last Time? Animation, Motion Capture, & Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation

Applications. Human and animal motion Robotics control Hair Plants Molecular motion

Documents. OpenSim Tutorial. March 10, 2009 GCMAS Annual Meeting, Denver, CO. Jeff Reinbolt, Ajay Seth, Scott Delp. Website: SimTK.

Torque-Position Transformer for Task Control of Position Controlled Robots

HUMANOID robots have recently attracted growing social

Simulating the task-level control of human motion: a methodology and framework for implementation

Using Artificial Neural Networks for Prediction Of Dynamic Human Motion

Virtual Interaction System Based on Optical Capture

Announcements: Quiz. Animation, Motion Capture, & Inverse Kinematics. Last Time? Today: How do we Animate? Keyframing. Procedural Animation

Documents. OpenSim Workshop. September 20-22, 2011 HYPER Summer School, La Alberca, Spain. Jeff Reinbolt, Jen Hicks. Website: SimTK.

Intermediate Desired Value Approach for Continuous Transition among Multiple Tasks of Robots

OpenSim Tutorial #3 Scaling, Inverse Kinematics, and Inverse Dynamics

MOTION capture is a technique and a process that

Animating Non-Human Characters using Human Motion Capture Data

EFFICIENT ALGORITHMS FOR ARTICULATED BRANCHING MECHANISMS: DYNAMIC MODELING, CONTROL, AND SIMULATION

WHOLE-BODY DYNAMIC BEHAVIOR AND CONTROL OF HUMAN-LIKE ROBOTS

Simulation. x i. x i+1. degrees of freedom equations of motion. Newtonian laws gravity. ground contact forces

David Galdeano. LIRMM-UM2, Montpellier, France. Members of CST: Philippe Fraisse, Ahmed Chemori, Sébatien Krut and André Crosnier

Thiruvarangan Ramaraj CS525 Graphics & Scientific Visualization Spring 2007, Presentation I, February 28 th 2007, 14:10 15:00. Topic (Research Paper):

TRAINING A ROBOTIC MANIPULATOR

Motion capture: An evaluation of Kinect V2 body tracking for upper limb motion analysis

Animation, Motion Capture, & Inverse Kinematics. Announcements: Quiz

Style-based Inverse Kinematics

Smooth Transition Between Tasks on a Kinematic Control Level: Application to Self Collision Avoidance for Two Kuka LWR Robots

Written exams of Robotics 2

Homework 2 Questions? Animation, Motion Capture, & Inverse Kinematics. Velocity Interpolation. Handing Free Surface with MAC

Artificial Neural Network-Based Prediction of Human Posture

Human-Centered Robotics and Interactive Haptic Simulation

Optimal motion trajectories. Physically based motion transformation. Realistic character animation with control. Highly dynamic motion

Redundancy Resolution by Minimization of Joint Disturbance Torque for Independent Joint Controlled Kinematically Redundant Manipulators

Reproduction of Human Arm Movements Using Kinect-Based Motion Capture Data

AMSTERDAM BOSTON HEIDELBERG LONDON NEW YORK OXFORD PARIS SAN DIEGO SAN FRANCISCO SINGAPORE SYDNEY TOKYO F ^ k.^

A New Algorithm for Measuring and Optimizing the Manipulability Index

Computer Animation. Algorithms and Techniques. z< MORGAN KAUFMANN PUBLISHERS. Rick Parent Ohio State University AN IMPRINT OF ELSEVIER SCIENCE

Robots are built to accomplish complex and difficult tasks that require highly non-linear motions.

Dynamics modeling of structure-varying kinematic chains for free-flying robots

A New Algorithm for Measuring and Optimizing the Manipulability Index

Optimization of a two-link Robotic Manipulator

A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models

Handwriting Trajectory Movements Controlled by a Bêta-Elliptic Model

Human Motion Synthesis by Motion Manifold Learning and Motion Primitive Segmentation

Inverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector

Using Classical Mechanism Concepts to Motivate Modern Mechanism Analysis and Synthesis Methods

Table of Contents Introduction Historical Review of Robotic Orienting Devices Kinematic Position Analysis Instantaneous Kinematic Analysis

animation projects in digital art animation 2009 fabio pellacini 1

Applications. Systems. Motion capture pipeline. Biomechanical analysis. Graphics research

Chapter 1: Introduction

Learning Deformations of Human Arm Movement to Adapt to Environmental Constraints

Controlling Humanoid Robots with Human Motion Data: Experimental Validation

Animation Lecture 10 Slide Fall 2003

Automated Parameterization of the Joint Space Dynamics of a Robotic Arm. Josh Petersen

Dynamic Human Shape Description and Characterization

Data-driven Approaches to Simulation (Motion Capture)

Motion Capture & Simulation

Vision-based Manipulator Navigation. using Mixtures of RBF Neural Networks. Wolfram Blase, Josef Pauli, and Jorg Bruske

Dynamic Simulation of Interactive Robotic Environment

Triangulation: A new algorithm for Inverse Kinematics

SCREW-BASED RELATIVE JACOBIAN FOR MANIPULATORS COOPERATING IN A TASK

AC : ADAPTIVE ROBOT MANIPULATORS IN GLOBAL TECHNOLOGY

A NOUVELLE MOTION STATE-FEEDBACK CONTROL SCHEME FOR RIGID ROBOTIC MANIPULATORS

Motion Capture, Motion Edition

Algorithm for Arm Position Reconstruction from Optical Motion Capture Data with Noisy or Missing Data

Compliant Control of Whole-body Multi-contact Behaviors in Humanoid Robots

Motion Capture. Motion Capture in Movies. Motion Capture in Games

Simulation based Design for Robotic Care Device: Optimizing Trajectory of Transfer Support Robot

Motion Control with Strokes

In-vivo Estimation of the Human Elbow Joint Dynamics during Passive Movements Based on the Musculo-skeletal Kinematics Computation

Term Project Final Report for CPSC526 Statistical Models of Poses Using Inverse Kinematics

Design & Kinematic Analysis of an Articulated Robotic Manipulator

10/25/2018. Robotics and automation. Dr. Ibrahim Al-Naimi. Chapter two. Introduction To Robot Manipulators

Automatic Generation of Animatable 3D Personalized Model Based on Multi-view Images

Constraint-Based Task Programming with CAD Semantics: From Intuitive Specification to Real-Time Control

Developing a Robot Model using System-Level Design

Physically Based Character Animation

Simulation and Modeling of 6-DOF Robot Manipulator Using Matlab Software

This week. CENG 732 Computer Animation. Warping an Object. Warping an Object. 2D Grid Deformation. Warping an Object.

Whole Body Humanoid Control From Human Motion Descriptors

KINEMATIC AND DYNAMIC SIMULATION OF A 3DOF PARALLEL ROBOT

Personal style & NMF-based Exaggerative Expressions of Face. Seongah Chin, Chung-yeon Lee, Jaedong Lee Multimedia Department of Sungkyul University

Imitation Learning of Robot Movement Using Evolutionary Algorithm

Visual Tracking of Unknown Moving Object by Adaptive Binocular Visual Servoing

CS770/870 Spring 2017 Animation Basics

CS770/870 Spring 2017 Animation Basics

Cecilia Laschi The BioRobotics Institute Scuola Superiore Sant Anna, Pisa

ARTICLE IN PRESS. Journal of Computational Science xxx (2012) xxx xxx. Contents lists available at SciVerse ScienceDirect

Computer Animation Fundamentals. Animation Methods Keyframing Interpolation Kinematics Inverse Kinematics

INPUT PARAMETERS FOR MODELS I

Analyzing and Segmenting Finger Gestures in Meaningful Phases

Animation. CS 465 Lecture 22

Transcription:

Human Motion Reconstruction by Direct Control of Marker Trajectories Emel Demircan, Luis Sentis, Vincent De Sapio and Oussama Khatib Artificial Intelligence Laboratory, Stanford University, Stanford, CA 94305, U.S.A., e-mail: {emeld, lsentis, vdesap, khatib}@stanford.edu Abstract. Understanding the basis of human movement and reproducing it in robotic environments is a compelling challenge that has engaged a multidisciplinary audience. In addressing this challenge, an important initial step involves reconstructing motion from experimental motion capture data. To this end we propose a new algorithm to reconstruct human motion from motion capture data through direct control of captured marker trajectories. This algorithm is based on a task/posture decomposition and prioritized control approach. This approach ensures smooth tracking of desired marker trajectories as well as the extraction of joint angles in real-time without the need for inverse kinematics. It also provides flexibility over traditional inverse kinematic approaches. Our algorithm was validated on a sequence of tai chi motions. The results demonstrate the efficacy of the direct marker control approach for motion reconstruction from experimental marker data. Key words: human motion synthesis, operational space formulation, task/posture decomposition, prioritization, marker space. 1 Introduction The central nervous system (CNS) is able to elegantly coordinate the complex structure of the human body to perform movements of great agility and sophistication. An effective way of understanding human movement involves mimicking motions which are optimal in performance. Such optimal movements include those exhibited by highly skilled practitioners in sports and the martial arts. Human motor performance depends on skilled motor coordination and posture control as well as physical strength and perception. Drawing inspiration from their biological counterparts humanoid robots are being imbued with skilled dynamic behaviors. Enhancing the authenticity of synthesized human motion in robotic systems has been a continuing challenge that draws together researchers from the fields of physiology, neuroscience, biomechanics, and robotics. This challenge has been addressed by researchers through retargetting methods [1, 10]. Additionally, adapting existing motion for a human character with a given set of constraints has been explored [4,9]. However, 1

2 E. Demircan et al. these techniques, which map motion-capture data to different characters and scenes, require inverse kinematic computations. We propose a new algorithm to reconstruct human movement through direct control of optical marker trajectory data. This approach directly projects marker points onto a simulated human model and tracks the trajectory in Cartesian space. Tracking the desired trajectories is accomplished using the operational space control framework [5]. With our algorithm we can drive a simulated model of the human body to experimental marker locations in real-time. This allows smooth tracking of the desired marker trajectories and extraction of joint angles within a reasonable error boundary. Further, the task/posture decomposition used in the operational space method constitutes a natural decomposition for dealing with marker data, thus avoiding the performance of inverse kinematics. The algorithm presented here is validated through a sequence of slow tai chi motions. Tai chi motions are light, lively, and balanced, and they constitute a rich variety of motions useful for testing purposes. With our new algorithm we show that human movement can be controlled and reconstructed in real-time. This facilitates the investigation of other high performance dynamic skills. 2 Direct Marker Control Approach Optical motion capture constitutes a common and effective method for capturing human motion. A series of markers attached to a subject s body are imaged by a set of cameras and the spatial positions of the markers are triangulated from the image data. A number of post processing steps need to be performed to convert the raw marker positions into useful kinematic data. The most significant step is to convert the marker trajectories into joint space trajectories. This has commonly been done using inverse kinematic techniques. As an alternative to performing inverse kinematics on marker data we propose to dynamically track the markers using a task-level control approach. We will refer to this approach as direct marker control. For the purposes of our direct marker control application, we will define task space as the space of Cartesian coordinates for the motion capture markers. However, it must be noted that marker trajectories obtained through motion capture are not independent. For example, markers on the same body link are rigidly constrained to each other and the relative motion between markers on adjacent links is limited by the freedom in the connecting joints. To accommodate for motion dependencies the markers are grouped into independent subsets, {m 1,...,m n }. Each subset, m i, is represented by a single task vector, x mi, that is formed by concatenating the Cartesian coordinates of the individual markers contained within that subset. Using a prioritized control approach a hierarchy of marker task vectors is formed where the tasks that are lower in the hierarchy are projected into the null space of the tasks that are higher in the hierarchy. At the end of this recursive process, independent subsets of marker tasks are obtained, ensuring

Human Motion Reconstruction by Direct Control of Marker Trajectories 3 the overall feasibility of the marker control. The operational space formulation [5] is then used to directly control the marker trajectories. 2.1 Prioritized Control in Marker Space In this section, we develop the proposed framework for direct control of marker trajectories. The need to address high performance dynamic behaviors and the need for flexibility over traditional inverse kinematic approaches constitute the motivation for our approach. A behavioral task generally involves descriptions of various parts of the multibody mechanism, each represented by an operational point x t(i). The full task is represented as an m 1 vector, x t, formed by vertically concatenating the coordinates of all operational points. The Jacobian associated with this task is denoted as J t. The derivation of the operational space formulation begins with the joint space dynamics of the robot [8] A q + b + g = Ɣ (1) where q is the the vector of n generalized coordinates of the articulated system, A is the n n kinetic energy matrix, b is the vector of centrifugal and Coriolis generalized forces, g is the vector of gravity forces, and Ɣ is the vector of generalized control forces. Task dynamic behavior is obtained by projecting (1) into the space associated with the task, which can be done with the following operation J T t [A q + b + g = Ɣ] = tẍ t + µ t + p t = J T t Ɣ (2) Here, J T t is the dynamically-consistent generalized inverse of J t [8], t is the m m kinetic energy matrix associated with the task and µ t and p t are the associated centrifugal/coriolis and gravity force vectors. The operational space framework [5] is used as the basis for our direct marker control algorithm. In this formulation, the task behavior is divided into a set of independent task points and the torque component for the task is determined in a manner that compensates for the dynamics in task space. For a task behavior, x t, with decoupled dynamics and unit inertial properties ẍ t = Ft, this torque is given by the force transformation Ɣ task = Jt T F t (3) where J t is the Jacobian of the task and F t is the operational space force. This operational space force is given by F t = t F t + µ t + p t (4) A task/posture decomposition allows us to represent the dynamics of a simulated human subject in a relevant task space that is complemented by a posture space. The

4 E. Demircan et al. total control torque is decomposed into two dynamically decoupled torque vectors: the torque corresponding to the commanded task behavior and the torque that only affects posture behavior in the null space of the task Ɣ = Ɣ task + Ɣ posture = J T t F t + N T t Ɣ p (5) In this expression Nt T is the null space projection matrix, and Ɣ p is the torque projected into the null space. The prioritized control framework [7, 11] is used to control the collection of marker task vectors. In this framework the torque decomposition is: where the posture torque can be rewritten as Ɣ = J T t F t + N T t (J T p F p) (6) Ɣ posture = (J p N t ) T F p = J T p t F p t (7) Consequently, Equation (5) can be represented as: Ɣ = J T t F t + J T p t F p t (8) Alternately, if an additional task is projected into the posture we express this as Ɣ = J T t 1 F t1 + J T t 2 t 1 F t2 t 1 (9) This is generalized for an arbitrary number of additional tasks Ɣ = J T t 1 F t1 + J T t 2 t 1 F t2 t 1 + +J T t n t n 1 t 1 F tn t n 1 t 1 (10) 2.2 Direct Marker Control Formulation For the application to marker space we will use m i to denote the task for a particular marker subset. Equation (10) then becomes: Ɣ = J T m 1 F m1 + J T m 2 m 1 F m2 m 1 + +J T m n m n 1 m 1 F mn m n 1 m 1 (11) The Jacobian and the force associated with marker space are deduced from the above equation as follows: J m1 F m1 J m2 m 1 F m2 m 1 J and F (12).. J mn m n 1 m 1 F mn m n 1 m 1 The overall control torque is then

Human Motion Reconstruction by Direct Control of Marker Trajectories 5 Ɣ = J T F (13) An analysis on the bounds of the joint space errors can be performed using the Jacobian associated with the marker space. We note The inverse of this relationship is x = J q (14) q = J x (15) where J is the dynamically-consistent generalized inverse of J. Joint angles obtained through prioritized control in marker space deviate from the actual values but are bounded by: q J x (16) This allows us to tune the prioritized marker controller to accommodate the desired accuracy, for given configurations. 3 Simulation and Validation A series of real-time movements performed by a tai chi master were recently captured using an optical marker system to provide validation for our new algorithm. The subject was a fifty-five year old male of average build. An eight-camera retroreflective motion capture system was used to capture his movements at a rate of 60 Hz. Among the motions performed, a sequence of slow movements were chosen for validation and real-time simulation. Marker data of the recorded motion were then segmented and smoothed using the EvART software (Motion Analysis Corporation). Marker trajectories were imported into the SAI environment [6] which allows dynamic 3D simulations. Our existing human model which consists of 25 joints, was scaled to match the anthropometry of the tai chi master. The data used in this model have been derived from SIMM models [2]. The skeleton has been modeled as a multibody system within SAI and scaled based on body segment mass-center locations [3]. Figure 1 depicts the scaled human model simulated in the SAI environment. The human motion reconstruction described in Section 2 was executed using crucial upper body joint marker trajectories (e.g. shoulder, elbow and wrist of both right and left arms). Sets of two decoupled markers were grouped into distinct marker subsets. The first subset consisted of the right shoulder and the left wrist markers and the second set consisted of the left elbow and the right wrist markers. The markers were then directly tracked through the entire movement sequence using prioritization.

6 E. Demircan et al. Fig. 1 The scaled human model of the tai chi master simulated in the SAI environment. Markers of the right shoulder and the left wrist are selected to form the first marker set to be controlled (dark spheres). The second subset is formed by the left elbow and the right wrist markers (light spheres). Joint angles over the entire trajectory were directly obtained as a natural consequence of the direct marker control approach, thus avoiding any need for computing inverse kinematics. Figure 2 illustrates the joint angles obtained through direct marker tracking. The commanded and tracked positions of the controlled markers, as well as the joint angles, were recorded during real-time simulation. Figure 3 shows the comparison between tracked and commanded positions of each tracked marker. The consistency between the two curves in each plot suggests the efficacy of the human motion reconstruction algorithm proposed. Bounds on the joint angle errors can be addressed using Equation (15). We can compute the maximum and minimum joint angle error bounds among all the joints. Thus q max = max( q) and q min = min( q) (17) where q = J x. Figure 4 shows the margin of marker position errors and the margin of joint angle errors respectively. Maximum and minimum joint angle error magnitudes vary stably over the trajectory, suggesting well bounded errors on the joint angles.

Human Motion Reconstruction by Direct Control of Marker Trajectories 7 Fig. 2 Joint angles obtained through direct control of marker data. The joint angles for the shoulder, elbow and wrist segments are shown for the corresponding arm. Smooth joint space trajectories are generated without inverse kinematics. 4 Conclusions In this paper, we presented a new algorithm to control marker trajectories based on the operational space method of Khatib [5], and task/posture decomposition using prioritization. The algorithm reconstructs human motion in real-time through smooth tracking of marker trajectories. This facilitates the extraction of joint angles without the need for inverse kinematic calculations. The algorithm also easily accommodates anthropometric scaling of the simulation model to the human subject. We validated our new algorithm through a set of tai chi movement data. The results illustrate smooth tracking of the marker trajectories in marker space. Smooth joint angles trajectories were obtained as a natural output of the marker tracking methodology. A bound on the joint space error was obtained and the results of this analysis indicated stable error bounds over the trajectory. The errors can further be decreased with a more precise camera calibration during motion capture experiment and a more accurate model scaling of the simulation. There are plans to extend our algorithm to address task/posture decomposition in the camera space of the optical motion capture system. This may exploit further advantages of our approach by accommodating arbitrary operational spaces. Additionally, where precise knowledge of the subject anthropometry is not known a priori, our approach can be adapted so that limb lengths can be adjusted until they optimally track the marker data. This provides a way of inferring more precise anthropometry through direct tracking of marker data. By extending our new algorithm

8 E. Demircan et al. Fig. 3 Tracked and goal trajectories of markers. The tracked trajectories (solid lines) are shown for markers attached to the wrist, shoulder, and elbow segments. It can be seen that the generated trajectories closely track the corresponding goal trajectories (dotted lines).

Human Motion Reconstruction by Direct Control of Marker Trajectories 9 Fig. 4 Margin of marker position errors and margin of joint angle errors over the trajectory. Joint angle error magnitudes show a stable variation over the trajectory, thus ensuring well bounded errors on the joint angles. to incorporate other operational spaces, we hope to synthesize and investigate motor control models for skilled human movement using a task-level framework. Acknowledgements The contributions of Dr. Jessica Rose, Director of Lucile Packard Children s Hospital Motion and Gait Analysis Laboratory, and the financial support of the Simbios National Center for Biomedical Computing Grant (http://simbios.stanford.edu/, NIH GM072970) are gratefully acknowledged. References 1. Choi, K. and Ko, H. (1999), On-line motion retargetting, in Proceedings Seventh Pacific Conference on Computer Graphics and Applications, Editor + Publishers?, pp. 32. 2. Delp, S. and Loan, P. (1995), A software system to develop and analyse models of musculoskeletal structures, Computers in Biology and Medicine 25, 21 34. 3. Dempster, W.T., (1955), Space requirements of the seated operator, WADC Technical Report 55-159. 4. Grochow, K., Martin, S.L., Hertzmann, A. and Popovic, Z. (2004), Style-based inverse kinematics, in ACM Transactions on Graphics (TOG), Proceedings of the 2004 SIGGRAPH Conference 23(3), 522 531. 5. Khatib, O. (1987), A unified approach for motion and force control of robot manipulators: The operational space formulation, International Journal of Robotics Research 3(1), 43Ű-53. 6. Khatib, O., Brock, O., Chang, K., Conti, F., Ruspini, D. and Sentis, L. (2002), Robotics and interactive simulation, Communications of the ACM 45(3), 46 51. 7. Khatib, O., Sentis, L., Park, J. and Warren, J. (2004), Whole-body dynamic behavior and control of human-like robots, International Journal of Humanoid Robotics 1(1), 29 43.

10 E. Demircan et al. 8. Khatib, O. (2004), Advanced Robotic Manipulation, Class Notes, Stanford University, Stanford, USA. 9. Lee, J. and Shin, S.Y. (1999), A hierarchical approach tointeractive motion editing for humanlike figures, in Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, Editor + Publishers?, pp. 39 48. 10. Savenko, A. and Clapworthy, G. (2002), Using motion analysis techniques for motion retargetting, in Proceedings Sixth International Conference on Information Visualization, IV, Editor + Publishers?, pp. 110. 11. Sentis, L. and Khatib, O. (2005), Synthesis of whole-body behaviors through hierarchical control of behavioral primitives, International Journal of Humanoid Robotics Volume+page nos.?