Towards an Improved Understanding of Human In-hand Manipulation Capabilities

Size: px
Start display at page:

Download "Towards an Improved Understanding of Human In-hand Manipulation Capabilities"

Transcription

1 Towards an Improved Understanding of Human In-hand Manipulation Capabilities Learning from Demonstration Eugen Richter Faculty of Mathematics, Informatics and Natural Sciences Technical Aspects of Multimodal Systems 20. December 2016 E. Richter 1 / 33

2 Outline 1. Introduction 2. Capturing human motion 3. Integration of object feedback 4. Data analysis 5. The road ahead E. Richter 2 / 33

3 Introduction In-hand manipulation In-hand manipulation of objects is one of the most complex and expressive biological capabilities of the human being It requires: Stabilization of manipulated object through contact Effective control of a large number of degrees of freedom In-hand manipulation can be described as repositioning of a grasped object within a hand while maintaining a stable grasp at any given time E. Richter 3 / 33

4 Introduction Manipulation sequence In-hand manipulation consists of a sequence of stable grasps Transition between grasps must be possible (synthesis vs. demonstration) Dynamic nature of in-hand manipulation can lead to unstable states of object grasp during transition Adaptive strategy required A few of the categories of in-hand manipulation movements are: Rolling Sliding Finger gaiting E. Richter 4 / 33

5 Introduction Robotic hands E. Richter 5 / 33

6 Introduction Challenges Unstable states during grasp transitions High dimensionality (at least 20 DOF just for the hand) Non-linearity and high coupling of finger motion Generalization of capabilities to objects of different size Generalization to unknown objects Very difficult to acquire good tactile feedback Human finger motion is fast A lot of possibility for occlusion Real-time requirements E. Richter 6 / 33

7 Introduction Potential applications Robotics / Automation Dexterous object manipulation Human-Robot Interaction Manufacturing / Assembly Medical Prosthetics Exoskeletal prosthetics Assistance E. Richter 7 / 33

8 Introduction Natural hand motion Hand/finger motion synthesis is a highly active research field Primary focus is grasp stability However, not every stable grasp is natural or efficient Collecting data from human demonstration makes sense Human in-hand manipulation capabilities are highly evolved Captured motion sequences constrain the problem search space Recorded data is mostly collision free Good starting point for grasp planning, grasp optimization (adjustment) E. Richter 8 / 33

9 Introduction Curse of dimensionality The human hand has 27 DOF (including the wrist) Robotic hands usually have less Not possible to extract all degrees of freedom (especially of the thumb) Natural human hand motion is highly coupled allowing dimensionality reduction E. Richter 9 / 33

10 Introduction Trends in related work Lots of research on grasping, not so much on in-hand manipulation Despite availability of dexterous robot hands several research focused on more limited mechanisms Less fingers Underactuated systems Some focus on data from the human, others on synthetic data Available results are often presented using task-specific manipulators Applicable in strictly controlled environments Real-time requirements are still hard to fullfil Difficult to adapt to uncertain tasks or unstructured environments E. Richter 10 / 33

11 Capturing human motion Motion capture Marker-based, passive (optical) Usually done with retro-reflective marker objects attached to parts of human body But anything else is good too if the image processing approach can handle it Less invasive, yet still restrictive Marker-based, active (optical) Actively driven LEDs, pulsed or non-pulsed Unique pulse sequences simplify identification, non-pulsed LEDs processed similar to passive markers More invasive due to wiring and potentialy larger markers E. Richter 11 / 33

12 Capturing human motion Motion capture Markerless (optical) Based on 2D/3D information from stereo vision setups or RGB-D sensors Processing usually requires an existing dataset Non-invasive, but computationaly more involved Other means of motion capture: Magnetic systems (e.g. Polhemus, Ascension) Mechanical systems (e.g. CyberGlove) Inertial systems (e.g. Xsens) E. Richter 12 / 33

13 Capturing human motion Motion capture issues Occlusion and accuracy (relative, absolute) Ambient conditions (lighting), noise Sensor resolution, range of motion Calibration accuracy Broken trajectories (Occlusion) Phantom markers (Reflection) Blooming (two-for-one), flickering (intensity, volume boundaries), merging Crossover (close passing markers) Tracking continuity E. Richter 13 / 33

14 Capturing human motion Motion capture issues Marker identification problem. Loss of marker data (Noise/Occlusion). E. Richter 14 / 33

15 Capturing human motion PhaseSpace Impulse X2E Active marker-based motion capture/tracking system Very high frame capture rate (up to 960 Hz) Detector FOV is close to 60 Almost instant identification Tracking of markers and rigids Real-time filtering, cleaner data Not very user friendly API largely undocumented E. Richter 15 / 33

16 Integration of object feedback Role of tactile feedback Touch is one of the main senses used by human beings for interaction with their environment [WJ84] have shown lack of touch to introduce great difficulty of maintaining stable grasps Tactile information is indispensable for adaptive manipulation behavior Detection of object slippage Handling of deformable objects Lack of tactile feedback in today s industrial/service robots restricts their use to highly structured environments E. Richter 16 / 33

17 Integration of object feedback On-hand sensing Sensing touch/force using sensors on the hand is difficult HANDLE project data glove [HA1] Sensing patches (Tekscan) [HA1] E. Richter 17 / 33

18 Integration of object feedback Instrumented objects Estimation of contact points with object is also difficult Idea is not new - but there is not much research Different shapes resembling everyday objects IMU for orientation information Markers (AprilTag, LEDs,... ) for position tracking Tactile sensors (e.g. FSR, optical?) for force sensing Force-Torque sensor(s) for wrench information between parts... E. Richter 18 / 33

19 Integration of object feedback Shadow Can Cylindrical object (from the HANDLE project) Upgraded with IMU for orientation and relative motion 40x FSR sensors (mostly non-linear response, hysteresis) Coarse resolution of the surface area Close enough to cups, bottles, object handles, canisters, etc. Too large for true in-hand manipulation experiments E. Richter 19 / 33

20 Integration of object feedback Custom-built objects Everyday objects (e.g. kitchen) are difficult to customize Primary issues Integration of force sensing elements, F/T sensors Tracking marker integration 3D-printed objects present a convenient option Generically-shaped objects should be a good start Cylinder, cube, tetrahedron,... Good spatial resolution is key Difficult issue on curved surfaces 3D-printed object [WA16] E. Richter 20 / 33

21 Data analysis What to do with all the data? Dataset built from series of in-hand manipulation sequences Pose information of arm, hand, fingers and objects Finger contact point information Forces applied at contact points Perform hand/finger motion segmentation Search for manipulation actions, motion primitives Combine motion primitives for a full manipulation task e.g. through a graph-based approach Map motion primitives to robotic hand Current focus: Automatic segmentation (and annotation) E. Richter 21 / 33

22 Data analysis Automatic motion segmentation Inspired by [ZTH08] and [KTWZ10] and [VKK14] In-hand manipulation sequence as input Temporal segmentation in an unsupervised fashion Identify distinct motion (manipulation) actions in recorded sequence e.g. reach, pre-grasp, initial grasp, pick-up, object rotation (in-hand), put-down, release Classify segments inbetween actions as transitions Decompose manipulation actions into motion primitives Identification of manipulation actions and motion primitives using the concept of self-similarity E. Richter 22 / 33

23 Data analysis Segmentation into actions Input: Sequence M as points p i,..., p m, each represented by F = (f 1,...f N ) F is a vector of multi-modal features (pose information, contact forces, etc.) k-d tree search for local neighbors of each p i in feature space S i is the resulting neighborhood of p i with distance given in frames Sets S i are converted into a self-similarity matrix (SSM) SSM is an M x M matrix containing local neigborhood distances for each frame E. Richter 23 / 33

24 Data analysis Self-similarity Self-similarity matrix representation is used to determine action boundaries Region growing applied to self-similarity matrix through: Forward step starting at M1,1 Backward step starting at Mn,n Areas in between actions are considered transitions Example: SSM / Region growing [VKK14]. E. Richter 24 / 33

25 Data analysis Segmentation into motion primitives Construction of a neighborhood graph G M for the set of all neighborhoods S Every neighbor represents a single node Connections between nodes are established using dynamic time warping Each diagonal in the SSM translates to a connected component in the graph Adding one start/end node connected to first/last reduces warping path search to shortest path problem Each connected component in G M represents a local similarity and is thus considered a motion primitive E. Richter 25 / 33

26 Data analysis Resulting data Motion primitives can be clustered Determines frequency Indicates temporal relationship Cluster graph G C with motion primitives as nodes Search for valid warping paths between pairs m i and m j of motion primitives Construction of neighborhood graph using neighbors in the rectangle spanned around m i and m j Search for shortest path from top to bottom of spanned submatrix Shortest path satisfying inclusion criteria (length, slope) results in edge between nodes Each strongly connected component represents a cluster E. Richter 26 / 33

27 The road ahead Outlook: Motion tracking PhaseSpace tracking system is installed and operational Main LED driver is currently causing problems (Re-) Calibration is non-trivial Tests show accuracy issues outside the sweet-spot of the calibrated capture volume Standard markers too large for finger placement Custom calibration object would be awesome (in the making) So would be a custom alignment object (in the making) E. Richter 27 / 33

28 The road ahead Outlook: Motion tracking MultiLED boards required to drive LEDs directly Reverse engineering of the control sequence not very promising Placing markers on skin is very inconvenient Hand glove as thin and robust as possible Current choice: cheap running gloves Has some stretch, fits well, LEDs can be sewn onto the surface Thin enough, touch feels normal Potential issue: fingertip friction E. Richter 28 / 33

29 The road ahead Outlook: ROS integration Possible to use PhaseSpace API directly (e.g. standalone applications) C/C++, Python and even a Unity package Current state of ROS integration allows streaming of 3DOF marker and 6DOF rigid data API offers data filtering - highly undocumented (not integrated yet) Calibration must be done with PhaseSpace tools No plans to integrate into ROS - student project maybe? E. Richter 29 / 33

30 The road ahead Outlook: Instrumented objects Shadow Coke-Can too large, but still interesting for initial tests Needs to be upgraded due to unstable data from the on-board controllers Build a smaller version of the Coke-Can with higher surface resolution Minimal size is constrained by size of LED driver Placing AprilTags for pose tracking is difficult Possible to use further sensors (Kinect, Stereo vision) for object pose reconstruction E. Richter 30 / 33

31 The road ahead Outlook: Dataset recording ROS package for hand glove tracking is in the making Uses hand model from my diploma thesis Current work is focused on filtering of data dropout due to occlusion Measures to offset the joint marker positions based on data from medical studies Ultimate goal for the dataset is to make it public Several datasets [HBLS16] exist but are often not dealing with in-hand manipulation (at finger motion level) are unstructured (e.g. change of task domain, or even the sensors for capturing of hand/finger motion) E. Richter 31 / 33

32 The road ahead Outlook: Data analysis Presented approach more or less implemented, but not tested Trying to segment old data from my thesis did not perform well due to strong occlusion yielding hard to understand SSMs Exploit multi-modal nature of the capture framework to minimize data dropout Additional information obtained from the instrumented object should improve rate of discrimination during segmentation of motion primitives Will need subjects for experiments :) E. Richter 32 / 33

33 The road ahead References [HA1] [HBLS16] [KTWZ10] [VKK14] [WJ84] [ZTH08] HANDLE project (eu/fp7), Y. Huang, M. Bianchi, M. Liarokapis, and Y. Sun, Recent datasets on object manipulation: A survey, Big Data, B. Krueger, J. Tautges, A. Weber, and A. Zinke, Fast local and global similarity searches in large motion capture databases, 2010 ACM SIGGRAPH/Eurographics Symposium on Computer Animation, A. Voegele, B. Krueger, and R. Klein, Efficient unsupervised temporal segmentation of human motion, 2014 ACM SIGGRAPH/Eurographics Symposium on Computer Animation, G. Westling and R. S. Johansson, Factors influencing the force control during precision grip, Experimental Brain Research 53 (1984), no. 2, F. Zhou, F. De La Torre, and J. K. Hodgins, Aligned cluster analysis for temporal segmentation of human motion, IEEE Conference on Automatic Face and Gestures Recognition, E. Richter 33 / 33

Simulation of Deformation in Robotics

Simulation of Deformation in Robotics Asia Simulation Conference 2009 Oct. 7, 2009 Robots vs Creatures Simulation of Deformation in Robotics Shinichi Hirai Dept. Robotics Ritsumeikan Univ. http://www.ritsumei.ac.jp/se/~hirai/ Robots rigid

More information

Gesture Recognition: Hand Pose Estimation. Adrian Spurr Ubiquitous Computing Seminar FS

Gesture Recognition: Hand Pose Estimation. Adrian Spurr Ubiquitous Computing Seminar FS Gesture Recognition: Hand Pose Estimation Adrian Spurr Ubiquitous Computing Seminar FS2014 27.05.2014 1 What is hand pose estimation? Input Computer-usable form 2 Augmented Reality Gaming Robot Control

More information

Gesture Recognition Technique:A Review

Gesture Recognition Technique:A Review Gesture Recognition Technique:A Review Nishi Shah 1, Jignesh Patel 2 1 Student, Indus University, Ahmedabad 2 Assistant Professor,Indus University,Ahmadabad Abstract Gesture Recognition means identification

More information

Data-driven Approaches to Simulation (Motion Capture)

Data-driven Approaches to Simulation (Motion Capture) 1 Data-driven Approaches to Simulation (Motion Capture) Ting-Chun Sun tingchun.sun@usc.edu Preface The lecture slides [1] are made by Jessica Hodgins [2], who is a professor in Computer Science Department

More information

Available online at ScienceDirect. Procedia Computer Science 105 (2017 )

Available online at   ScienceDirect. Procedia Computer Science 105 (2017 ) Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 105 (2017 ) 264 269 2016 IEEE International Symposium on Robotics and Intelligent Sensors, IRIS 2016, 17-20 December 2016,

More information

10/25/2018. Robotics and automation. Dr. Ibrahim Al-Naimi. Chapter two. Introduction To Robot Manipulators

10/25/2018. Robotics and automation. Dr. Ibrahim Al-Naimi. Chapter two. Introduction To Robot Manipulators Robotics and automation Dr. Ibrahim Al-Naimi Chapter two Introduction To Robot Manipulators 1 Robotic Industrial Manipulators A robot manipulator is an electronically controlled mechanism, consisting of

More information

An Interactive Technique for Robot Control by Using Image Processing Method

An Interactive Technique for Robot Control by Using Image Processing Method An Interactive Technique for Robot Control by Using Image Processing Method Mr. Raskar D. S 1., Prof. Mrs. Belagali P. P 2 1, E&TC Dept. Dr. JJMCOE., Jaysingpur. Maharashtra., India. 2 Associate Prof.

More information

CS 775: Advanced Computer Graphics. Lecture 17 : Motion Capture

CS 775: Advanced Computer Graphics. Lecture 17 : Motion Capture CS 775: Advanced Computer Graphics Lecture 17 : History Study of human motion Leonardo da Vinci (1452 1519) History Study of human motion Edward J. Muybridge, 1830 1904 http://en.wikipedia.org/wiki/eadweard_muybridge

More information

Keywords: clustering, construction, machine vision

Keywords: clustering, construction, machine vision CS4758: Robot Construction Worker Alycia Gailey, biomedical engineering, graduate student: asg47@cornell.edu Alex Slover, computer science, junior: ais46@cornell.edu Abstract: Progress has been made in

More information

Stable Grasp and Manipulation in 3D Space with 2-Soft-Fingered Robot Hand

Stable Grasp and Manipulation in 3D Space with 2-Soft-Fingered Robot Hand Stable Grasp and Manipulation in 3D Space with 2-Soft-Fingered Robot Hand Tsuneo Yoshikawa 1, Masanao Koeda 1, Haruki Fukuchi 1, and Atsushi Hirakawa 2 1 Ritsumeikan University, College of Information

More information

Depth Sensors Kinect V2 A. Fornaser

Depth Sensors Kinect V2 A. Fornaser Depth Sensors Kinect V2 A. Fornaser alberto.fornaser@unitn.it Vision Depth data It is not a 3D data, It is a map of distances Not a 3D, not a 2D it is a 2.5D or Perspective 3D Complete 3D - Tomography

More information

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The

More information

Physiological Motion Compensation in Minimally Invasive Robotic Surgery Part I

Physiological Motion Compensation in Minimally Invasive Robotic Surgery Part I Physiological Motion Compensation in Minimally Invasive Robotic Surgery Part I Tobias Ortmaier Laboratoire de Robotique de Paris 18, route du Panorama - BP 61 92265 Fontenay-aux-Roses Cedex France Tobias.Ortmaier@alumni.tum.de

More information

ACE Project Report. December 10, Reid Simmons, Sanjiv Singh Robotics Institute Carnegie Mellon University

ACE Project Report. December 10, Reid Simmons, Sanjiv Singh Robotics Institute Carnegie Mellon University ACE Project Report December 10, 2007 Reid Simmons, Sanjiv Singh Robotics Institute Carnegie Mellon University 1. Introduction This report covers the period from September 20, 2007 through December 10,

More information

Why study Computer Vision?

Why study Computer Vision? Why study Computer Vision? Images and movies are everywhere Fast-growing collection of useful applications building representations of the 3D world from pictures automated surveillance (who s doing what)

More information

Intelligent Robotics

Intelligent Robotics 64-424 Intelligent Robotics 64-424 Intelligent Robotics http://tams.informatik.uni-hamburg.de/ lectures/2013ws/vorlesung/ir Jianwei Zhang / Eugen Richter University of Hamburg Faculty of Mathematics, Informatics

More information

Human body animation. Computer Animation. Human Body Animation. Skeletal Animation

Human body animation. Computer Animation. Human Body Animation. Skeletal Animation Computer Animation Aitor Rovira March 2010 Human body animation Based on slides by Marco Gillies Human Body Animation Skeletal Animation Skeletal Animation (FK, IK) Motion Capture Motion Editing (retargeting,

More information

A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots

A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots Jurgen Leitner, Simon Harding, Alexander Forster and Peter Corke Presentation: Hana Fusman Introduction/ Overview The goal of their

More information

3D Scanning. Qixing Huang Feb. 9 th Slide Credit: Yasutaka Furukawa

3D Scanning. Qixing Huang Feb. 9 th Slide Credit: Yasutaka Furukawa 3D Scanning Qixing Huang Feb. 9 th 2017 Slide Credit: Yasutaka Furukawa Geometry Reconstruction Pipeline This Lecture Depth Sensing ICP for Pair-wise Alignment Next Lecture Global Alignment Pairwise Multiple

More information

Grasp Recognition using a 3D Articulated Model and Infrared Images

Grasp Recognition using a 3D Articulated Model and Infrared Images Grasp Recognition using a 3D Articulated Model and Infrared Images Koichi Ogawara Institute of Industrial Science, Univ. of Tokyo, Tokyo, Japan Jun Takamatsu Institute of Industrial Science, Univ. of Tokyo,

More information

crops BGU Main Researchers

crops BGU Main Researchers crops BGU Main Researchers WP2 detection tasks: Ohad Ben Shahar WP4: Yael Edan Tasks T4.3, T3.7: Sigal Berman WP9 & T3.7: Amir Shapiro 1 Efi Vitzrabin Ben-Gurion University of the Negev Adaptive Sensor

More information

Online Interactive 4D Character Animation

Online Interactive 4D Character Animation Online Interactive 4D Character Animation Marco Volino, Peng Huang and Adrian Hilton Web3D 2015 Outline 4D Performance Capture - 3D Reconstruction, Alignment, Texture Maps Animation - Parametric Motion

More information

Motion Capture & Simulation

Motion Capture & Simulation Motion Capture & Simulation Motion Capture Character Reconstructions Joint Angles Need 3 points to compute a rigid body coordinate frame 1 st point gives 3D translation, 2 nd point gives 2 angles, 3 rd

More information

Animations. Hakan Bilen University of Edinburgh. Computer Graphics Fall Some slides are courtesy of Steve Marschner and Kavita Bala

Animations. Hakan Bilen University of Edinburgh. Computer Graphics Fall Some slides are courtesy of Steve Marschner and Kavita Bala Animations Hakan Bilen University of Edinburgh Computer Graphics Fall 2017 Some slides are courtesy of Steve Marschner and Kavita Bala Animation Artistic process What are animators trying to do? What tools

More information

Research Subject. Dynamics Computation and Behavior Capture of Human Figures (Nakamura Group)

Research Subject. Dynamics Computation and Behavior Capture of Human Figures (Nakamura Group) Research Subject Dynamics Computation and Behavior Capture of Human Figures (Nakamura Group) (1) Goal and summary Introduction Humanoid has less actuators than its movable degrees of freedom (DOF) which

More information

Fingertips Tracking based on Gradient Vector

Fingertips Tracking based on Gradient Vector Int. J. Advance Soft Compu. Appl, Vol. 7, No. 3, November 2015 ISSN 2074-8523 Fingertips Tracking based on Gradient Vector Ahmad Yahya Dawod 1, Md Jan Nordin 1, and Junaidi Abdullah 2 1 Pattern Recognition

More information

Animation. CS 465 Lecture 22

Animation. CS 465 Lecture 22 Animation CS 465 Lecture 22 Animation Industry production process leading up to animation What animation is How animation works (very generally) Artistic process of animation Further topics in how it works

More information

ArchGenTool: A System-Independent Collaborative Tool for Robotic Architecture Design

ArchGenTool: A System-Independent Collaborative Tool for Robotic Architecture Design ArchGenTool: A System-Independent Collaborative Tool for Robotic Architecture Design Emanuele Ruffaldi (SSSA) I. Kostavelis, D. Giakoumis, D. Tzovaras (CERTH) Overview Problem Statement Existing Solutions

More information

Anibal Ollero Professor and head of GRVC University of Seville (Spain)

Anibal Ollero Professor and head of GRVC University of Seville (Spain) Aerial Manipulation Anibal Ollero Professor and head of GRVC University of Seville (Spain) aollero@us.es Scientific Advisor of the Center for Advanced Aerospace Technologies (Seville, Spain) aollero@catec.aero

More information

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level

More information

ECG782: Multidimensional Digital Signal Processing

ECG782: Multidimensional Digital Signal Processing Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu ECG782: Multidimensional Digital Signal Processing Lecture 01 Introduction http://www.ee.unlv.edu/~b1morris/ecg782/ 2 Outline Computer Vision

More information

CS 231. Inverse Kinematics Intro to Motion Capture

CS 231. Inverse Kinematics Intro to Motion Capture CS 231 Inverse Kinematics Intro to Motion Capture Representation 1) Skeleton Origin (root) Joint centers/ bones lengths 2) Keyframes Pos/Rot Root (x) Joint Angles (q) 3D characters Kinematics study of

More information

Outline Sensors. EE Sensors. H.I. Bozma. Electric Electronic Engineering Bogazici University. December 13, 2017

Outline Sensors. EE Sensors. H.I. Bozma. Electric Electronic Engineering Bogazici University. December 13, 2017 Electric Electronic Engineering Bogazici University December 13, 2017 Absolute position measurement Outline Motion Odometry Inertial systems Environmental Tactile Proximity Sensing Ground-Based RF Beacons

More information

Computer Animation and Visualisation. Lecture 3. Motion capture and physically-based animation of characters

Computer Animation and Visualisation. Lecture 3. Motion capture and physically-based animation of characters Computer Animation and Visualisation Lecture 3. Motion capture and physically-based animation of characters Character Animation There are three methods Create them manually Use real human / animal motions

More information

A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models

A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models Emanuele Ruffaldi Lorenzo Peppoloni Alessandro Filippeschi Carlo Alberto Avizzano 2014 IEEE International

More information

Toward data-driven contact mechanics

Toward data-driven contact mechanics Toward data-driven contact mechanics KRIS HAUSER Shiquan Wang and Mark Cutkosky Joint work with: Alessio Rocchi NSF NRI #1527826 Analytical contact models are awesome Side Force closure but not form closure

More information

Three-dimensional nondestructive evaluation of cylindrical objects (pipe) using an infrared camera coupled to a 3D scanner

Three-dimensional nondestructive evaluation of cylindrical objects (pipe) using an infrared camera coupled to a 3D scanner Three-dimensional nondestructive evaluation of cylindrical objects (pipe) using an infrared camera coupled to a 3D scanner F. B. Djupkep Dizeu, S. Hesabi, D. Laurendeau, A. Bendada Computer Vision and

More information

Dynamic Time Warping for Binocular Hand Tracking and Reconstruction

Dynamic Time Warping for Binocular Hand Tracking and Reconstruction Dynamic Time Warping for Binocular Hand Tracking and Reconstruction Javier Romero, Danica Kragic Ville Kyrki Antonis Argyros CAS-CVAP-CSC Dept. of Information Technology Institute of Computer Science KTH,

More information

Chapter 1: Introduction

Chapter 1: Introduction Chapter 1: Introduction This dissertation will describe the mathematical modeling and development of an innovative, three degree-of-freedom robotic manipulator. The new device, which has been named the

More information

Computer Vision: Reconstruction, Recognition, and Interaction

Computer Vision: Reconstruction, Recognition, and Interaction Computer Vision: Reconstruction, Recognition, and Interaction Linda G. Shapiro Department of Computer Science and Engineering University of Washington 1 3 OUTLINE Reconstruction - Kari Pulli s Reconstruction

More information

Object re-grasping with Dexterous Robotic Hands

Object re-grasping with Dexterous Robotic Hands 1 Object re-grasping with Dexterous Robotic Hands Ana Carolina Grilo da Silva Januário Instituto Superior Técnico, Lisboa, Portugal Email: ana.januario@tecnico.ulisboa.pt Abstract This thesis addresses

More information

Method for designing and controlling compliant gripper

Method for designing and controlling compliant gripper IOP Conference Series: Materials Science and Engineering PAPER OPEN ACCESS Method for designing and controlling compliant gripper To cite this article: A R Spanu et al 2016 IOP Conf. Ser.: Mater. Sci.

More information

Assignment 3. Position of the center +/- 0.1 inches Orientation +/- 1 degree. Decal, marker Stereo, matching algorithms Pose estimation

Assignment 3. Position of the center +/- 0.1 inches Orientation +/- 1 degree. Decal, marker Stereo, matching algorithms Pose estimation Assignment 3 1. You are required to analyze the feasibility of designing a vision system for the robot gas station attendant. Assume that the driver parks the car so that the flap and the cap are in a

More information

Applications. Systems. Motion capture pipeline. Biomechanical analysis. Graphics research

Applications. Systems. Motion capture pipeline. Biomechanical analysis. Graphics research Motion capture Applications Systems Motion capture pipeline Biomechanical analysis Graphics research Applications Computer animation Biomechanics Robotics Cinema Video games Anthropology What is captured?

More information

Real-time gesture recognition for controlling a virtual hand

Real-time gesture recognition for controlling a virtual hand ISBN 978-1-84626-xxx-x Proceedings of 2011 International Conference on Optimization of the Robots and Manipulators (OPTIROB 2011) Sinaia, Romania, 26-28 Mai, 2011, pp. xxx-xxx Real-time gesture recognition

More information

Adaptive Gesture Recognition System Integrating Multiple Inputs

Adaptive Gesture Recognition System Integrating Multiple Inputs Adaptive Gesture Recognition System Integrating Multiple Inputs Master Thesis - Colloquium Tobias Staron University of Hamburg Faculty of Mathematics, Informatics and Natural Sciences Technical Aspects

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

CMU Facilities. Motion Capture Lab. Panoptic Studio

CMU Facilities. Motion Capture Lab. Panoptic Studio CMU Facilities Motion Capture Lab The 1700 square foot Motion Capture Lab provides a resource for behavior capture of humans as well as measuring and controlling robot behavior in real time. It includes

More information

Graph-based High Level Motion Segmentation using Normalized Cuts

Graph-based High Level Motion Segmentation using Normalized Cuts Graph-based High Level Motion Segmentation using Normalized Cuts Sungju Yun, Anjin Park and Keechul Jung Abstract Motion capture devices have been utilized in producing several contents, such as movies

More information

Mobile Manipulator Design

Mobile Manipulator Design Mobile Manipulator Design December 10, 2007 Reid Simmons, Sanjiv Singh Robotics Institute Carnegie Mellon University 1. Introduction This report provides a preliminary design for two mobile manipulators

More information

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers

More information

3D Face and Hand Tracking for American Sign Language Recognition

3D Face and Hand Tracking for American Sign Language Recognition 3D Face and Hand Tracking for American Sign Language Recognition NSF-ITR (2004-2008) D. Metaxas, A. Elgammal, V. Pavlovic (Rutgers Univ.) C. Neidle (Boston Univ.) C. Vogler (Gallaudet) The need for automated

More information

DEVELOPMENT OF THE FORWARD KINEMATICS FOR ROBOT FINGERS BY USING ROBOREALM

DEVELOPMENT OF THE FORWARD KINEMATICS FOR ROBOT FINGERS BY USING ROBOREALM IJRRAS 7 () May www.arpapress.com/volumes/vol7issue/ijrras_7.pdf DEVELOPMENT OF THE FORWARD KINEMATICS FOR ROBOT FINGERS BY USING ROBOREALM Jamaludin Jalani Dept. of Mechanical Engineering, University

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 WRI C225 Lecture 02 130124 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Basics Image Formation Image Processing 3 Intelligent

More information

A Real-Time Hand Gesture Recognition for Dynamic Applications

A Real-Time Hand Gesture Recognition for Dynamic Applications e-issn 2455 1392 Volume 2 Issue 2, February 2016 pp. 41-45 Scientific Journal Impact Factor : 3.468 http://www.ijcter.com A Real-Time Hand Gesture Recognition for Dynamic Applications Aishwarya Mandlik

More information

Analyzing and Segmenting Finger Gestures in Meaningful Phases

Analyzing and Segmenting Finger Gestures in Meaningful Phases 2014 11th International Conference on Computer Graphics, Imaging and Visualization Analyzing and Segmenting Finger Gestures in Meaningful Phases Christos Mousas Paul Newbury Dept. of Informatics University

More information

CS 775: Advanced Computer Graphics. Lecture 8 : Motion Capture

CS 775: Advanced Computer Graphics. Lecture 8 : Motion Capture CS 775: Advanced Computer Graphics Lecture 8 : History Study of human motion Leonardo da Vinci (1452 1519) History Study of human motion Edward J. Muybridge, 1830 1904 http://en.wikipedia.org/wiki/eadweard_muybridge

More information

International Journal of Advanced Research in Computer Science and Software Engineering

International Journal of Advanced Research in Computer Science and Software Engineering Volume 2, Issue 10, October 2012 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Force Closure

More information

CS 231. Inverse Kinematics Intro to Motion Capture. 3D characters. Representation. 1) Skeleton Origin (root) Joint centers/ bones lengths

CS 231. Inverse Kinematics Intro to Motion Capture. 3D characters. Representation. 1) Skeleton Origin (root) Joint centers/ bones lengths CS Inverse Kinematics Intro to Motion Capture Representation D characters ) Skeleton Origin (root) Joint centers/ bones lengths ) Keyframes Pos/Rot Root (x) Joint Angles (q) Kinematics study of static

More information

Mechanical structure of a robot=skeleton of human body Study of structure of a robot=physical structure of the manipulator structure

Mechanical structure of a robot=skeleton of human body Study of structure of a robot=physical structure of the manipulator structure UNIT I FUNDAMENTALS OF ROBOT Part A 1. Define Robot. An industrial robot is a re-programmable, multifunctional manipulator designed to move materials, parts, tools, or specialized devices through variable

More information

Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements

Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements M. Lourakis and E. Hourdakis Institute of Computer Science Foundation for Research and Technology Hellas

More information

Advanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation

Advanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation Advanced Vision Guided Robotics David Bruce Engineering Manager FANUC America Corporation Traditional Vision vs. Vision based Robot Guidance Traditional Machine Vision Determine if a product passes or

More information

ROBOT SENSORS. 1. Proprioceptors

ROBOT SENSORS. 1. Proprioceptors ROBOT SENSORS Since the action capability is physically interacting with the environment, two types of sensors have to be used in any robotic system: - proprioceptors for the measurement of the robot s

More information

Optimal motion trajectories. Physically based motion transformation. Realistic character animation with control. Highly dynamic motion

Optimal motion trajectories. Physically based motion transformation. Realistic character animation with control. Highly dynamic motion Realistic character animation with control Optimal motion trajectories Physically based motion transformation, Popovi! and Witkin Synthesis of complex dynamic character motion from simple animation, Liu

More information

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 1: Introduction

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 1: Introduction MCE/EEC 647/747: Robot Dynamics and Control Lecture 1: Introduction Reading: SHV Chapter 1 Robotics and Automation Handbook, Chapter 1 Assigned readings from several articles. Cleveland State University

More information

Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images

Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images 1 Introduction - Steve Chuang and Eric Shan - Determining object orientation in images is a well-established topic

More information

Intelligent Robotics

Intelligent Robotics 64-424 Intelligent Robotics 64-424 Intelligent Robotics http://tams.informatik.uni-hamburg.de/ lectures/2014ws/vorlesung/ir Jianwei Zhang / Eugen Richter University of Hamburg Faculty of Mathematics, Informatics

More information

Efficient Grasping from RGBD Images: Learning Using a New Rectangle Representation. Yun Jiang, Stephen Moseson, Ashutosh Saxena Cornell University

Efficient Grasping from RGBD Images: Learning Using a New Rectangle Representation. Yun Jiang, Stephen Moseson, Ashutosh Saxena Cornell University Efficient Grasping from RGBD Images: Learning Using a New Rectangle Representation Yun Jiang, Stephen Moseson, Ashutosh Saxena Cornell University Problem Goal: Figure out a way to pick up the object. Approach

More information

LUMS Mine Detector Project

LUMS Mine Detector Project LUMS Mine Detector Project Using visual information to control a robot (Hutchinson et al. 1996). Vision may or may not be used in the feedback loop. Visual (image based) features such as points, lines

More information

AMR 2011/2012: Final Projects

AMR 2011/2012: Final Projects AMR 2011/2012: Final Projects 0. General Information A final project includes: studying some literature (typically, 1-2 papers) on a specific subject performing some simulations or numerical tests on an

More information

Generality and Simple Hands

Generality and Simple Hands Generality and Simple Hands with Grasp Invariance Matthew T. Mason and Alberto Rodriguez February 18, 2011 Outline 1. Overview of Robotics Institute (Mason) 2. Simple hands project (Mason) 3. Grasp invariance

More information

Dynamic Controllers in Character Animation. Jack Wang

Dynamic Controllers in Character Animation. Jack Wang Dynamic Controllers in Character Animation Jack Wang Overview Definition Related Work Composable Controllers Framework (2001) Results Future Work 2 Physics-based Animation Dynamic Controllers vs. Simulation

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational

More information

HUMAN COMPUTER INTERFACE BASED ON HAND TRACKING

HUMAN COMPUTER INTERFACE BASED ON HAND TRACKING Proceedings of MUSME 2011, the International Symposium on Multibody Systems and Mechatronics Valencia, Spain, 25-28 October 2011 HUMAN COMPUTER INTERFACE BASED ON HAND TRACKING Pedro Achanccaray, Cristian

More information

Feature Tracking and Optical Flow

Feature Tracking and Optical Flow Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who 1 in turn adapted slides from Steve Seitz, Rick Szeliski,

More information

Registration of Dynamic Range Images

Registration of Dynamic Range Images Registration of Dynamic Range Images Tan-Chi Ho 1,2 Jung-Hong Chuang 1 Wen-Wei Lin 2 Song-Sun Lin 2 1 Department of Computer Science National Chiao-Tung University 2 Department of Applied Mathematics National

More information

Learning Articulated Skeletons From Motion

Learning Articulated Skeletons From Motion Learning Articulated Skeletons From Motion Danny Tarlow University of Toronto, Machine Learning with David Ross and Richard Zemel (and Brendan Frey) August 6, 2007 Point Light Displays It's easy for humans

More information

Research and Literature Review on Developing Motion Capture System for Analyzing Athletes Action

Research and Literature Review on Developing Motion Capture System for Analyzing Athletes Action International Conference on Education Technology, Management and Humanities Science (ETMHS 2015) Research and Literature Review on Developing Motion Capture System for Analyzing Athletes Action HAN Fang

More information

Path Planning. Marcello Restelli. Dipartimento di Elettronica e Informazione Politecnico di Milano tel:

Path Planning. Marcello Restelli. Dipartimento di Elettronica e Informazione Politecnico di Milano   tel: Marcello Restelli Dipartimento di Elettronica e Informazione Politecnico di Milano email: restelli@elet.polimi.it tel: 02 2399 3470 Path Planning Robotica for Computer Engineering students A.A. 2006/2007

More information

Margarita Grinvald. Gesture recognition for Smartphones/Wearables

Margarita Grinvald. Gesture recognition for Smartphones/Wearables Margarita Grinvald Gesture recognition for Smartphones/Wearables Gestures hands, face, body movements non-verbal communication human interaction 2 Gesture recognition interface with computers increase

More information

Vision-Based Hand Detection for Registration of Virtual Objects in Augmented Reality

Vision-Based Hand Detection for Registration of Virtual Objects in Augmented Reality International Journal of Future Computer and Communication, Vol. 2, No. 5, October 213 Vision-Based Hand Detection for Registration of Virtual Objects in Augmented Reality Kah Pin Ng, Guat Yew Tan, and

More information

16-662: Robot Autonomy Project Bi-Manual Manipulation Final Report

16-662: Robot Autonomy Project Bi-Manual Manipulation Final Report 16-662: Robot Autonomy Project Bi-Manual Manipulation Final Report Oliver Krengel, Abdul Zafar, Chien Chih Ho, Rohit Murthy, Pengsheng Guo 1 Introduction 1.1 Goal We aim to use the PR2 robot in the search-based-planning

More information

CS283: Robotics Fall 2016: Sensors

CS283: Robotics Fall 2016: Sensors CS283: Robotics Fall 2016: Sensors Sören Schwertfeger / 师泽仁 ShanghaiTech University Robotics ShanghaiTech University - SIST - 23.09.2016 2 REVIEW TRANSFORMS Robotics ShanghaiTech University - SIST - 23.09.2016

More information

Comparison of Reconstruction Methods for Computed Tomography with Industrial Robots using Automatic Object Position Recognition

Comparison of Reconstruction Methods for Computed Tomography with Industrial Robots using Automatic Object Position Recognition 19 th World Conference on Non-Destructive Testing 2016 Comparison of Reconstruction Methods for Computed Tomography with Industrial Robots using Automatic Object Position Recognition Philipp KLEIN 1, Frank

More information

Inverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector

Inverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector Inverse Kinematics Given a desired position (p) & orientation (R) of the end-effector q ( q, q, q ) 1 2 n Find the joint variables which can bring the robot the desired configuration z y x 1 The Inverse

More information

Markerless human motion capture through visual hull and articulated ICP

Markerless human motion capture through visual hull and articulated ICP Markerless human motion capture through visual hull and articulated ICP Lars Mündermann lmuender@stanford.edu Stefano Corazza Stanford, CA 93405 stefanoc@stanford.edu Thomas. P. Andriacchi Bone and Joint

More information

High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond

High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond New generation of interfaces Instead of interacting through indirect input devices (mice and keyboard), the user is interacting

More information

A MULTI-ROBOT SYSTEM FOR ASSEMBLY TASKS IN AUTOMOTIVE INDUSTRY

A MULTI-ROBOT SYSTEM FOR ASSEMBLY TASKS IN AUTOMOTIVE INDUSTRY The 4th International Conference Computational Mechanics and Virtual Engineering COMEC 2011 20-22 OCTOBER 2011, Brasov, Romania A MULTI-ROBOT SYSTEM FOR ASSEMBLY TASKS IN AUTOMOTIVE INDUSTRY A. Fratu 1

More information

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery 1 Charles TOTH, 1 Dorota BRZEZINSKA, USA 2 Allison KEALY, Australia, 3 Guenther RETSCHER,

More information

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate

More information

PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO

PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO Stefan Krauß, Juliane Hüttl SE, SoSe 2011, HU-Berlin PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO 1 Uses of Motion/Performance Capture movies games, virtual environments biomechanics, sports science,

More information

Last Time? Animation, Motion Capture, & Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation

Last Time? Animation, Motion Capture, & Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation Last Time? Animation, Motion Capture, & Inverse Kinematics Navier-Stokes Equations Conservation of Momentum & Mass Incompressible Flow Today How do we animate? Keyframing Procedural Animation Physically-Based

More information

WIRELESS VEHICLE WITH ANIMATRONIC ROBOTIC ARM

WIRELESS VEHICLE WITH ANIMATRONIC ROBOTIC ARM WIRELESS VEHICLE WITH ANIMATRONIC ROBOTIC ARM PROJECT REFERENCE NO. : 37S0918 COLLEGE : P A COLLEGE OF ENGINEERING, MANGALORE BRANCH : ELECTRONICS & COMMUNICATION GUIDE : MOHAMMAD RAFEEQ STUDENTS : CHARANENDRA

More information

Introduction to Digitization Techniques for Surgical Guidance

Introduction to Digitization Techniques for Surgical Guidance Introduction to Digitization Techniques for Surgical Guidance Rebekah H. Conley, MS and Logan W. Clements, PhD Department of Biomedical Engineering Biomedical Modeling Laboratory Outline Overview of Tracking

More information

animation projects in digital art animation 2009 fabio pellacini 1

animation projects in digital art animation 2009 fabio pellacini 1 animation projects in digital art animation 2009 fabio pellacini 1 animation shape specification as a function of time projects in digital art animation 2009 fabio pellacini 2 how animation works? flip

More information

Final Project Report: Mobile Pick and Place

Final Project Report: Mobile Pick and Place Final Project Report: Mobile Pick and Place Xiaoyang Liu (xiaoyan1) Juncheng Zhang (junchen1) Karthik Ramachandran (kramacha) Sumit Saxena (sumits1) Yihao Qian (yihaoq) Adviser: Dr Matthew Travers Carnegie

More information

Spatial R-C-C-R Mechanism for a Single DOF Gripper

Spatial R-C-C-R Mechanism for a Single DOF Gripper NaCoMM-2009-ASMRL28 Spatial R-C-C-R Mechanism for a Single DOF Gripper Rajeev Lochana C.G * Mechanical Engineering Department Indian Institute of Technology Delhi, New Delhi, India * Email: rajeev@ar-cad.com

More information

BIN PICKING APPLICATIONS AND TECHNOLOGIES

BIN PICKING APPLICATIONS AND TECHNOLOGIES BIN PICKING APPLICATIONS AND TECHNOLOGIES TABLE OF CONTENTS INTRODUCTION... 3 TYPES OF MATERIAL HANDLING... 3 WHOLE BIN PICKING PROCESS... 4 VISION SYSTEM: HARDWARE... 4 VISION SYSTEM: SOFTWARE... 5 END

More information

Introduction to Medical Imaging (5XSA0) Module 5

Introduction to Medical Imaging (5XSA0) Module 5 Introduction to Medical Imaging (5XSA0) Module 5 Segmentation Jungong Han, Dirk Farin, Sveta Zinger ( s.zinger@tue.nl ) 1 Outline Introduction Color Segmentation region-growing region-merging watershed

More information

Last Time? Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation

Last Time? Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation Last Time? Inverse Kinematics Navier-Stokes Equations Conservation of Momentum & Mass Incompressible Flow Today How do we animate? Keyframing Procedural Animation Physically-Based Animation Forward and

More information