Lighting- and Occlusion-robust View-based Teaching/Playback for Model-free Robot Programming

Similar documents
View-based Programming with Reinforcement Learning for Robotic Manipulation

View-Based Teaching/Playback for Industrial Manipulators

Industrial Calibration. Chris Lewis Southwest Research Institute

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Canny Edge Based Self-localization of a RoboCup Middle-sized League Robot

Object Classification in Domestic Environments

Motion Planning of Robot Fingertips for Graspless Manipulation

Collecting outdoor datasets for benchmarking vision based robot localization

Intern Presentation:

Efficient Grasping from RGBD Images: Learning Using a New Rectangle Representation. Yun Jiang, Stephen Moseson, Ashutosh Saxena Cornell University

3D Modeling of Objects Using Laser Scanning

Motion and Tracking. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Seminar Dept. Automação e Sistemas - UFSC Scan-to-Map Matching Using the Hausdorff Distance for Robust Mobile Robot Localization

Robotic Visual Servoing. & RTX robot control in Matlab. RTX control in Matlab. Robotic Visual Servoing Overview. Robotic Visual Servoing Reminder

TEXTURE OVERLAY ONTO NON-RIGID SURFACE USING COMMODITY DEPTH CAMERA

THE FACTORY AUTOMATION COMPANY. Force Sensor. Tactile intelligence. Assembling Contouring Measuring

Task analysis based on observing hands and objects by vision

3D Editing System for Captured Real Scenes

Improving Vision-based Topological Localization by Combining Local and Global Image Features

Structured light 3D reconstruction

Robot Vision Control of robot motion from video. M. Jagersand

AUTOMATIC 3D HUMAN ACTION RECOGNITION Ajmal Mian Associate Professor Computer Science & Software Engineering

Jo-Car2 Autonomous Mode. Path Planning (Cost Matrix Algorithm)

FOREGROUND DETECTION ON DEPTH MAPS USING SKELETAL REPRESENTATION OF OBJECT SILHOUETTES

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

CS4495/6495 Introduction to Computer Vision

FIDUCIAL BASED POSE ESTIMATION ADEWOLE AYOADE ALEX YEARSLEY

High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond

Integrating Grasp Planning and Visual Feedback for Reliable Manipulation

CHAPTER 1 INTRODUCTION

Image Transformations & Camera Calibration. Mašinska vizija, 2018.

Capture of Arm-Muscle deformations using a Depth Camera

A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots

Perception and Control for a Humanoid Robot using Vision

Adaptive Gesture Recognition System Integrating Multiple Inputs

CRF Based Point Cloud Segmentation Jonathan Nation

A Simple 3D Scanning System of the Human Foot Using a Smartphone with Depth Camera

RGBD Occlusion Detection via Deep Convolutional Neural Networks

UAV Position and Attitude Sensoring in Indoor Environment Using Cameras

Fast trajectory matching using small binary images

As stated in the main document, in each exercise, the required movements of the upper

3D Shape and Indirect Appearance By Structured Light Transport

Field-of-view dependent registration of point clouds and incremental segmentation of table-tops using time-offlight

Autonomous Mobile Robot Design

Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization

Direct Methods in Visual Odometry

Shortest Path Homography-Based Visual Control for Differential Drive Robots

Scanning and Printing Objects in 3D

Wide Area 2D/3D Imaging

Final Project Report: Mobile Pick and Place

Factorization Method Using Interpolated Feature Tracking via Projective Geometry

MONOCULAR OBJECT LOCALIZATION BY SUPERQUADRICS CURVATURE REPROJECTION AND MATCHING

A model-based approach for tool tracking in laparoscopy

Depth Sensors Kinect V2 A. Fornaser

Manipulating a Large Variety of Objects and Tool Use in Domestic Service, Industrial Automation, Search and Rescue, and Space Exploration

Active Recognition and Manipulation of Simple Parts Exploiting 3D Information

What is computer vision?

Object Shape Reconstruction and Pose Estimation by a Camera Mounted on a Mobile Robot

Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement

Robot Programming for Assembly Task

Fully Automatic Endoscope Calibration for Intraoperative Use

CS 112 The Rendering Pipeline. Slide 1

Gesture Recognition: Hand Pose Estimation. Adrian Spurr Ubiquitous Computing Seminar FS

3D Perception. CS 4495 Computer Vision K. Hawkins. CS 4495 Computer Vision. 3D Perception. Kelsey Hawkins Robotics

Learning from Successes and Failures to Grasp Objects with a Vacuum Gripper

Simultaneous Localization and Mapping (SLAM)

TEXTURE OVERLAY ONTO NON-RIGID SURFACE USING COMMODITY DEPTH CAMERA

Omnidirectional Visual Control of Mobile Robots based on the 1D Trifocal Tensor

Robot localization method based on visual features and their geometric relationship

output vector controller y s... y 2 y 1

Small Object Manipulation in 3D Perception Robotic Systems Using Visual Servoing

Dr. Enrique Cabello Pardos July

A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models

Computing the relations among three views based on artificial neural network

EE795: Computer Vision and Intelligent Systems

VISION-BASED HANDLING WITH A MOBILE ROBOT

Rotation Invariant Image Registration using Robust Shape Matching

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

Virtual Testbeds for Planetary Exploration: The Self Localization Aspect

Guided Image Super-Resolution: A New Technique for Photogeometric Super-Resolution in Hybrid 3-D Range Imaging

EE 264: Image Processing and Reconstruction. Image Motion Estimation I. EE 264: Image Processing and Reconstruction. Outline

Dept. of Mechanical and Materials Engineering University of Western Ontario London, Ontario, Canada

3D Modeling from Range Images

PROBLEM FORMULATION AND RESEARCH METHODOLOGY

Filtering and mapping systems for underwater 3D imaging sonar

Automatic 3-D 3 D Model Acquisition from Range Images. Michael K. Reed and Peter K. Allen Computer Science Department Columbia University

Localization and Map Building

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW

BabyTigers-98: Osaka Legged Robot Team

Topological Mapping. Discrete Bayes Filter

Advances in 3D data processing and 3D cameras

Robot Vision without Calibration

REFINEMENT OF COLORED MOBILE MAPPING DATA USING INTENSITY IMAGES

User-assisted Segmentation and 3D Motion Tracking. Michael Fleder Sudeep Pillai Jeremy Scott

Hand-Eye Calibration from Image Derivatives

Fast and robust techniques for 3D/2D registration and photo blending on massive point clouds

Occlusion Detection of Real Objects using Contour Based Stereo Matching

Reinforcement Learning for Appearance Based Visual Servoing in Robotic Manipulation

Dense 3D Reconstruction. Christiano Gava

Transcription:

Lighting- and Occlusion-robust View-based Teaching/Playback for Model-free Robot Programming *Yusuke MAEDA (Yokohama National University) Yoshito SAITO (Ricoh Corp)

Background Conventional Teaching/Playback still widely used model-free: neither task-specific models nor object-specific models are necessary for constant task conditions eg) initial pose of object does not change 2

When the initial object pose is not constant Object localization with cameras Model-based image processing Geometric feature extraction: edge, vertex, Pattern matching Object-specific: model-freeness is lost camera 3

Motivation To develop a model-free robot programming method that can cope with change of task conditions View-based teaching/playback : robot programming with view-based image processing [Maeda 2011 ICRA] 4

Model-based vs View-based Model-based approach with object-specific models accurate cumbersome View-based (Appearance-based) approach without object-specific models versatile 5

View-based Teaching/Playback [Maeda 2011 ICRA] View-based image processing using PCA not object-specific no need for camera calibration Adaptability to change of initial object pose using the generalization ability of neural networks generalization from multiple demonstrations 6

Related Works [Zhang et al 2000]: View-based fine positioning for grippers [Zhao et al 2008]: View-based visual servoing for relative positioning [Levine et al 2016]: View-based grasping of novel objects through massive learning 7

Overview of View-based Teaching/Playback (1/3) Human Demonstration Record all the scene images and corresponding robot motions Goal 8

Overview of View-based Teaching/Playback (2/3) Mapping Acquisition From scene image to robot motion Robot motion Mapping Image 9

Overview of View-based Teaching/Playback (3/3) View-based Playback Autonomous task execution using the acquired image-to-motion mapping Goal 10

Neural Network for Mapping numerous pixels Raw Pixel Data PCA Factor Scores FS (t) Hand Configuration Config (t) Output Layer Robot Motion Move (t) Move (t-δt) Hidden Layer Input Layer 11

Advantage of View-based Teaching/Playback Possible to cope with changes of task conditions (to some extent) Use of generalization ability of NN Model-free Multiple Demonstrations Interpolative Playback Teach2 Playback Teach1 12

Objective Original view-based teaching/playback [Maeda 2011 ICRA] sensitive to lighting conditions sensitive to occlusions To make our view-based teaching/playback lighting- and occulusion-robust Use of range images Use of subimages 13

Experimental Setup Target task: Pushing on a plane Kinect for grayscale and range images Teaching (x2) 14

Change of Lighting Condition (playback with grayscale images) Teaching (x2) Playback (x2) 15

Change of Lighting Condition (playback with range images) Teaching (x2) Playback (x2) 16

Position Errors at Goal for Each Initial Positions 260 Grayscale images used 260 Range images used x [mm] 270 280 290 300 Demonstration Error < 20 mm Error < 30 mm Error < 40 mm Error >= 40 mm Goal x [mm] 270 280 290 300 Demonstration Error < 20 mm Error < 30 mm Error < 40 mm Error >= 40 mm Goal 310 310 320 320 330 5 25 45 65 85 105 125 y [mm] 330 5 25 45 65 85 105 125 y [mm] Larger errors found for range images due to their noise 17

Grayscale Images vs Range Images Grayscale Images Less noisy, but less robust Range Images Noisy, but robust Adaptive online switching of used images Achieve both accuracy and robustness in view-based teaching/playback 18

Switching Grayscale/Range Images (1/2) Neural Network for Robot Motion Factor Scores FS (t) Hand Configuration Config (t) Move (t-δt) Output Layer Hidden Layer Relative Motion Robot Motion Move (t) Additional Neural Network for Inconsistency Detection Factor Scores FS (t) Hand Configuration Config (t) Output Layer Hidden Layer Input Layer Absolute Position Hand Configuration Config (t+δt) Input Layer Hand Configuration Config (t+δt) Consistent? 19

Switching Grayscale/Range Images (2/2) NN for Robot Motion Config(t+Δt) NN for Inconsistency Detection Config(t+Δt) Grayscale image is usable yes Error<T A T A : threshold no Unexpected situation detected 20

Lighting-robust View-based Playback (light added) Playback (x2) Used Images (x2) Range images were used automatically due to change of lighting condition 21

Lighting-robust View-based Playback (light on/off) Playback (x2) Used Images (x2) Successfully adapted to light on/off by switching neural networks 22

Making View-based Teaching/Playback Occlusion-robust Neural networks are trained for not only full images but also subimages to overcome partial occlusions 23

Switching Full/Sub Images Loop Start Inconsistency found for full image? no yes* Loop End *subimage with minimum inconsistency used 24

Occlusion-robust View-based Playback Use of subimage-based NNs Playback (x2) Used Images (x2) 25

Switching Grayscale Image, Range Full image and Range Subimages Loop Start Inconsistency found for grayscale full image? yes Inconsistency found for range full image? yes no no Loop End 26

Lighting- and Occlusion-robust View-based Playback Playback (x2) Used Images (x2) 27

Conclusion Lighting-robust view-based teaching/playback Use of range images Switching between range and grayscale images for better accuracy Occlusion-robust view-based teaching/playback Use of subimages 28

Future Work Application to various robotic tasks that require higher DOF Incorporation of various sensor modalities Force information [Nakagawa et al 2016 IAS-14] 29