Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Similar documents
Table of Contents. Chapter 1. Modeling and Identification of Serial Robots... 1 Wisama KHALIL and Etienne DOMBRE

Written exams of Robotics 2

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

Robotics 2 Visual servoing

Robotics 2 Iterative Learning for Gravity Compensation

CS4733 Class Notes, Computer Vision

Region-based Segmentation

Multiple View Geometry

Parallel Robots. Mechanics and Control H AMID D. TAG HI RAD. CRC Press. Taylor & Francis Group. Taylor & Francis Croup, Boca Raton London NewYoric

What is Computer Vision?

ROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino

Moving Object Tracking in Video Using MATLAB

Robot vision review. Martin Jagersand

COS Lecture 10 Autonomous Robot Navigation

Automatic Control Industrial robotics

A comparison between Position Based and Image Based Visual Servoing on a 3 DOFs translating robot

Dynamic Time Warping for Binocular Hand Tracking and Reconstruction

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

calibrated coordinates Linear transformation pixel coordinates

EE 701 ROBOT VISION. Segmentation

Robot Vision Control of robot motion from video. M. Jagersand

Lecture «Robot Dynamics»: Kinematic Control

Finite Element Method. Chapter 7. Practical considerations in FEM modeling

Image processing and features

Keeping features in the camera s field of view: a visual servoing strategy

Computer Vision & Digital Image Processing. Image segmentation: thresholding

Visual Recognition: Image Formation

Image Processing Fundamentals. Nicolas Vazquez Principal Software Engineer National Instruments

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 1: Introduction

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR

Hand-Eye Calibration from Image Derivatives

Outline 7/2/201011/6/

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

Experiments with Edge Detection using One-dimensional Surface Fitting

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

A NOUVELLE MOTION STATE-FEEDBACK CONTROL SCHEME FOR RIGID ROBOTIC MANIPULATORS

10/25/2018. Robotics and automation. Dr. Ibrahim Al-Naimi. Chapter two. Introduction To Robot Manipulators

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Camera calibration. Robotic vision. Ville Kyrki

Visual Tracking of Unknown Moving Object by Adaptive Binocular Visual Servoing

Dense 3D Reconstruction. Christiano Gava

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Inverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector

3D Geometry and Camera Calibration

MAPI Computer Vision. Multiple View Geometry

Computer and Machine Vision

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

Feature Detectors and Descriptors: Corners, Lines, etc.

Structured light 3D reconstruction

Physiological Motion Compensation in Minimally Invasive Robotic Surgery Part I

Digital Image Processing COSC 6380/4393

HOUGH TRANSFORM CS 6350 C V

Gaze interaction (2): models and technologies

Geometric camera models and calibration

Perception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich.

Requirements for region detection

Binary Image Processing. Introduction to Computer Vision CSE 152 Lecture 5

Local qualitative shape from stereo. without detailed correspondence. Extended Abstract. Shimon Edelman. Internet:

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Towards the completion of assignment 1

Centre for Digital Image Measurement and Analysis, School of Engineering, City University, Northampton Square, London, ECIV OHB

Robots are built to accomplish complex and difficult tasks that require highly non-linear motions.

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.

Vision Review: Image Formation. Course web page:

Prof. Fanny Ficuciello Robotics for Bioengineering Trajectory planning

ELEC Dr Reji Mathew Electrical Engineering UNSW

Perception and Action using Multilinear Forms

Miniature faking. In close-up photo, the depth of field is limited.

Task analysis based on observing hands and objects by vision

Robot mechanics and kinematics

Lecture 16: Computer Vision

Lecture 16: Computer Vision

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Cecilia Laschi The BioRobotics Institute Scuola Superiore Sant Anna, Pisa

Structure from Motion. Prof. Marco Marcon

CHAPTER 5 MOTION DETECTION AND ANALYSIS

Stereo Image Rectification for Simple Panoramic Image Generation

EE795: Computer Vision and Intelligent Systems

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Jane Li. Assistant Professor Mechanical Engineering Department, Robotic Engineering Program Worcester Polytechnic Institute

Operation of machine vision system

2 1/2 D visual servoing with respect to planar. contours having complex and unknown shapes

Surface Registration. Gianpaolo Palma

High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond

On-line and Off-line 3D Reconstruction for Crisis Management Applications

Dense 3D Reconstruction. Christiano Gava

Image Segmentation. Selim Aksoy. Bilkent University

Image Segmentation. Selim Aksoy. Bilkent University

Basilio Bona DAUIN Politecnico di Torino

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Unit 3 Multiple View Geometry

1724. Mobile manipulators collision-free trajectory planning with regard to end-effector vibrations elimination

Editorial Manager(tm) for International Journal of. Title: A Minimalist Approach to 3D Visually-Guided Grasping for Service Robots

Range Sensors (time of flight) (1)

Computer vision: models, learning and inference. Chapter 13 Image preprocessing and feature extraction

1 Trajectories. Class Notes, Trajectory Planning, COMS4733. Figure 1: Robot control system.

Computational Foundations of Cognitive Science

Transcription:

Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level control measures used in the control loop visual servoing control is based on feedback of visual measurements position-based visual servoing image-based visual servoing hybrid visual servoing image processing is aimed at extracting numerical information referred to as image feature parameters pose estimation methods are based on the measurement of a certain number of points or correspondences numerical pose estimation methods are based on the integration of the linear mapping between the camera velocity in the operational space and the time derivative of the feature parameters in the image plane camera calibration is based on the measurement of a certain number of correspondences

Configuration of the visual system multi-camera systems information about its depth by evaluating its distance with respect to the visual system 3D vision or stereo vision mono-camera systems two images of the same object from two different poses if only a single image is available, the depth can be estimated on the basis of geometrical characteristics of the object known in advance. This cheaper and easier to calibrate lower accuracy eye-to-hand fixed location advantage is that the camera field of view does not change during the execution of the task, implying that the accuracy of such measurements is constant the manipulator occludes, in part or in whole, the view of the objects eye-in-hand mobile configuration the camera is placed on the manipulator high variability in the accuracy of measurements the accuracy becomes almost constant and is usually higher than that achievable with eye-tohand cameras hybrid configuration consisting of one or more cameras in eye-to-hand configuration, and one or more cameras in eye-in-hand configuration ensures a good accuracy throughout the workspace, while avoiding the problems of occlusions

Image processing visual information is very rich and varied complex and computational expensive transformations before it can be used for controlling a robotic system extraction of numerical information from the image image feature parameters two basic operations segmentation a representation suitable for the identification of measurable features of the image interpretation measurement of the feature parameters of the image the source information is contained in a two dimensional memory array representing the spatial sample of the image image function I(X,Y) is a vector function whose components represent the values of one or more physical quantities related to the pixel in a sampled and quantized form light intensity in the wavelengths of red, green and blue or in shades of gray (number of gray levels depends on resolution 256 gray levels)

Perspective transformation

Gray-level histogram provides the frequency of occurrence of each gray level in the image the gray levels are quantized from 0 to 255 the value h(p) of the histogram at a particular gray level p [0, 255] is the number of image pixels with gray level p if this value is divided by the total number of pixels, the histogram is termed normalized histogram

Image segmentation consists of a grouping process, by which the image is divided into a certain number of groups, referred to as segments (component of each group similar with respect to one or more characteristics) distinct objects of the environment or homogeneous object parts finding connected regions of the image grouping sets of pixels sharing common features into two-dimensional connected areas high memory usage low computational load detection of boundaries identifying the pixels corresponding to object contours and isolating them from the rest of the image the boundary of an object, once extracted, can be used to define the position and shape of the object itself complementary

Region-based segmentation obtaining connected regions by continuous merging of initially small groups of adjacent pixels into larger ones if the pixels belonging to these regions satisfy a common property, termed uniformity predicate (verifying gray level) binary segmentation or image binarization by comparing the gray level of each pixel with a threshold l the peaks of the histogram are termed modes (for the dark objects the closest minimum to the left) in the presence of multiple objects, a further elaboration is required to separate the connected regions corresponding to the single objects the gray-scale histogram is noisy and the modes are difficult to identify various techniques have been developed to increase the robustness of binary segmentation appropriate filtering of the image before binarization algorithms for automatic selection of the threshold

Boundary-based segmentation boundary-based segmentation techniques usually obtain a boundary by grouping many single local edges corresponding to local discontinuities of image gray level local edges are sets of pixels where the light intensity changes abruptly the algorithms for boundary detection derive an intermediate image based on local edges from the original gray-scale image construct short-curve segments by edge linking obtain the boundaries by joining these curve segments through geometric primitives often known in advance edge detection is essentially a filtering process whereas boundary detection is a higher level task usually requiring more sophisticated software edge detection can be performed by grouping the pixels where the magnitude of the gradient is greater than a threshold in case of simple and well-defined shapes, boundary detection becomes straightforward and segmentation reduces to the sole edge detection several edge detection techniques exist, most of them require the calculation of the gradient or of the Laplacian of function I(XI, YI )

image interpretation is the process of calculating the image feature parameters from the segments, represented in terms of boundaries or in terms of regions characterize the position, orientation and shape of the two-dimensional object Moment of a region Image interpretation area of the region, total number of pixels coordinates of the centroid central moments second order central moments inertia moments inertia tensor relative to the center of mass (eigenvalues define the principal moments of inertia, eigenvectors define the principal axes of inertia)

Pose estimation visual servoing is based on the mapping between the feature parameters of an object measured in the image plane of the camera and the operational space variables defining the relative pose of the object with respect to the camera often it is sufficient to derive a differential mapping in terms of velocity (easier to solve linear, numerical integration algoritms) (k 1) vector s, termed feature vector (normalized pixel coordinates) the problem to solve is of computing the elements of from the measurements of object feature parameters in the camera image plane consider n points on the object and the corresponding position vectors with respect to the object frame projections of points on the image plane defines the feature vector

Pose estimation the homogeneous coordinates of the points of the object with respect to the camera frame can be expressed as the homogeneous coordinates of the projections of these points on the image plane are given by n correspondences are available for n points of the object, whose coordinates are known both in the object frame and in the image plane these correspondences define a system of equations to be solved for the unknown elements of matrix if the object is in motion with respect to the camera, the feature vector s is time-varying the interaction matrix defines the linear mapping between the absolute velocity of the camera and the corresponding image plane velocity (object is fixed with respect to the base frame)

The visual servoing problem visual measurements are used to compute the end-effector pose with respect to an object observed by the camera on the basis of visual measurements elaborated in real time the goal is to reach and keep a (constant or time-varying) desired pose with respect to the observed object the direct measurements provided by the visual system are concerned with feature parameters in the image plane the robotic task is defined in the operational space position-based visual servoing termed visual servoing in the operational space image-based visual servoing termed visual servoing in the image space

Position-based visual servoing conceptually similar to the operational space control but feedback is based on the real-time estimation of the pose of the observed object with respect to the camera using visual measurements the estimation can be performed analytically iterative numerical algorithms advantage possibility of acting directly on operational space variables drawback due to the absence of direct control of the image features, the object may exit from the camera field of view during the transient or as a consequence of planning errors the feedback loop is open due to lack of visual measurements and instability may occur

Image-space visual servoing the error is defined as the difference between the value of the image feature parameters in the desired configuration and the value of the parameters measured with the camera in the current pose advantage real-time estimate of the pose of the object with respect to the camera is not required it is possible to keep the object within the camera field of view during the motion disadvantage due to the nonlinearity of the mapping between the image feature parameters and the operational space variables, singular configurations may occur, which cause instability or saturation of the control action the end-effector trajectories cannot be easily predicted in advance and may produce collisions with obstacles or joint limits violation

Comparison camera calibration position-based visual servoing is more sensitive to camera calibration errors compared to image-based visual servoing for the first approach, the presence of uncertainties on calibration parameters, both intrinsic and extrinsic, produces errors on the estimate of operational space variables that may be seen as an external disturbance in the image-based visual servoing approach, the quantities used for the computation of the control action are directly defined in the image plane and measured in pixel units the desired value of the feature parameters is measured using the camera the uncertainty affecting calibration parameters can be seen as a disturbance acting on the forward path of the control loop, where disturbance rejection capability is high geometric model of the object for position-based visual servoing, the object geometry must be known if only one camera is used, because it is necessary for pose estimation, while it may be unknown when a stereo camera system is used image-based visual servoing does not require, in principle, knowledge of the object geometry, even for mono-camera systems for both approaches, the problem of regulation to a constant set-point is presented the object is assumed to be fixed with respect to the base frame single calibrated camera, mounted on the manipulator s end-effector the end-effector frame is chosen so as to coincide with the camera frame

Position-based visual servoing visual measurements are used to estimate in real time the homogeneous transformation matrix (relative pose of object with respect to camera frame) the object is fixed with respect to the base frame the position-based visual servoing problem can be formulated by imposing a desired value to the relative pose of the object frame with respect to the camera frame homogeneous transformation matrix expressing the position and orientation displacement of the camera frame in the current pose with respect to the desired pose the error does not depend on the object pose and represents the error between the desired pose and the current pose of the camera frame

PD control with gravity compensation the control has to be designed so that the operational space error tends to zero asymptotically by deriving the error considering that

PD control with gravity compensation

Resolved-velocity control the information deriving from visual measurements is computed at a frequency lower or equal to the camera frame rate one order of magnitude lower than the typical frequencies used for motion control of robot manipulators to preserve stability the control gains must be set to values much lower than those typically used for motion control the performance of the closed-loop system in terms of speed of convergence and disturbance rejection capability turns out to be poor solution: high-gain motion controller in the joint space or in the operational space neglecting the effects on the tracking errors deriving from manipulator dynamics and disturbances the controlled manipulator can be considered as an ideal positioning device in the case of joint space motion control, the following equality holds:

Resolved-velocity control visual servoing can be achieved by computing the trajectory visual measurements on the basis of so that the operational space tracking error goes asymptotically to zero the possibility of knowing in advance the trajectory of the camera is important because, during the motion, the object may exit from the camera field of view, making visual measurements unavailable

Image-based visual servoing the object is fixed with respect to the base frame image-based visual servoing can be formulated by stipulating that the vector of the object feature parameters has a desired constant value corresponding to the desired pose of the camera a desired pose exists so that the camera pose belongs to the dexterous workspace of the manipulator and the feature parameters can be chosen as the coordinates of n points of the object n 4 for coplanar points (and no triplets of collinear points) n 6 in case of noncoplanar points the interaction matrix Ls(s, zc) depends on variables zc with zc = [ zc,1... zc,n]t, zc,i being the third coordinate of the generic feature point of the object and s the control law must be designed to guarantee that the image space error tends asymptotically to zero

PD control with gravity compensation the object is fixed with respect to the base frame

Resolved-velocity control the concept of resolved-velocity control can easily be extended to the image space the convergence to zero of the image space error ensures the asymptotic convergence of to the desired pose