Proceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns. Direct Obstacle Detection and Motion. from Spatio-Temporal Derivatives

Size: px
Start display at page:

Download "Proceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns. Direct Obstacle Detection and Motion. from Spatio-Temporal Derivatives"

Transcription

1 Proceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns CAIP'95, pp , Prague, Czech Republic, Sep 1995 Direct Obstacle Detection and Motion from Spatio-Temporal Derivatives Par Fornland Computational Vision and Active Perception Laboratory (CVAP) Department of Numerical Analysis and Computing Science Royal Institute of Technology, S Stockholm, Sweden Abstract. Autonomous vehicles need a means of detecting obstructions on its path, to avoid collision. In this paper, a novel approach to obstacle detection is presented. A camera moves on a visible ground plane with the optical axis parallel to the ground. Camera motion parameters are linearly related to rst order spatio-temporal derivatives of the taken image sequence; image ow is not needed. Motion is robustly estimated using RANSAC. An error measure for each image point corresponds to the likelihood of an obstacle in that point. 1 Introduction Visual systems must detect when there is a risk of colliding with obstructions. Many such situations can be conceptually reduced to obstacles protruding above a at ground; the task of nding these protrusions is referred to as obstacle detection, for which various schemes using a moving monocular camera have been presented in the literature. The obstacle detection is performed by studying image motion, which dier between regions viewing obstacles and ground plane. In [3] a calibrated visual system was used, assuming translational motion. A reference ow calculated from the calibration information was compared to the estimated ow. But one may not know the calibration parameters, therefore all algorithms cannot rely on this assumption. It was indeed suggested [7] to nd the reference ow by moving the camera over a ground plane without obstacles. The motion and geometry are then assumed constant during the obstacle detection. Approaches not requiring a reference ow have been suggested, e.g. examining the divergence [6] and other qualities [8] of the ow. Both detected obstacles without specic assumptions about the camera motion or the scene structure, but both required dense ow estimates. Flow estimation is not yet robust [1] despite considerable eorts by the computer vision community. In [2], a representation of a translational motion and the equation of a plane were continuously updated, using long sequences. The estimated parameters enabled a prediction of image intensities of points viewing the plane. Obstacles cause bad predictions. In this paper is presented a direct framework for obstacle detection using rst-order spatio-temporal derivatives of an image sequence. World structure is neither required nor reconstructed from the images. Similar to [2], a short circuit

2 is introduced into the previously suggested procedures for obstacle detection, since no optical ow is required. Care has been taken not to restrict the motion to pure translation; instead a general framework is developed. The optical axis is parallel to the ground, which is an often avoided case. From the gradient constraint equation [1] and the second order ow equation [5], together with geometry assumptions, a linear equation relating motion parameters to image derivatives is derived. A here presented version of RANSAC [4] robustly estimates the motion from the over-constrained system of equations by disregarding a number of observations as outliers. Each observation is dened by coordinates and spatio-temporal derivatives of one point in the image plane. Two approaches are suggested, one globally estimating motion from all available observations, forming a histogram from the error measure. The other approach locally estimates the motion parameters in a number of small windows in the image. A multi-dimensional histogram is formed from the estimated parameters. In both cases, a peak is normally found at the correct motion, corresponding to ground plane points. This segments the image into obstacles and ground plane. 2 Theories and Algorithms Assume that a camera is positioned above a ground plane, as in Fig. 1 in which the X-axis is orthogonal to the paper. A 3D point P = (X; Y; Z) projects on the image at p = (fx=z; fy=z) = (x; y). If length measures are expressed in terms of the focal length, we can set f = 1. An image point may be projected from a point P on the ground plane or an obstacle Q. Also projected 3D velocities are dierent for obstacles [3] than for ground plane points. Y f Y0 (x,y) Z Q Fig. 1. The camera, the ground plane, and an obstacle P A camera moving through a static 3D world is equivalent to a rigid world moving in front of a static camera, a view which is used in this discussion. A constant rigid 3D motion can be resolved into translational (V X ; V Y ; V Z ) and rotational ( X ; Y ; Z ) components, the latter with angular velocities around an axis through the origin. This forms an ane transformation: _ P _X _Y A Z _ V X V Y V Z Y Z? Z Y A Z X? X Z A (1) X Y? Y X

3 Projecting P _ on the image leads to the second order [5] image ow, characterizing the motion of image structures with respect to three dimensional motion:!! _x _y = Z?1 (V X? xv Z )? xy X + (1 + x 2 ) Y? y Z Z?1 (V Y? yv Z )? (1 + y 2 ) X + xy Y + x Z Frequently in motion analysis, the camera is tilted as in Fig. 1 in order to avoid image points projected from close to the horizon, causing high and dicult spatial frequencies. In this research the tilt angle is zero, and the diculties must therefore be handled. Ground plane points satisfy the relation Z = Y 0 =y, and since the translational velocities only can be retrieved up to a scale factor, no restriction is imposed by scaling them with 1=Y 0, forming translational parameters C X, C Y and C Z. Figure 2 shows how the ground plane image ow depends on single motion parameters of _x = _y y 0? xy 0 y? y 2 C X C Y C Z 1 A +?xy 1 + x 2?y?(1 + y 2 ) xy x (2) 0 1 Y A (3) Z Fig. 2. The V X, V Z and Y components respectively 2.1 Motion Estimation A common assumption in computer vision research is that the intensity of any moving image structure, I(x(t); y(t); t), is constant over time, di=dt = 0. The dierential theory chain rule gives di=dt = ri _p + I t where _p = ( _x; _y) t is the image ow. The resulting equation is ri _p + I t = 0 (4) which is referred to as the gradient constraint equation (GCE). This equation is the foundation of a wide variety of techniques [1] for estimating image ow. The ow function (3) is now inserted into the GCE (4), resulting in yi x C X + yi y C Y? (xyi x + y 2 I y )C Z? (I y + xyi x + y 2 I y ) X + +(I x + x 2 I x + xyi y ) Y + (xi y? yi x ) Z +I t = 0 (5) where the coecients are the camera motion parameters, and the observations are taken from the image. The assumption that the optical axis and the X- axis are always parallel to the ground plane is violated by X- and Z-rotations, respectively. A Y -translation violates the assumption about constant height. These arguments lead to the constraints C Y = X = Z = 0.

4 2.2 The RANSAC method The RANSAC method [4] has been applied to computer vision before, but not to this problem. Given a set of observations of a parametrised function, it nds a subset of observations tting the function, disregarding other observations. A simple version ts a line to a set of points (x; y). Figure 3 (a) exemplies a dicult line-tting with synthetic data, where RANSAC performs well. The dashed line is estimated using all observations in a least squares error method. y Freq Freq 0 0 (a) x (b) Error (c) Error Fig. 3. (a) RANSAC (b) Error histogram for ground and (c) ground with obstacle A generalisation uses M noisy observations x i 2 R N of a hyperplane A x = A 1 x 1 + : : : + A N x N = 1, where A is a normal to the hyperplane, and x are coordinates. First, an integer index vector i = (i1 : : : i N ) is chosen randomly from 1 i k M, dening a preliminary hyperplane through? xi1 : : : x in t A =? 1 : : : 1 The signed orthogonal distance between a point x and a hyperplane is d =? x A? 1 = A, i.e. points on opposite sides of the plane have dierent signs. The set of observation points lying within a predened distance from the hyperplane is remembered; remaining points are disregarded as outliers. The procedure is repeated n times, each time for dierent i. The nal hyperplane is estimated with a Least Squares t of the largest remembered set. (6) 2.3 Obstacle Detection The obstacles are detected in the following manner. The RANSAC method rst globally estimates the motion parameter hyperplane. For each observation derived from the ground plane, the signed distance d (see above) is found to have an approximately Gaussian distribution N(0; ). But points viewing obstacles correspond to observation variables with dierent statistics. The RANSAC method is robust against outliers, in this case corresponding to obstacles. Histogramming techniques can be applied to provide thresholds, indirectly segmenting the image into ground and obstacles. The threshold will be delicate to estimate if the obstacles are small, as their histogram peaks then might be below the noise level. But a large obstacle corresponds to a discernible peak in the histogram. The largest allowed obstacle size, where the method starts breaking down, is investigated in a coming, longer version of this paper.

5 The motion estimation can also be performed separately in small windows of the image. For windows containing only the ground plane, the motion parameters are expected to be well estimated. But for windows viewing obstacles, the estimated motion will be erroneous. Multi-dimensional histograms are formed from the set of estimated motion parameters of the image windows, and a prominent peak should normally be found for the correct motion. Histogram bins in a neighborhood of the peak correspond to image windows viewing the ground plane, and the other bins correspond to potential obstacles. 3 Experiments Experiments were performed both for the motion estimation and the obstacle detection. Three consecutive spatially smoothed images were used. The rst experiments conrm that direct estimation of the camera motion is possible. Synthetic images with only two motion parameters, and the geometry described in Section 2 were used. Inspired by [2], only the lower part of the image was considered. The randomness of RANSAC causes a small noise, see Table 1. Ground truth Estimates C Z Y C Z Y Table 1. Motion estimation experiments without obstacles Histograms of the distance from each observation point to the estimated motion parameter hyperplane were examined. Synthetic images were produced, and as indicated by Fig. 3 (b), the shape of the histogram formed from a ground plane resembles a Gaussian distribution. Figure 3 (c) shows a histogram formed from a mix of a ground plane and a constant height obstacle. Real images were used to evaluate the obstacle detection. The exact motion parameters were unknown, but the motion was approximately translational in the (X; Z)-plane. The images view a box on a ground plane. As real images are noisy due to motion blur, sensor noise etc, and the box only have small motion parallax close to the ground, the segmentation is not expected to be very robust. The stronger local window strategy was therefore employed. The resulting binary image indicating obstacles is eroded to remove smaller regions. Thresholds and RANSAC-parameters are scene-dependent, and therefore selected accordingly. The obstacles are marked in Fig. 4, where the top of the box is not detected since only the lower part of the image is used, where the ground plane is visible. The detection works well for obstacle parts suciently high above the ground.

6 Fig. 4. Box scenes with the detected obstacles Acknowledgements: The author wishes to thank Dr. Bergholm for providing valuable experience in motion estimation. References 1. J.L. Barron, D.J. Fleet, and S.S. Beauchemin, \Performance of optical ow techniques", International Journal of Computer Vision, vol. 12, no. 1, pp. 43{77, S. Carlsson and J-O Eklundh, \Object detection using model based prediction and motion parallax", in Proceedings of the First European Conference on Computer Vision, pp. 134{138, Springer-Verlag, Apr (Antibes, France). 3. W. Enkelmann, \Obstacle detection by evaluation of optical ow elds from image sequences", Image and Vision Computing, vol. 9, no. 3, pp. 160{168, M.A. Fischler and R.C. Bolles, \Random sample consensus: A paradigm for model tting with applications to image analysis and automated cartography", Commun. ACM, vol. 24, pp. 381{395, H.C. Lounguet-Higgins and K. Prazdny, \The interpretation of a moving retinal image", in Proc. Royal Society London B-208, pp. 385{397, R. C. Nelson and Y. Aloimonos, \Obstacle avoidance using ow eld divergence", IEEE Trans. on PAMI, vol. 11, pp. 1102{1106, M. Tistarelli and G. Sandini, \Dynamic aspects in active vision", CVGIP: Image Understanding, vol. 56, pp. 108{129, July G-S Young, T-H Hong, M. Herman, and J.C.S. Yang, \Safe navigation for autonomous vehicles: A purposive and direct solution", in SPIE Int. Conf. on Intelligent Robots and Computer Vision XII, pp. 31{42, 1993.

Hand-Eye Calibration from Image Derivatives

Hand-Eye Calibration from Image Derivatives Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed

More information

A Robust Two Feature Points Based Depth Estimation Method 1)

A Robust Two Feature Points Based Depth Estimation Method 1) Vol.31, No.5 ACTA AUTOMATICA SINICA September, 2005 A Robust Two Feature Points Based Depth Estimation Method 1) ZHONG Zhi-Guang YI Jian-Qiang ZHAO Dong-Bin (Laboratory of Complex Systems and Intelligence

More information

Department of Electrical Engineering, Keio University Hiyoshi Kouhoku-ku Yokohama 223, Japan

Department of Electrical Engineering, Keio University Hiyoshi Kouhoku-ku Yokohama 223, Japan Shape Modeling from Multiple View Images Using GAs Satoshi KIRIHARA and Hideo SAITO Department of Electrical Engineering, Keio University 3-14-1 Hiyoshi Kouhoku-ku Yokohama 223, Japan TEL +81-45-563-1141

More information

3D Motion from Image Derivatives Using the Least Trimmed Square Regression

3D Motion from Image Derivatives Using the Least Trimmed Square Regression 3D Motion from Image Derivatives Using the Least Trimmed Square Regression Fadi Dornaika and Angel D. Sappa Computer Vision Center Edifici O, Campus UAB 08193 Bellaterra, Barcelona, Spain {dornaika, sappa}@cvc.uab.es

More information

Dominant plane detection using optical flow and Independent Component Analysis

Dominant plane detection using optical flow and Independent Component Analysis Dominant plane detection using optical flow and Independent Component Analysis Naoya OHNISHI 1 and Atsushi IMIYA 2 1 School of Science and Technology, Chiba University, Japan Yayoicho 1-33, Inage-ku, 263-8522,

More information

extracted occurring from the spatial and temporal changes in an image sequence. An image sequence

extracted occurring from the spatial and temporal changes in an image sequence. An image sequence Motion: Introduction are interested in the visual information that can be We from the spatial and temporal changes extracted in an image sequence. An image sequence occurring of a series of images (frames)

More information

first order approx. u+d second order approx. (S)

first order approx. u+d second order approx. (S) Computing Dierential Properties of 3-D Shapes from Stereoscopic Images without 3-D Models F. Devernay and O. D. Faugeras INRIA. 2004, route des Lucioles. B.P. 93. 06902 Sophia-Antipolis. FRANCE. Abstract

More information

Center for Automation Research, University of Maryland. The independence measure is the residual normal

Center for Automation Research, University of Maryland. The independence measure is the residual normal Independent Motion: The Importance of History Robert Pless, Tomas Brodsky, and Yiannis Aloimonos Center for Automation Research, University of Maryland College Park, MD, 74-375 Abstract We consider a problem

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Hand-Eye Calibration from Image Derivatives

Hand-Eye Calibration from Image Derivatives Hand-Eye Calibration from Image Derivatives Henrik Malm, Anders Heyden Centre for Mathematical Sciences, Lund University Box 118, SE-221 00 Lund, Sweden email: henrik,heyden@maths.lth.se Abstract. In this

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

A robust and convergent iterative approach for determining the dominant plane from two views without correspondence and calibration

A robust and convergent iterative approach for determining the dominant plane from two views without correspondence and calibration Proc. Computer Vision and Pattern Recognition (CVPR 97), pp. 508-513, June 17-19, 1997, San Juan, Puerto Rico robust and convergent iterative approach for determining the dominant plane from two views

More information

Local qualitative shape from stereo. without detailed correspondence. Extended Abstract. Shimon Edelman. Internet:

Local qualitative shape from stereo. without detailed correspondence. Extended Abstract. Shimon Edelman. Internet: Local qualitative shape from stereo without detailed correspondence Extended Abstract Shimon Edelman Center for Biological Information Processing MIT E25-201, Cambridge MA 02139 Internet: edelman@ai.mit.edu

More information

Transactions on Information and Communications Technologies vol 19, 1997 WIT Press, ISSN

Transactions on Information and Communications Technologies vol 19, 1997 WIT Press,   ISSN Hopeld Network for Stereo Correspondence Using Block-Matching Techniques Dimitrios Tzovaras and Michael G. Strintzis Information Processing Laboratory, Electrical and Computer Engineering Department, Aristotle

More information

Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp , Kobe, Japan, September 1992

Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp , Kobe, Japan, September 1992 Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp.957-962, Kobe, Japan, September 1992 Tracking a Moving Object by an Active Vision System: PANTHER-VZ Jun Miura, Hideharu Kawarabayashi,

More information

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW Thorsten Thormählen, Hellward Broszio, Ingolf Wassermann thormae@tnt.uni-hannover.de University of Hannover, Information Technology Laboratory,

More information

Real Time Biologically-Inspired Depth Maps from Spherical Flow

Real Time Biologically-Inspired Depth Maps from Spherical Flow Real Time Biologically-Inspired Depth Maps from Spherical Flow Chris McCarthy, Nick Barnes and Mandyam Srinivasan Abstract We present a strategy for generating real-time relative depth maps of an environment

More information

DEPTH AND GEOMETRY FROM A SINGLE 2D IMAGE USING TRIANGULATION

DEPTH AND GEOMETRY FROM A SINGLE 2D IMAGE USING TRIANGULATION 2012 IEEE International Conference on Multimedia and Expo Workshops DEPTH AND GEOMETRY FROM A SINGLE 2D IMAGE USING TRIANGULATION Yasir Salih and Aamir S. Malik, Senior Member IEEE Centre for Intelligent

More information

MOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE

MOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE Head-Eye Coordination: A Closed-Form Solution M. Xie School of Mechanical & Production Engineering Nanyang Technological University, Singapore 639798 Email: mmxie@ntuix.ntu.ac.sg ABSTRACT In this paper,

More information

Perceptual Grouping from Motion Cues Using Tensor Voting

Perceptual Grouping from Motion Cues Using Tensor Voting Perceptual Grouping from Motion Cues Using Tensor Voting 1. Research Team Project Leader: Graduate Students: Prof. Gérard Medioni, Computer Science Mircea Nicolescu, Changki Min 2. Statement of Project

More information

1998 IEEE International Conference on Intelligent Vehicles 587

1998 IEEE International Conference on Intelligent Vehicles 587 Ground Plane Obstacle Detection using Projective Geometry A.Branca, E.Stella, A.Distante Istituto Elaborazione Segnali ed Immagini - CNR Via Amendola 166/5, 70126 Bari Italy e-mail: [branca,stella,distante]@iesi.ba.cnr.it

More information

Optical Flow-Based Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides.

Optical Flow-Based Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides. Optical Flow-Based Motion Estimation Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides. 1 Why estimate motion? We live in a 4-D world Wide applications Object

More information

Tracking Multiple Objects in 3D. Coimbra 3030 Coimbra target projection in the same image position (usually

Tracking Multiple Objects in 3D. Coimbra 3030 Coimbra target projection in the same image position (usually Tracking Multiple Objects in 3D Jo~ao P. Barreto Paulo Peioto Coimbra 33 Coimbra 33 Jorge Batista Helder Araujo Coimbra 33 Coimbra 33 Abstract In this paper a system for tracking multiple targets in 3D

More information

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies M. Lourakis, S. Tzurbakis, A. Argyros, S. Orphanoudakis Computer Vision and Robotics Lab (CVRL) Institute of

More information

Homographies and RANSAC

Homographies and RANSAC Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position

More information

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM Karthik Krish Stuart Heinrich Wesley E. Snyder Halil Cakir Siamak Khorram North Carolina State University Raleigh, 27695 kkrish@ncsu.edu sbheinri@ncsu.edu

More information

Comparison of Temporal Filters for Optical Flow Estimation in Continuous Mobile Robot Navigation

Comparison of Temporal Filters for Optical Flow Estimation in Continuous Mobile Robot Navigation Comparison of Temporal Filters for Optical Flow Estimation in Continuous Mobile Robot Navigation Chris McCarthy 1 and Nick Barnes 2 1 Dept. Computer Science and Software Engineering The University of Melbourne

More information

PRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1. We present some preliminary results on a system for tracking 3D motion using

PRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1. We present some preliminary results on a system for tracking 3D motion using PRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1 Tak-keung CHENG derek@cs.mu.oz.au Leslie KITCHEN ljk@cs.mu.oz.au Computer Vision and Pattern Recognition Laboratory, Department of Computer Science,

More information

Optical Flow-Based Person Tracking by Multiple Cameras

Optical Flow-Based Person Tracking by Multiple Cameras Proc. IEEE Int. Conf. on Multisensor Fusion and Integration in Intelligent Systems, Baden-Baden, Germany, Aug. 2001. Optical Flow-Based Person Tracking by Multiple Cameras Hideki Tsutsui, Jun Miura, and

More information

A Summary of Projective Geometry

A Summary of Projective Geometry A Summary of Projective Geometry Copyright 22 Acuity Technologies Inc. In the last years a unified approach to creating D models from multiple images has been developed by Beardsley[],Hartley[4,5,9],Torr[,6]

More information

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences Jean-François Lalonde, Srinivasa G. Narasimhan and Alexei A. Efros {jlalonde,srinivas,efros}@cs.cmu.edu CMU-RI-TR-8-32 July

More information

British Machine Vision Conference 2 The established approach for automatic model construction begins by taking surface measurements from a number of v

British Machine Vision Conference 2 The established approach for automatic model construction begins by taking surface measurements from a number of v Segmentation of Range Data into Rigid Subsets using Planar Surface Patches A. P. Ashbrook, R. B. Fisher, C. Robertson and N. Wergi Department of Articial Intelligence The University of Edinburgh 5, Forrest

More information

LOCAL-GLOBAL OPTICAL FLOW FOR IMAGE REGISTRATION

LOCAL-GLOBAL OPTICAL FLOW FOR IMAGE REGISTRATION LOCAL-GLOBAL OPTICAL FLOW FOR IMAGE REGISTRATION Ammar Zayouna Richard Comley Daming Shi Middlesex University School of Engineering and Information Sciences Middlesex University, London NW4 4BT, UK A.Zayouna@mdx.ac.uk

More information

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press,   ISSN ransactions on Information and Communications echnologies vol 6, 996 WI Press, www.witpress.com, ISSN 743-357 Obstacle detection using stereo without correspondence L. X. Zhou & W. K. Gu Institute of Information

More information

3D object recognition used by team robotto

3D object recognition used by team robotto 3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object

More information

Structure from Motion. Prof. Marco Marcon

Structure from Motion. Prof. Marco Marcon Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)

More information

Calibrating a Structured Light System Dr Alan M. McIvor Robert J. Valkenburg Machine Vision Team, Industrial Research Limited P.O. Box 2225, Auckland

Calibrating a Structured Light System Dr Alan M. McIvor Robert J. Valkenburg Machine Vision Team, Industrial Research Limited P.O. Box 2225, Auckland Calibrating a Structured Light System Dr Alan M. McIvor Robert J. Valkenburg Machine Vision Team, Industrial Research Limited P.O. Box 2225, Auckland New Zealand Tel: +64 9 3034116, Fax: +64 9 302 8106

More information

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA

More information

A Quantitative Comparison of 4 Algorithms for Recovering Dense Accurate Depth

A Quantitative Comparison of 4 Algorithms for Recovering Dense Accurate Depth A Quantitative Comparison o 4 Algorithms or Recovering Dense Accurate Depth Baozhong Tian and John L. Barron Dept. o Computer Science University o Western Ontario London, Ontario, Canada {btian,barron}@csd.uwo.ca

More information

Marcel Worring Intelligent Sensory Information Systems

Marcel Worring Intelligent Sensory Information Systems Marcel Worring worring@science.uva.nl Intelligent Sensory Information Systems University of Amsterdam Information and Communication Technology archives of documentaries, film, or training material, video

More information

On-line and Off-line 3D Reconstruction for Crisis Management Applications

On-line and Off-line 3D Reconstruction for Crisis Management Applications On-line and Off-line 3D Reconstruction for Crisis Management Applications Geert De Cubber Royal Military Academy, Department of Mechanical Engineering (MSTA) Av. de la Renaissance 30, 1000 Brussels geert.de.cubber@rma.ac.be

More information

Stereoscopic Tracking of Bodies in Motion

Stereoscopic Tracking of Bodies in Motion Stereoscopic Tracking of Bodies in Motion Roberto CIPOLLA * Masanobu YAMAMOTO Robotics Research Group, University of Oxford Electrotechnical Laboratory, Japan In this paper we present a fast, highly efficient

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Dynamic visual attention: competitive versus motion priority scheme

Dynamic visual attention: competitive versus motion priority scheme Dynamic visual attention: competitive versus motion priority scheme Bur A. 1, Wurtz P. 2, Müri R.M. 2 and Hügli H. 1 1 Institute of Microtechnology, University of Neuchâtel, Neuchâtel, Switzerland 2 Perception

More information

Lucas-Kanade Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides.

Lucas-Kanade Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides. Lucas-Kanade Motion Estimation Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides. 1 Why estimate motion? We live in a 4-D world Wide applications Object

More information

Comparison Between The Optical Flow Computational Techniques

Comparison Between The Optical Flow Computational Techniques Comparison Between The Optical Flow Computational Techniques Sri Devi Thota #1, Kanaka Sunanda Vemulapalli* 2, Kartheek Chintalapati* 3, Phanindra Sai Srinivas Gudipudi* 4 # Associate Professor, Dept.

More information

Computing Slow Optical Flow By Interpolated Quadratic Surface Matching

Computing Slow Optical Flow By Interpolated Quadratic Surface Matching Computing Slow Optical Flow By Interpolated Quadratic Surface Matching Takashi KUREMOTO Faculty of Engineering Yamaguchi University Tokiwadai --, Ube, 755-8 Japan wu@csse.yamaguchi-u.ac.jp Kazutoshi KOGA

More information

Introduction to Computer Vision

Introduction to Computer Vision Introduction to Computer Vision Michael J. Black Nov 2009 Perspective projection and affine motion Goals Today Perspective projection 3D motion Wed Projects Friday Regularization and robust statistics

More information

Combining Appearance and Topology for Wide

Combining Appearance and Topology for Wide Combining Appearance and Topology for Wide Baseline Matching Dennis Tell and Stefan Carlsson Presented by: Josh Wills Image Point Correspondences Critical foundation for many vision applications 3-D reconstruction,

More information

Vision-based Manipulator Navigation. using Mixtures of RBF Neural Networks. Wolfram Blase, Josef Pauli, and Jorg Bruske

Vision-based Manipulator Navigation. using Mixtures of RBF Neural Networks. Wolfram Blase, Josef Pauli, and Jorg Bruske Vision-based Manipulator Navigation using Mixtures of RBF Neural Networks Wolfram Blase, Josef Pauli, and Jorg Bruske Christian{Albrechts{Universitat zu Kiel, Institut fur Informatik Preusserstrasse 1-9,

More information

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 Image Features: Local Descriptors Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 [Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 2/ 58 Local Features Detection: Identify

More information

Interpretation of Urban Surface Models using 2D Building Information Norbert Haala and Claus Brenner Institut fur Photogrammetrie Universitat Stuttgar

Interpretation of Urban Surface Models using 2D Building Information Norbert Haala and Claus Brenner Institut fur Photogrammetrie Universitat Stuttgar Interpretation of Urban Surface Models using 2D Building Information Norbert Haala and Claus Brenner Institut fur Photogrammetrie Universitat Stuttgart Geschwister-Scholl-Strae 24, 70174 Stuttgart, Germany

More information

Experiments with Edge Detection using One-dimensional Surface Fitting

Experiments with Edge Detection using One-dimensional Surface Fitting Experiments with Edge Detection using One-dimensional Surface Fitting Gabor Terei, Jorge Luis Nunes e Silva Brito The Ohio State University, Department of Geodetic Science and Surveying 1958 Neil Avenue,

More information

CH2605-4/88/0000/0082$ IEEE DETERMINATION OF CAMERA LOCATION FROM 2D TO 3D LINE AND POINT CORRESPONDENCES

CH2605-4/88/0000/0082$ IEEE DETERMINATION OF CAMERA LOCATION FROM 2D TO 3D LINE AND POINT CORRESPONDENCES DETERMINATION OF CAMERA LOCATION FROM 2D TO 3D LINE AND POINT CORRESPONDENCES Yuncai Liu Thomas S. Huang and 0. D. Faugeras Coordinated Science Laboratory University of Illinois at Urbana-Champaign 1101

More information

Real-time target tracking using a Pan and Tilt platform

Real-time target tracking using a Pan and Tilt platform Real-time target tracking using a Pan and Tilt platform Moulay A. Akhloufi Abstract In recent years, we see an increase of interest for efficient tracking systems in surveillance applications. Many of

More information

Fast and Reliable Two-View Translation Estimation

Fast and Reliable Two-View Translation Estimation Fast and Reliable Two-View Translation Estimation Johan Fredriksson 1 1 Centre for Mathematical Sciences Lund University, Sweden johanf@maths.lth.se Olof Enqvist 2 Fredrik Kahl 1,2 2 Department of Signals

More information

Mixture Models and EM

Mixture Models and EM Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering

More information

Fast Outlier Rejection by Using Parallax-Based Rigidity Constraint for Epipolar Geometry Estimation

Fast Outlier Rejection by Using Parallax-Based Rigidity Constraint for Epipolar Geometry Estimation Fast Outlier Rejection by Using Parallax-Based Rigidity Constraint for Epipolar Geometry Estimation Engin Tola 1 and A. Aydın Alatan 2 1 Computer Vision Laboratory, Ecóle Polytechnique Fédéral de Lausanne

More information

Stereo-Based Obstacle Avoidance in Indoor Environments with Active Sensor Re-Calibration

Stereo-Based Obstacle Avoidance in Indoor Environments with Active Sensor Re-Calibration Stereo-Based Obstacle Avoidance in Indoor Environments with Active Sensor Re-Calibration Darius Burschka, Stephen Lee and Gregory Hager Computational Interaction and Robotics Laboratory Johns Hopkins University

More information

CHAPTER 5 MOTION DETECTION AND ANALYSIS

CHAPTER 5 MOTION DETECTION AND ANALYSIS CHAPTER 5 MOTION DETECTION AND ANALYSIS 5.1. Introduction: Motion processing is gaining an intense attention from the researchers with the progress in motion studies and processing competence. A series

More information

An Improved Evolutionary Algorithm for Fundamental Matrix Estimation

An Improved Evolutionary Algorithm for Fundamental Matrix Estimation 03 0th IEEE International Conference on Advanced Video and Signal Based Surveillance An Improved Evolutionary Algorithm for Fundamental Matrix Estimation Yi Li, Senem Velipasalar and M. Cenk Gursoy Department

More information

3D Environment Reconstruction

3D Environment Reconstruction 3D Environment Reconstruction Using Modified Color ICP Algorithm by Fusion of a Camera and a 3D Laser Range Finder The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October 11-15,

More information

Local Image Registration: An Adaptive Filtering Framework

Local Image Registration: An Adaptive Filtering Framework Local Image Registration: An Adaptive Filtering Framework Gulcin Caner a,a.murattekalp a,b, Gaurav Sharma a and Wendi Heinzelman a a Electrical and Computer Engineering Dept.,University of Rochester, Rochester,

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

To appear in ECCV-94, Stockholm, Sweden, May 2-6, 1994.

To appear in ECCV-94, Stockholm, Sweden, May 2-6, 1994. To appear in ECCV-94, Stockholm, Sweden, May 2-6, 994. Recognizing Hand Gestures? James Davis and Mubarak Shah?? Computer Vision Laboratory, University of Central Florida, Orlando FL 3286, USA Abstract.

More information

arxiv: v1 [cs.cv] 18 Sep 2017

arxiv: v1 [cs.cv] 18 Sep 2017 Direct Pose Estimation with a Monocular Camera Darius Burschka and Elmar Mair arxiv:1709.05815v1 [cs.cv] 18 Sep 2017 Department of Informatics Technische Universität München, Germany {burschka elmar.mair}@mytum.de

More information

Binocular Stereo Vision. System 6 Introduction Is there a Wedge in this 3D scene?

Binocular Stereo Vision. System 6 Introduction Is there a Wedge in this 3D scene? System 6 Introduction Is there a Wedge in this 3D scene? Binocular Stereo Vision Data a stereo pair of images! Given two 2D images of an object, how can we reconstruct 3D awareness of it? AV: 3D recognition

More information

Passing Vehicle Detection from Dynamic Background Using Robust Information Fusion

Passing Vehicle Detection from Dynamic Background Using Robust Information Fusion Passing Vehicle Detection from Dynamic Background Using Robust Information Fusion Ying Zhu Dorin Comaniciu Martin Pellkofer Thorsten Koehler Siemens Corporate Research, Inc. Siemens VDO Automotive AG Princeton,

More information

Compositing a bird's eye view mosaic

Compositing a bird's eye view mosaic Compositing a bird's eye view mosaic Robert Laganiere School of Information Technology and Engineering University of Ottawa Ottawa, Ont KN 6N Abstract This paper describes a method that allows the composition

More information

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires

More information

An Event-based Optical Flow Algorithm for Dynamic Vision Sensors

An Event-based Optical Flow Algorithm for Dynamic Vision Sensors An Event-based Optical Flow Algorithm for Dynamic Vision Sensors Iffatur Ridwan and Howard Cheng Department of Mathematics and Computer Science University of Lethbridge, Canada iffatur.ridwan@uleth.ca,howard.cheng@uleth.ca

More information

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices Technical University of Cluj-Napoca Image Processing and Pattern Recognition Research Center www.cv.utcluj.ro Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving

More information

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level

More information

Estimating Camera Position And Posture by Using Feature Landmark Database

Estimating Camera Position And Posture by Using Feature Landmark Database Estimating Camera Position And Posture by Using Feature Landmark Database Motoko Oe 1, Tomokazu Sato 2 and Naokazu Yokoya 2 1 IBM Japan 2 Nara Institute of Science and Technology, Japan Abstract. Estimating

More information

Recognition. Clark F. Olson. Cornell University. work on separate feature sets can be performed in

Recognition. Clark F. Olson. Cornell University. work on separate feature sets can be performed in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 907-912, 1996. Connectionist Networks for Feature Indexing and Object Recognition Clark F. Olson Department of Computer

More information

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Zhiyan Zhang 1, Wei Qian 1, Lei Pan 1 & Yanjun Li 1 1 University of Shanghai for Science and Technology, China

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Can Lucas-Kanade be used to estimate motion parallax in 3D cluttered scenes?

Can Lucas-Kanade be used to estimate motion parallax in 3D cluttered scenes? Can Lucas-Kanade be used to estimate motion parallax in 3D cluttered scenes? V. Couture and M. S. Langer McGill University (School of Computer Science), Montreal, Quebec, H3A 2A7 Canada email: vincent.couture@mail.mcgill.ca

More information

Tracking of Human Body using Multiple Predictors

Tracking of Human Body using Multiple Predictors Tracking of Human Body using Multiple Predictors Rui M Jesus 1, Arnaldo J Abrantes 1, and Jorge S Marques 2 1 Instituto Superior de Engenharia de Lisboa, Postfach 351-218317001, Rua Conselheiro Emído Navarro,

More information

Robotics Programming Laboratory

Robotics Programming Laboratory Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car

More information

Time-to-Contact from Image Intensity

Time-to-Contact from Image Intensity Time-to-Contact from Image Intensity Yukitoshi Watanabe Fumihiko Sakaue Jun Sato Nagoya Institute of Technology Gokiso, Showa, Nagoya, 466-8555, Japan {yukitoshi@cv.,sakaue@,junsato@}nitech.ac.jp Abstract

More information

Target Tracking Using Mean-Shift And Affine Structure

Target Tracking Using Mean-Shift And Affine Structure Target Tracking Using Mean-Shift And Affine Structure Chuan Zhao, Andrew Knight and Ian Reid Department of Engineering Science, University of Oxford, Oxford, UK {zhao, ian}@robots.ox.ac.uk Abstract Inthispaper,wepresentanewapproachfortracking

More information

Vision par ordinateur

Vision par ordinateur Epipolar geometry π Vision par ordinateur Underlying structure in set of matches for rigid scenes l T 1 l 2 C1 m1 l1 e1 M L2 L1 e2 Géométrie épipolaire Fundamental matrix (x rank 2 matrix) m2 C2 l2 Frédéric

More information

An Automatic Method for Adjustment of a Camera Calibration Room

An Automatic Method for Adjustment of a Camera Calibration Room An Automatic Method for Adjustment of a Camera Calibration Room Presented at the FIG Working Week 2017, May 29 - June 2, 2017 in Helsinki, Finland Theory, algorithms, implementation, and two advanced applications.

More information

All human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them.

All human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them. All human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them. - Aristotle University of Texas at Arlington Introduction

More information

Motion in 2D image sequences

Motion in 2D image sequences Motion in 2D image sequences Definitely used in human vision Object detection and tracking Navigation and obstacle avoidance Analysis of actions or activities Segmentation and understanding of video sequences

More information

Autonomous Mobile Robot Design

Autonomous Mobile Robot Design Autonomous Mobile Robot Design Topic: EKF-based SLAM Dr. Kostas Alexis (CSE) These slides have partially relied on the course of C. Stachniss, Robot Mapping - WS 2013/14 Autonomous Robot Challenges Where

More information

Multiple Plane Segmentation Using Optical Flow

Multiple Plane Segmentation Using Optical Flow Multiple Plane Segmentation Using Optical Flow Marco Zucchelli José Santos-Victor Henrik I. Christensen CVAP & CAS, Royal Institute of Technology, Stockholm, Sweden S100 44 Vislab, Instituto Superior Técnico,

More information

Subpixel Corner Detection Using Spatial Moment 1)

Subpixel Corner Detection Using Spatial Moment 1) Vol.31, No.5 ACTA AUTOMATICA SINICA September, 25 Subpixel Corner Detection Using Spatial Moment 1) WANG She-Yang SONG Shen-Min QIANG Wen-Yi CHEN Xing-Lin (Department of Control Engineering, Harbin Institute

More information

Using temporal seeding to constrain the disparity search range in stereo matching

Using temporal seeding to constrain the disparity search range in stereo matching Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department

More information

A Stochastic Environment Modeling Method for Mobile Robot by using 2-D Laser scanner Young D. Kwon,Jin.S Lee Department of Electrical Engineering, Poh

A Stochastic Environment Modeling Method for Mobile Robot by using 2-D Laser scanner Young D. Kwon,Jin.S Lee Department of Electrical Engineering, Poh A Stochastic Environment Modeling Method for Mobile Robot by using -D Laser scanner Young D. Kwon,Jin.S Lee Department of Electrical Engineering, Pohang University of Science and Technology, e-mail: jsoo@vision.postech.ac.kr

More information

Background Initialization with A New Robust Statistical Approach

Background Initialization with A New Robust Statistical Approach Background Initialization with A New Robust Statistical Approach Hanzi Wang and David Suter Institute for Vision System Engineering Department of. Electrical. and Computer Systems Engineering Monash University,

More information

Real Time Biologically-Inspired Depth Maps from Spherical Flow

Real Time Biologically-Inspired Depth Maps from Spherical Flow Real Time Biologically-Inspired Depth Maps from Spherical Flow Chris McCarthy, Nick Barnes and Mandyam Srinivasan Abstract We present a strategy for generating real-time relative depth maps of an environment

More information

Robot localization method based on visual features and their geometric relationship

Robot localization method based on visual features and their geometric relationship , pp.46-50 http://dx.doi.org/10.14257/astl.2015.85.11 Robot localization method based on visual features and their geometric relationship Sangyun Lee 1, Changkyung Eem 2, and Hyunki Hong 3 1 Department

More information

of human activities. Our research is motivated by considerations of a ground-based mobile surveillance system that monitors an extended area for

of human activities. Our research is motivated by considerations of a ground-based mobile surveillance system that monitors an extended area for To Appear in ACCV-98, Mumbai-India, Material Subject to ACCV Copy-Rights Visual Surveillance of Human Activity Larry Davis 1 Sandor Fejes 1 David Harwood 1 Yaser Yacoob 1 Ismail Hariatoglu 1 Michael J.

More information

Absolute Scale Structure from Motion Using a Refractive Plate

Absolute Scale Structure from Motion Using a Refractive Plate Absolute Scale Structure from Motion Using a Refractive Plate Akira Shibata, Hiromitsu Fujii, Atsushi Yamashita and Hajime Asama Abstract Three-dimensional (3D) measurement methods are becoming more and

More information

VISION-BASED HANDLING WITH A MOBILE ROBOT

VISION-BASED HANDLING WITH A MOBILE ROBOT VISION-BASED HANDLING WITH A MOBILE ROBOT STEFAN BLESSING TU München, Institut für Werkzeugmaschinen und Betriebswissenschaften (iwb), 80290 München, Germany, e-mail: bl@iwb.mw.tu-muenchen.de STEFAN LANSER,

More information

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 11, November 2014,

More information

2 The MiníMax Principle First consider a simple problem. This problem will address the tradeos involved in a two-objective optimiation problem, where

2 The MiníMax Principle First consider a simple problem. This problem will address the tradeos involved in a two-objective optimiation problem, where Determining the Optimal Weights in Multiple Objective Function Optimiation Michael A. Gennert Alan L. Yuille Department of Computer Science Division of Applied Sciences Worcester Polytechnic Institute

More information

Arm coordinate system. View 1. View 1 View 2. View 2 R, T R, T R, T R, T. 12 t 1. u_ 1 u_ 2. Coordinate system of a robot

Arm coordinate system. View 1. View 1 View 2. View 2 R, T R, T R, T R, T. 12 t 1. u_ 1 u_ 2. Coordinate system of a robot Czech Technical University, Prague The Center for Machine Perception Camera Calibration and Euclidean Reconstruction from Known Translations Tomas Pajdla and Vaclav Hlavac Computer Vision Laboratory Czech

More information