Calibrating a Structured Light System Dr Alan M. McIvor Robert J. Valkenburg Machine Vision Team, Industrial Research Limited P.O. Box 2225, Auckland
|
|
- Loren Shelton
- 5 years ago
- Views:
Transcription
1 Calibrating a Structured Light System Dr Alan M. McIvor Robert J. Valkenburg Machine Vision Team, Industrial Research Limited P.O. Box 2225, Auckland New Zealand Tel: , Fax: a.mcivor@irl.cri.nz, b.valkenburg@irl.cri.nz Abstract Calibration of a structured light system is considered in this paper. The system consists of a projector whose output is controlled by a liquid crystal shutter, and a camera to capture the images. A sequence of patterns is projected to temporally encode 256 stripes on the visible surfaces. The transformation from 3D world coordinates to camera image plane coordinates and projector stripe ids are both modelled using perspective transformation matrices. The space of such matrices is characterised and a singular value decomposition procedure is described for estimating their elements. The accuracy of the calibration system so attained is investigated and improvements to the method are suggested. 1. Introduction The structured light system consists of a liquid crystal shutter controlling the output of a light source, and a CCD video camera to capture image of the scene. The liquid crystal shutter generates 8 dierent band images. The combination of these produces a temporal encoding of 256 dierent stripes. The raw output is thus an 8-bit stripe id for each pixel location in the image plane. Given the ray (in the world) that projects onto a particular image pixel, and the plane (in the world) projected by a stripe, the 3D coordinates of their intersection can be calculated. By this triangulation method, the raw sensor data is converted into 3D data. Calibration is the estimation of the parameters of a model that denes these rays and planes. The calibration method used is based on pin-hole optical models of the camera and projector. The transformation from 3D coordinates to camera image plane coordinates is commonly described by a 3 4 perspective transformation matrix (PTM). One of the contributions of this paper is to show that the projector can be similarly modelled by a 2 4 PTM describing the transformation from 3D coordinates to stripe ids. Another contribution is two theorems which describe the subset of 3 4 and 2 4 matrices which are PTMs.
2 2. System Model This section describes models of the camera and projector based on perspective transformation matrices. The physical interpretation of these models is also discussed. 2.1 Camera Model Given a homogeneous coordinate system, X w, in which world 3D positions are specied, and a homogeneous coordinate system, x c, in which pixel locations on the camera sensor are specied, the transformation from world coordinates to image coordinates (i.e., the \imaging" process) can be modelled by a 3 4 PTM [1]: The matrix C c has the following form: x c = C c X w (1) Denition 1 A 3 4 perspective transformation matrix is a matrix of the form C = f x kf y x o c 0 f y y o c h i 7 5 R t where R is a 3 3 rotation matrix, and, f x, f y are non-zero. (2) Physically, R and t dene the coordinate transformation between the world coordinate frame and a camera centric coordinate frame. The parameters f x and f y are the combination of camera focal length and pixel grid scaling factors, and x o c and y o c give the optic centre in image plane coordinates. The parameter k can be interpreted as a shear of the image plane coordinate system. The following theorem is useful in estimating a 3 4 PTM from observed data. Theorem 1 A 3 4 matrix is a 3 4 PTM i the submatrix formed by the rst three columns has rank 3. The condition being necessary follows because, from (2), the submatrix is the product of two matrices with non-zero determinant. Suciency follows from the observation that these two matrices form a QR decomposition of the 3 3 submatrix [2]. Further details of the proof are given in [3]. Given a PTM, there are 4 possible solutions for the decomposition in (2). 2.2 Projector Model The stripe ids can be thought of as a 1D coordinate frame on the projector plane. Then, the transformation from world coordinates to stripe ids (i.e., the inverse of the projection process) can be modelled by a 2 4 PTM between homogeneous world coordinates X w and homogeneous stripe ids x p : x p = C p X w (3) The matrix C p has the following form: Denition 2 A 2 4 perspective transformation matrix is a matrix of the form " # f C = p 0 x o h i p Q s where Q is a 3 3 rotation matrix, and and f p are non-zero. (4)
3 Figure 1: An image of the calibration target showing the fuducial marks and the outer band used for initial position estimation. Physically, Q and s dene the coordinate transformation between the world coordinate system and a projector centric one. The parameter f p is the projector focal length and x o p is the x-coordinate of the intersection of the optic axis and the liquid crystal shutter. Theorem 2 A 2 4 matrix is a 2 4 PTM i the submatrix consisting of the rst three columns is of rank 2. The proof of this theorem follows similarly to the camera case [3]. However, in this case, one parameter of the decomposition (s2) is indeterminant. 3. Implementation Issues Calibration involves estimating the PTMs of the camera and projector from a number of known world points and their corresponding image points and stripe ids. Physically, the world points are implemented as 72 circular fuducial marks (spots), 24 on each of three faces of a cube (Figure 1). The image sequence is processed to obtain the image location and stripe id of each spot center. 3.1 Estimating the Perspective Transformation Matrices We will only consider the task of estimating the camera PTM as the procedure for estimating the projector PTM is identical. Let (X i w ; xi c); i = 1 : : : n be a set of known world-image point pairs. We wish to nd an estimate of the PTM which is consistent with these data points in some optimal sense. Initially we assume that the data points come from a perfect camera which is noiseless and can be modelled exactly by a PTM.
4 Let x c = then where h x c y c 1 B = i T be an image point and ci be the i th row of a 3 4 matrix C, X w 0 0 X w?x c X w?y c X w x c = CX w, Bl = 0 (5) 3 T 7 5 h i T and l = c1 c 2 c 3 (6) If we associate the design matrix B i with the i th world-image point pair and dene A = h i T B T 1 : : : B T n we arrive at the following homogeneous equation. Al = 0 (7) It can be shown that in general rank(a) 11 and that, for n 6 suitably selected world points (non-coplanar and not lying on a twisted cubic passing through the origin of the camera coordinate system), rank(a) = 11 [1]. Assume that X i w ; i = 1 : : : n satisfy these conditions so nullity(a) = 1. Let ^l be any non-zero element of NS(A) and ^C be the associated 3 4 matrix. Let (X w ; x c ) be an arbitrary world-image point pair, B be the associated design matrix and h i T dene A = A T B T. It follows that B^l = 0 as nullity( A) = 1 and NS( A) NS(A). Equation (5) implies that x c = ^CX w and hence ^C is the required PTM. We know that the PTM, C, is only dened up to a scale factor. Hence, we may arbitrarily impose the constraint kck F = 1 (, klk 2 = 1 ) in order to make the problem of nding the PTM (nearly) well dened. Due to deciencies in the camera model, noise on the image coordinates, and noise (imprecision in measuring) on the world coordinates, the matrix A will in general be of full rank. Let S m denote the space generated by the singular vector associated with the smallest singular value of A. S m provides a generalisation of NS(A) in following the sense If l 2 S m then kalk kaqk for all q such that klk = kqk (8) In particular, if A has a non-trivial null-space then S m = NS(A). As previously mentioned the PTM, and hence l, is only required to a scale factor. However kalk depends on klk (i.e. l = kl o ) kalk = jkj kal o k) so it only makes sense to compare solution vectors of xed norm. Equation (8) shows that given a specied norm a vector in S m provides an optimal solution. The following equivalent formulation allows comparison with alternative techniques more readily. If we impose the constraint that klk = 1 then l 2 S m solves min l kalk such that klk = 1 (9) In the above we have ignored the structure of a PTM and just found a matrix which is consistent with the data. We actually want to nd the solution of min kalk such that klk = 1; det l 2 l1 l2 l3 6 4 l5 l6 l7 l9 l10 l = 0 (10) The condition for a matrix to be a PTM stated in Theorem 1 is weak in that the possibility of violating the constraint is very improbable (because the subset of matrices which satisfy
5 the constraint forms a sparse surface in R 12 ). Hence in practice we ignore the constraint, nd a matrix by solving (9) and then check that the matrix satises the determinant constraint and is therefore a PTM. If the solution of (9) satises the constraint it is obvious that it is also the solution of (10). 3.2 Localisation of marks Three types of information are required to perform calibration. It is necessary to know the world location, image location and stripe id of the fuducial spot centers. It is assumed that the cube has been measured and world location of the spots are known to sucient accuracy. The image location and stripe id of all spots are extracted from images taken of the cube. Currently we only use subpixel techniques to extract the image location of the spots and the centroid algorithm is used for this purpose [4]. The task of extracting the subpixel stripe id is more dicult and will be implemented in future versions. 4. Accuracy In [3], details are given of how the PTMs can be decomposed into the intrinsic and extrinsic parameters which describe the camera and projector transformations. The intrinsic parameters obtained in this way for the structured light system agree closely with those derived from the data sheets supplied by the component manufacturers. The PTM equations for the camera and the projector can be combined into TX w = " xc 0 # (11) where T = " c p 11? x pc p 21 c p 12? x pc p 22 C c c p 13? x pc p 23 c p 14? x pc p 24 # (12) and x p is the observed stripe id. Hence, the 3D position of the surface point can be determined by inverting T. Tests of this triangulation procedure show that the 3D data so generated from at surfaces has an RMS error of the order of 1mm and maximum errors of 4.6mm. This is for a variety of surface orientations within the 140mm140mm140mm working volume of the system. See [3] for details. 5. Discussion This paper has considered the calibration of a structured light system. The transformation from 3D to camera image plane coordinates is characterised by a 3 4 PTM and the transformation from 3D coordinates to projector stripe ids is characterised by a 24 PTM. After giving a denition of a PTM, the subset of matrices that are PTMs is determined. The components of the PTMs were determined using a singular value decomposition. It is noted that the subset of (suitably sized) matrices which are not PTMs is small. Therefore, the solution strategy is to solve over the space of all matrices, and check that the solution is indeed a PTM. One of the sources of errors in determining the PTMs is that the minimisation criteria used to determine the elements of the PTM does not have a good physical interpretation.
6 The selection of the PTM can be recast as a nonlinear least squares problem, where the observed image coordinates and stripe ids are compared with those generated by projecting the known 3D world coordinates, given the current PTM estimates. The objective would be to minimise the discrepancy between these. The calibration method described herein would be used as an initial estimate for the nonlinear optimisation scheme. This type of approach is currently the subject of on-going research. A major source of error is the fact that the linear pin-hole optics model does not accurately model an actual physical lens system. The latter suer from radial distortion, etc. Bundle Adjustment techniques exist for incorporating such factors into the optical models [5]. Applying these to structured light system calibration is not straight forward because the distortion model depends on components that are not directly observable for the projector 1. Accurate subpixel stripe estimates are also required. These areas will be considered in future work. References 1 Olivier Faugeras. Three-Dimensional Computer Vision. The MIT Press, Martin Armstrong, Andrew Zisserman, and Paul Beardsley. Euclidean structure from uncalibrated images. In Edwin Hancock, editor, BMVC94: Proceedings of the 5th British Machine Vision Conference, pages 509{518. BMVA Press, September Alan M. McIvor and Robert J. Valkenburg. Calibrating a Structured Light System. Report 362, Industrial Research Limited, February Robert J. Valkenburg, Alan M. McIvor, and P. Wayne Power. An evaluation of subpixel feature localisation methods for precision measurement. In International Symposium on Photonic Sensors and Control for Commercial Applications: Videometrics III. SPIE, November H. M. Karara, editor. Non-Topograhic Photogrammetry. American Society for Photogrammetry and Remote Sensing, 2 edition, The image plane coordinate parallel to the direction of the stripes is not measurable.
Vision Review: Image Formation. Course web page:
Vision Review: Image Formation Course web page: www.cis.udel.edu/~cer/arv September 10, 2002 Announcements Lecture on Thursday will be about Matlab; next Tuesday will be Image Processing The dates some
More information55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography Estimating homography from point correspondence
More informationHomogeneous Coordinates. Lecture18: Camera Models. Representation of Line and Point in 2D. Cross Product. Overall scaling is NOT important.
Homogeneous Coordinates Overall scaling is NOT important. CSED44:Introduction to Computer Vision (207F) Lecture8: Camera Models Bohyung Han CSE, POSTECH bhhan@postech.ac.kr (",, ) ()", ), )) ) 0 It is
More informationmeasurements using a structured light system (SLS). Structured light is a general concept and there are a number of ways of exploiting it to obtain 3D
Accepted for publication Image and Vision Computing 1997 1 Accurate 3D measurement using a Structured Light System R.J. Valkenburg A.M. M c Ivor Industrial Research Limited P.O. Box 2225 Auckland New Zealand
More informationIntroduction to Homogeneous coordinates
Last class we considered smooth translations and rotations of the camera coordinate system and the resulting motions of points in the image projection plane. These two transformations were expressed mathematically
More informationGeometric camera models and calibration
Geometric camera models and calibration http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 13 Course announcements Homework 3 is out. - Due October
More informationPin Hole Cameras & Warp Functions
Pin Hole Cameras & Warp Functions Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Pinhole Camera. Homogenous Coordinates. Planar Warp Functions. Motivation Taken from: http://img.gawkerassets.com/img/18w7i1umpzoa9jpg/original.jpg
More informationTwo-view geometry Computer Vision Spring 2018, Lecture 10
Two-view geometry http://www.cs.cmu.edu/~16385/ 16-385 Computer Vision Spring 2018, Lecture 10 Course announcements Homework 2 is due on February 23 rd. - Any questions about the homework? - How many of
More informationMultiple Views Geometry
Multiple Views Geometry Subhashis Banerjee Dept. Computer Science and Engineering IIT Delhi email: suban@cse.iitd.ac.in January 2, 28 Epipolar geometry Fundamental geometric relationship between two perspective
More informationCSE 252B: Computer Vision II
CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribe: Sameer Agarwal LECTURE 1 Image Formation 1.1. The geometry of image formation We begin by considering the process of image formation when a
More informationCamera Model and Calibration
Camera Model and Calibration Lecture-10 Camera Calibration Determine extrinsic and intrinsic parameters of camera Extrinsic 3D location and orientation of camera Intrinsic Focal length The size of the
More informationZ (cm) Y (cm) X (cm)
Oceans'98 IEEE/OES Conference Uncalibrated Vision for 3-D Underwater Applications K. Plakas, E. Trucco Computer Vision Group and Ocean Systems Laboratory Dept. of Computing and Electrical Engineering Heriot-Watt
More informationEpipolar geometry. x x
Two-view geometry Epipolar geometry X x x Baseline line connecting the two camera centers Epipolar Plane plane containing baseline (1D family) Epipoles = intersections of baseline with image planes = projections
More informationCS231A Course Notes 4: Stereo Systems and Structure from Motion
CS231A Course Notes 4: Stereo Systems and Structure from Motion Kenji Hata and Silvio Savarese 1 Introduction In the previous notes, we covered how adding additional viewpoints of a scene can greatly enhance
More information3D Geometry and Camera Calibration
3D Geometry and Camera Calibration 3D Coordinate Systems Right-handed vs. left-handed x x y z z y 2D Coordinate Systems 3D Geometry Basics y axis up vs. y axis down Origin at center vs. corner Will often
More informationRecovering structure from a single view Pinhole perspective projection
EPIPOLAR GEOMETRY The slides are from several sources through James Hays (Brown); Silvio Savarese (U. of Michigan); Svetlana Lazebnik (U. Illinois); Bill Freeman and Antonio Torralba (MIT), including their
More informationPin Hole Cameras & Warp Functions
Pin Hole Cameras & Warp Functions Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Pinhole Camera. Homogenous Coordinates. Planar Warp Functions. Example of SLAM for AR Taken from:
More informationComputer Vision Project-1
University of Utah, School Of Computing Computer Vision Project- Singla, Sumedha sumedha.singla@utah.edu (00877456 February, 205 Theoretical Problems. Pinhole Camera (a A straight line in the world space
More informationA General Expression of the Fundamental Matrix for Both Perspective and Affine Cameras
A General Expression of the Fundamental Matrix for Both Perspective and Affine Cameras Zhengyou Zhang* ATR Human Information Processing Res. Lab. 2-2 Hikari-dai, Seika-cho, Soraku-gun Kyoto 619-02 Japan
More information55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography Estimating homography from point correspondence
More informationCamera Calibration. COS 429 Princeton University
Camera Calibration COS 429 Princeton University Point Correspondences What can you figure out from point correspondences? Noah Snavely Point Correspondences X 1 X 4 X 3 X 2 X 5 X 6 X 7 p 1,1 p 1,2 p 1,3
More information3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,
3D Sensing and Reconstruction Readings: Ch 12: 12.5-6, Ch 13: 13.1-3, 13.9.4 Perspective Geometry Camera Model Stereo Triangulation 3D Reconstruction by Space Carving 3D Shape from X means getting 3D coordinates
More information3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.
3D Sensing 3D Shape from X Perspective Geometry Camera Model Camera Calibration General Stereo Triangulation 3D Reconstruction 3D Shape from X shading silhouette texture stereo light striping motion mainly
More informationCamera Calibration Using Line Correspondences
Camera Calibration Using Line Correspondences Richard I. Hartley G.E. CRD, Schenectady, NY, 12301. Ph: (518)-387-7333 Fax: (518)-387-6845 Email : hartley@crd.ge.com Abstract In this paper, a method of
More informationComputer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.
Computer Vision Coordinates Prof. Flávio Cardeal DECOM / CEFET- MG cardeal@decom.cefetmg.br Abstract This lecture discusses world coordinates and homogeneous coordinates, as well as provides an overview
More informationDD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication
DD2423 Image Analysis and Computer Vision IMAGE FORMATION Mårten Björkman Computational Vision and Active Perception School of Computer Science and Communication November 8, 2013 1 Image formation Goal:
More informationStereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz
Stereo II CSE 576 Ali Farhadi Several slides from Larry Zitnick and Steve Seitz Camera parameters A camera is described by several parameters Translation T of the optical center from the origin of world
More informationA thesis submitted in partial fulllment of. the requirements for the degree of. Bachelor of Technology. Computer Science and Engineering
R N O C A thesis submitted in partial fulllment of the requirements for the degree of Bachelor of Technology in Computer Science and Engineering Rahul Bhotika Anurag Mittal Supervisor : Dr Subhashis Banerjee
More informationCOS429: COMPUTER VISON CAMERAS AND PROJECTIONS (2 lectures)
COS429: COMPUTER VISON CMERS ND PROJECTIONS (2 lectures) Pinhole cameras Camera with lenses Sensing nalytical Euclidean geometry The intrinsic parameters of a camera The extrinsic parameters of a camera
More informationVisual Recognition: Image Formation
Visual Recognition: Image Formation Raquel Urtasun TTI Chicago Jan 5, 2012 Raquel Urtasun (TTI-C) Visual Recognition Jan 5, 2012 1 / 61 Today s lecture... Fundamentals of image formation You should know
More informationFlexible Calibration of a Portable Structured Light System through Surface Plane
Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured
More informationAugmented Reality II - Camera Calibration - Gudrun Klinker May 11, 2004
Augmented Reality II - Camera Calibration - Gudrun Klinker May, 24 Literature Richard Hartley and Andrew Zisserman, Multiple View Geometry in Computer Vision, Cambridge University Press, 2. (Section 5,
More informationRectification and Distortion Correction
Rectification and Distortion Correction Hagen Spies March 12, 2003 Computer Vision Laboratory Department of Electrical Engineering Linköping University, Sweden Contents Distortion Correction Rectification
More information1 Projective Geometry
CIS8, Machine Perception Review Problem - SPRING 26 Instructions. All coordinate systems are right handed. Projective Geometry Figure : Facade rectification. I took an image of a rectangular object, and
More informationMultiple View Reconstruction of Calibrated Images using Singular Value Decomposition
Multiple View Reconstruction of Calibrated Images using Singular Value Decomposition Ayan Chaudhury, Abhishek Gupta, Sumita Manna, Subhadeep Mukherjee, Amlan Chakrabarti Abstract Calibration in a multi
More informationA COMPREHENSIVE SIMULATION SOFTWARE FOR TEACHING CAMERA CALIBRATION
XIX IMEKO World Congress Fundamental and Applied Metrology September 6 11, 2009, Lisbon, Portugal A COMPREHENSIVE SIMULATION SOFTWARE FOR TEACHING CAMERA CALIBRATION David Samper 1, Jorge Santolaria 1,
More informationcalibrated coordinates Linear transformation pixel coordinates
1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial
More informationMultiple View Geometry in Computer Vision
Multiple View Geometry in Computer Vision Prasanna Sahoo Department of Mathematics University of Louisville 1 Structure Computation Lecture 18 March 22, 2005 2 3D Reconstruction The goal of 3D reconstruction
More informationCamera calibration. Robotic vision. Ville Kyrki
Camera calibration Robotic vision 19.1.2017 Where are we? Images, imaging Image enhancement Feature extraction and matching Image-based tracking Camera models and calibration Pose estimation Motion analysis
More informationCS201 Computer Vision Camera Geometry
CS201 Computer Vision Camera Geometry John Magee 25 November, 2014 Slides Courtesy of: Diane H. Theriault (deht@bu.edu) Question of the Day: How can we represent the relationships between cameras and the
More informationComputer Vision Projective Geometry and Calibration. Pinhole cameras
Computer Vision Projective Geometry and Calibration Professor Hager http://www.cs.jhu.edu/~hager Jason Corso http://www.cs.jhu.edu/~jcorso. Pinhole cameras Abstract camera model - box with a small hole
More informationLecture 3: Camera Calibration, DLT, SVD
Computer Vision Lecture 3 23--28 Lecture 3: Camera Calibration, DL, SVD he Inner Parameters In this section we will introduce the inner parameters of the cameras Recall from the camera equations λx = P
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 7: Image Alignment and Panoramas What s inside your fridge? http://www.cs.washington.edu/education/courses/cse590ss/01wi/ Projection matrix intrinsics projection
More informationCamera Model and Calibration. Lecture-12
Camera Model and Calibration Lecture-12 Camera Calibration Determine extrinsic and intrinsic parameters of camera Extrinsic 3D location and orientation of camera Intrinsic Focal length The size of the
More informationMachine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy
1 Machine vision Summary # 11: Stereo vision and epipolar geometry STEREO VISION The goal of stereo vision is to use two cameras to capture 3D scenes. There are two important problems in stereo vision:
More informationArm coordinate system. View 1. View 1 View 2. View 2 R, T R, T R, T R, T. 12 t 1. u_ 1 u_ 2. Coordinate system of a robot
Czech Technical University, Prague The Center for Machine Perception Camera Calibration and Euclidean Reconstruction from Known Translations Tomas Pajdla and Vaclav Hlavac Computer Vision Laboratory Czech
More informationRigid Body Motion and Image Formation. Jana Kosecka, CS 482
Rigid Body Motion and Image Formation Jana Kosecka, CS 482 A free vector is defined by a pair of points : Coordinates of the vector : 1 3D Rotation of Points Euler angles Rotation Matrices in 3D 3 by 3
More informationMAPI Computer Vision. Multiple View Geometry
MAPI Computer Vision Multiple View Geometry Geometry o Multiple Views 2- and 3- view geometry p p Kpˆ [ K R t]p Geometry o Multiple Views 2- and 3- view geometry Epipolar Geometry The epipolar geometry
More informationCOMP 558 lecture 19 Nov. 17, 2010
COMP 558 lecture 9 Nov. 7, 2 Camera calibration To estimate the geometry of 3D scenes, it helps to know the camera parameters, both external and internal. The problem of finding all these parameters is
More informationMERGING POINT CLOUDS FROM MULTIPLE KINECTS. Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia
MERGING POINT CLOUDS FROM MULTIPLE KINECTS Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia Introduction What do we want to do? : Use information (point clouds) from multiple (2+) Kinects
More informationStructure from Motion and Multi- view Geometry. Last lecture
Structure from Motion and Multi- view Geometry Topics in Image-Based Modeling and Rendering CSE291 J00 Lecture 5 Last lecture S. J. Gortler, R. Grzeszczuk, R. Szeliski,M. F. Cohen The Lumigraph, SIGGRAPH,
More informationThe Geometry Behind the Numerical Reconstruction of Two Photos
The Geometry Behind the Numerical Reconstruction of Two Photos Hellmuth Stachel stachel@dmg.tuwien.ac.at http://www.geometrie.tuwien.ac.at/stachel ICEGD 2007, The 2 nd Internat. Conf. on Eng g Graphics
More informationComputer Vision: Lecture 3
Computer Vision: Lecture 3 Carl Olsson 2019-01-29 Carl Olsson Computer Vision: Lecture 3 2019-01-29 1 / 28 Todays Lecture Camera Calibration The inner parameters - K. Projective vs. Euclidean Reconstruction.
More informationECE Digital Image Processing and Introduction to Computer Vision. Outline
ECE592-064 Digital Image Processing and Introduction to Computer Vision Depart. of ECE, NC State University Instructor: Tianfu (Matt) Wu Spring 2017 1. Recap Outline 2. Modeling Projection and Projection
More informationA Method for Interactive 3D Reconstruction of Piecewise Planar Objects from Single Images
A Method for Interactive 3D Reconstruction of Piecewise Planar Objects from Single Images Peter F Sturm and Stephen J Maybank Computational Vision Group, Department of Computer Science The University of
More informationRobust Geometry Estimation from two Images
Robust Geometry Estimation from two Images Carsten Rother 09/12/2016 Computer Vision I: Image Formation Process Roadmap for next four lectures Computer Vision I: Image Formation Process 09/12/2016 2 Appearance-based
More informationStereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz
Stereo CSE 576 Ali Farhadi Several slides from Larry Zitnick and Steve Seitz Why do we perceive depth? What do humans use as depth cues? Motion Convergence When watching an object close to us, our eyes
More informationEpipolar Geometry and the Essential Matrix
Epipolar Geometry and the Essential Matrix Carlo Tomasi The epipolar geometry of a pair of cameras expresses the fundamental relationship between any two corresponding points in the two image planes, and
More informationModule 4F12: Computer Vision and Robotics Solutions to Examples Paper 2
Engineering Tripos Part IIB FOURTH YEAR Module 4F2: Computer Vision and Robotics Solutions to Examples Paper 2. Perspective projection and vanishing points (a) Consider a line in 3D space, defined in camera-centered
More informationStereo Vision. MAN-522 Computer Vision
Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 5: Projection Reading: Szeliski 2.1 Projection Reading: Szeliski 2.1 Projection Müller Lyer Illusion http://www.michaelbach.de/ot/sze_muelue/index.html Modeling
More informationCOSC579: Scene Geometry. Jeremy Bolton, PhD Assistant Teaching Professor
COSC579: Scene Geometry Jeremy Bolton, PhD Assistant Teaching Professor Overview Linear Algebra Review Homogeneous vs non-homogeneous representations Projections and Transformations Scene Geometry The
More informationCamera models and calibration
Camera models and calibration Read tutorial chapter 2 and 3. http://www.cs.unc.edu/~marc/tutorial/ Szeliski s book pp.29-73 Schedule (tentative) 2 # date topic Sep.8 Introduction and geometry 2 Sep.25
More informationA Method for Interactive 3D Reconstruction of Piecewise Planar Objects from Single Images
A Method for Interactive 3D Reconstruction of Piecewise Planar Objects from Single Images Peter Sturm Steve Maybank To cite this version: Peter Sturm Steve Maybank A Method for Interactive 3D Reconstruction
More informationViewing. Reading: Angel Ch.5
Viewing Reading: Angel Ch.5 What is Viewing? Viewing transform projects the 3D model to a 2D image plane 3D Objects (world frame) Model-view (camera frame) View transform (projection frame) 2D image View
More informationCameras and Stereo CSE 455. Linda Shapiro
Cameras and Stereo CSE 455 Linda Shapiro 1 Müller-Lyer Illusion http://www.michaelbach.de/ot/sze_muelue/index.html What do you know about perspective projection? Vertical lines? Other lines? 2 Image formation
More informationToday. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography
Computational Photography Matthias Zwicker University of Bern Fall 2009 Today From 2D to 3D using multiple views Introduction Geometry of two views Stereo matching Other applications Multiview geometry
More informationCamera model and multiple view geometry
Chapter Camera model and multiple view geometry Before discussing how D information can be obtained from images it is important to know how images are formed First the camera model is introduced and then
More informationChapter 3. Quadric hypersurfaces. 3.1 Quadric hypersurfaces Denition.
Chapter 3 Quadric hypersurfaces 3.1 Quadric hypersurfaces. 3.1.1 Denition. Denition 1. In an n-dimensional ane space A; given an ane frame fo;! e i g: A quadric hypersurface in A is a set S consisting
More informationImage Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania
Image Formation Antonino Furnari Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania furnari@dmi.unict.it 18/03/2014 Outline Introduction; Geometric Primitives
More informationStructure from motion
Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t 2 R 3,t 3 Camera 1 Camera
More informationStructure from Motion
Structure from Motion Outline Bundle Adjustment Ambguities in Reconstruction Affine Factorization Extensions Structure from motion Recover both 3D scene geoemetry and camera positions SLAM: Simultaneous
More informationMath 308 Autumn 2016 MIDTERM /18/2016
Name: Math 38 Autumn 26 MIDTERM - 2 /8/26 Instructions: The exam is 9 pages long, including this title page. The number of points each problem is worth is listed after the problem number. The exam totals
More informationIndex. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263
Index 3D reconstruction, 125 5+1-point algorithm, 284 5-point algorithm, 270 7-point algorithm, 265 8-point algorithm, 263 affine point, 45 affine transformation, 57 affine transformation group, 57 affine
More informationRobotics - Projective Geometry and Camera model. Marcello Restelli
Robotics - Projective Geometr and Camera model Marcello Restelli marcello.restelli@polimi.it Dipartimento di Elettronica, Informazione e Bioingegneria Politecnico di Milano Ma 2013 Inspired from Matteo
More informationReminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches
Reminder: Lecture 20: The Eight-Point Algorithm F = -0.00310695-0.0025646 2.96584-0.028094-0.00771621 56.3813 13.1905-29.2007-9999.79 Readings T&V 7.3 and 7.4 Essential/Fundamental Matrix E/F Matrix Summary
More informationSelf-calibration of Multiple Laser Planes for 3D Scene Reconstruction
Self-calibration of Multiple Laser Planes for 3D Scene Reconstruction Ryo Furukawa Faculty of Information Sciences, Hiroshima City University, Japan ryo-f@cs.hiroshima-cu.ac.jp Hiroshi Kawasaki Faculty
More informationRaycasting. Chapter Raycasting foundations. When you look at an object, like the ball in the picture to the left, what do
Chapter 4 Raycasting 4. Raycasting foundations When you look at an, like the ball in the picture to the left, what do lamp you see? You do not actually see the ball itself. Instead, what you see is the
More informationComputer Vision. Geometric Camera Calibration. Samer M Abdallah, PhD
Computer Vision Samer M Abdallah, PhD Faculty of Engineering and Architecture American University of Beirut Beirut, Lebanon Geometric Camera Calibration September 2, 2004 1 Computer Vision Geometric Camera
More informationStructure from motion
Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t R 2 3,t 3 Camera 1 Camera
More informationTwo-View Geometry (Course 23, Lecture D)
Two-View Geometry (Course 23, Lecture D) Jana Kosecka Department of Computer Science George Mason University http://www.cs.gmu.edu/~kosecka General Formulation Given two views of the scene recover the
More informationIndex. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253
Index 3D reconstruction, 123 5+1-point algorithm, 274 5-point algorithm, 260 7-point algorithm, 255 8-point algorithm, 253 affine point, 43 affine transformation, 55 affine transformation group, 55 affine
More informationThe end of affine cameras
The end of affine cameras Affine SFM revisited Epipolar geometry Two-view structure from motion Multi-view structure from motion Planches : http://www.di.ens.fr/~ponce/geomvis/lect3.pptx http://www.di.ens.fr/~ponce/geomvis/lect3.pdf
More informationUnit 3 Multiple View Geometry
Unit 3 Multiple View Geometry Relations between images of a scene Recovering the cameras Recovering the scene structure http://www.robots.ox.ac.uk/~vgg/hzbook/hzbook1.html 3D structure from images Recover
More informationCamera Models and Image Formation. Srikumar Ramalingam School of Computing University of Utah
Camera Models and Image Formation Srikumar Ramalingam School of Computing University of Utah srikumar@cs.utah.edu Reference Most slides are adapted from the following notes: Some lecture notes on geometric
More informationCamera Models and Image Formation. Srikumar Ramalingam School of Computing University of Utah
Camera Models and Image Formation Srikumar Ramalingam School of Computing University of Utah srikumar@cs.utah.edu VisualFunHouse.com 3D Street Art Image courtesy: Julian Beaver (VisualFunHouse.com) 3D
More informationCoplanar circles, quasi-affine invariance and calibration
Image and Vision Computing 24 (2006) 319 326 www.elsevier.com/locate/imavis Coplanar circles, quasi-affine invariance and calibration Yihong Wu *, Xinju Li, Fuchao Wu, Zhanyi Hu National Laboratory of
More informationCOMPARISON OF CAMERA CALIBRATION PARAMETERS USING PHOTOMODELER AND AUSTRALIS
COMPARISON OF CAMERA CALIBRATION PARAMETERS USING PHOTOMODELER AND AUSTRALIS Fazli Abd. Rahman*, Halim Setan*, Albert K.Chong**, Zulkepli Majid* & Anuar Ahmad* *Department of Geomatic Engineering, Faculty
More informationInterlude: Solving systems of Equations
Interlude: Solving systems of Equations Solving Ax = b What happens to x under Ax? The singular value decomposition Rotation matrices Singular matrices Condition number Null space Solving Ax = 0 under
More informationAnnouncements. Stereo
Announcements Stereo Homework 2 is due today, 11:59 PM Homework 3 will be assigned today Reading: Chapter 7: Stereopsis CSE 152 Lecture 8 Binocular Stereopsis: Mars Given two images of a scene where relative
More informationDescriptive Geometry Meets Computer Vision The Geometry of Two Images (# 82)
Descriptive Geometry Meets Computer Vision The Geometry of Two Images (# 8) Hellmuth Stachel stachel@dmg.tuwien.ac.at http://www.geometrie.tuwien.ac.at/stachel th International Conference on Geometry and
More informationComputer Vision I Name : CSE 252A, Fall 2012 Student ID : David Kriegman Assignment #1. (Due date: 10/23/2012) x P. = z
Computer Vision I Name : CSE 252A, Fall 202 Student ID : David Kriegman E-Mail : Assignment (Due date: 0/23/202). Perspective Projection [2pts] Consider a perspective projection where a point = z y x P
More informationCamera Calibration from the Quasi-affine Invariance of Two Parallel Circles
Camera Calibration from the Quasi-affine Invariance of Two Parallel Circles Yihong Wu, Haijiang Zhu, Zhanyi Hu, and Fuchao Wu National Laboratory of Pattern Recognition, Institute of Automation, Chinese
More informationEuclidean Reconstruction from Constant Intrinsic Parameters
uclidean Reconstruction from Constant ntrinsic Parameters nders Heyden, Kalle Åström Dept of Mathematics, Lund University Box 118, S-221 00 Lund, Sweden email: heyden@maths.lth.se, kalle@maths.lth.se bstract
More informationImage Formation I Chapter 2 (R. Szelisky)
Image Formation I Chapter 2 (R. Selisky) Guido Gerig CS 632 Spring 22 cknowledgements: Slides used from Prof. Trevor Darrell, (http://www.eecs.berkeley.edu/~trevor/cs28.html) Some slides modified from
More information3D Model Acquisition by Tracking 2D Wireframes
3D Model Acquisition by Tracking 2D Wireframes M. Brown, T. Drummond and R. Cipolla {96mab twd20 cipolla}@eng.cam.ac.uk Department of Engineering University of Cambridge Cambridge CB2 1PZ, UK Abstract
More informationEXPERIMENTAL RESULTS ON THE DETERMINATION OF THE TRIFOCAL TENSOR USING NEARLY COPLANAR POINT CORRESPONDENCES
EXPERIMENTAL RESULTS ON THE DETERMINATION OF THE TRIFOCAL TENSOR USING NEARLY COPLANAR POINT CORRESPONDENCES Camillo RESSL Institute of Photogrammetry and Remote Sensing University of Technology, Vienna,
More informationThe real voyage of discovery consists not in seeking new landscapes, but in having new eyes.
The real voyage of discovery consists not in seeking new landscapes, but in having new eyes. - Marcel Proust University of Texas at Arlington Camera Calibration (or Resectioning) CSE 4392-5369 Vision-based
More informationLecture No Perspective Transformation (course: Computer Vision)
Lecture No. 2-6 Perspective Transformation (course: Computer Vision) e-mail: naeemmahoto@gmail.com Department of Software Engineering, Mehran UET Jamshoro, Sind, Pakistan 3-D Imaging Transformation A 3D
More informationOutline. ETN-FPI Training School on Plenoptic Sensing
Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration
More information