Perspective Projection Describes Image Formation Berthold K.P. Horn

Similar documents
Camera Model and Calibration

Massachusetts Institute of Technology Department of Computer Science and Electrical Engineering 6.801/6.866 Machine Vision QUIZ II

Tsai s camera calibration method revisited

Camera Model and Calibration. Lecture-12

Visual Recognition: Image Formation

3D Geometry and Camera Calibration

Agenda. Rotations. Camera calibration. Homography. Ransac

Rigid Body Motion and Image Formation. Jana Kosecka, CS 482

3-D D Euclidean Space - Vectors

CS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka

Introduction to Computer Vision

Agenda. Rotations. Camera models. Camera calibration. Homographies

Chapters 1 7: Overview

3D Transformations. CS 4620 Lecture 10. Cornell CS4620 Fall 2014 Lecture Steve Marschner (with previous instructors James/Bala)

calibrated coordinates Linear transformation pixel coordinates

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Homogeneous Coordinates. Lecture18: Camera Models. Representation of Line and Point in 2D. Cross Product. Overall scaling is NOT important.

Homogeneous coordinates, lines, screws and twists

Rectification and Distortion Correction

Structure from Motion. Prof. Marco Marcon

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.

Unit 3 Multiple View Geometry

Module 4F12: Computer Vision and Robotics Solutions to Examples Paper 2

Model Based Perspective Inversion

Augmented Reality II - Camera Calibration - Gudrun Klinker May 11, 2004

Geometric Transformations

Camera Geometry II. COS 429 Princeton University

521466S Machine Vision Exercise #1 Camera models

3D Transformations. CS 4620 Lecture Kavita Bala w/ prior instructor Steve Marschner. Cornell CS4620 Fall 2015 Lecture 11

Vector Algebra Transformations. Lecture 4

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

Perspective Projection in Homogeneous Coordinates

Transforms. COMP 575/770 Spring 2013

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

Lecture 9: Epipolar Geometry

CS201 Computer Vision Camera Geometry

Camera Calibration. COS 429 Princeton University

Camera Models and Image Formation. Srikumar Ramalingam School of Computing University of Utah

CS612 - Algorithms in Bioinformatics

The end of affine cameras

Camera models and calibration

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

Camera model and multiple view geometry

Vision Review: Image Formation. Course web page:

Chapter 18. Geometric Operations

Computer Vision Projective Geometry and Calibration. Pinhole cameras

Introduction to Homogeneous coordinates

Camera Calibration for Video See-Through Head-Mounted Display. Abstract. 1.0 Introduction. Mike Bajura July 7, 1993

Geometric camera models and calibration

Exterior Orientation Parameters

CS 664 Slides #9 Multi-Camera Geometry. Prof. Dan Huttenlocher Fall 2003

Extraction of Focal Lengths from the Fundamental Matrix

High Altitude Balloon Localization from Photographs

CS231A Midterm Review. Friday 5/6/2016

Identifying Car Model from Photographs

Chapter 3 Image Registration. Chapter 3 Image Registration

Camera Models and Image Formation. Srikumar Ramalingam School of Computing University of Utah

COMPUTER AND ROBOT VISION

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Structure from motion

Computer Vision cmput 428/615

CS231A Course Notes 4: Stereo Systems and Structure from Motion

CHAPTER 3. Single-view Geometry. 1. Consequences of Projection

Robotics (Kinematics) Winter 1393 Bonab University

Smartphone Video Guidance Sensor for Small Satellites

Stereo Vision. MAN-522 Computer Vision

Hand-Eye Calibration from Image Derivatives

Camera Projection Models We will introduce different camera projection models that relate the location of an image point to the coordinates of the

Camera model and calibration

Structure from motion

Camera calibration. Robotic vision. Ville Kyrki

CS 184, Fall 1996 Midterm #1 Professor: unknown

Hartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne

Early Fundamentals of Coordinate Changes and Rotation Matrices for 3D Computer Vision

Autonomous on-orbit Calibration Of Star Trackers

DD2429 Computational Photography :00-19:00

Lecture 3: Camera Calibration, DLT, SVD

Visualizing Quaternions

An idea which can be used once is a trick. If it can be used more than once it becomes a method

CIS 580, Machine Perception, Spring 2015 Homework 1 Due: :59AM

CIS 580, Machine Perception, Spring 2016 Homework 2 Due: :59AM

METR Robotics Tutorial 2 Week 2: Homogeneous Coordinates

Articulated Characters

CS 184, Fall 1996 Midterm #1 Professor: unknown

Game Mathematics. (12 Week Lesson Plan)

Short on camera geometry and camera calibration

Today. Today. Introduction. Matrices. Matrices. Computergrafik. Transformations & matrices Introduction Matrices

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Epipolar Geometry and the Essential Matrix

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Robotics - Projective Geometry and Camera model. Marcello Restelli

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert

C / 35. C18 Computer Vision. David Murray. dwm/courses/4cv.

Autonomous Navigation for Flying Robots

Pin Hole Cameras & Warp Functions

BIL Computer Vision Apr 16, 2014

Today. Parity. General Polygons? Non-Zero Winding Rule. Winding Numbers. CS559 Lecture 11 Polygons, Transformations

Transcription:

Perspective Projection Describes Image Formation Berthold K.P. Horn

Wheel Alignment: Camber, Caster, Toe-In, SAI,

Camber: angle between axle and horizontal plane. Toe: angle between projection of axle on horizontal plane and a perpendicular to thrust line. Caster: angle in side elevation between steering axis and vertical. SAI: angle in frontal elevation between steering axis and vertical.

Planar Target Mounted Rigidly on Rim

CCD Cameras and LED Illumination

Determining Rolling Axis

Determining Steering Axis

Photogrammetric Problems: Interior Orientation (2D 3D) Exterior Orientation (3D 2D) Absolute Orientation (3D 3D) Relative Orientation (2D 2D)

Interior Orientation: Principal Point + Principal Distance Center of Projection: (u o, v o, f)

Exterior Orientation: Rotation + Translation of Camera: R and t

Rigid body transformation: rotation and translation of object coordinate system into camera coordinate system. x c y c z c = R x t y t z t + R is a 3 3 orthonormal matrix that represents rotation, while t = (x o,y o,z o ) T represents translation. x o y o (1) z o Constraints: Orthonormality: R T R = I (six independent second order constraints). Rotation rather than reflection: det(r) =+1 (one third order constraint).

Perspective projection: Camera coordinate system (3D) into image coordinate system (2D) (*) (**). f is principal distance effective focal length u = f(x c /z c ) + u o v = f(y c /z c ) + v o (2) while (u o,v o ) T is principal point image center Interior orientation of the camera is given by (u o,v o,f) T. (*) Camera coordinate system origin is at center of projection. (**) Camera coordinate system is aligned with image axes.

Planar object: Let z t = 0 in the plane of the object. x c y c = z c r 11 r 12 r 13 r 21 r 22 r 23 r 31 r 32 r 33 x t y t 0 + x o y o (3) z o Absorbing the translation (x o,y o,z o ) T into the 3 3 matrix, and dividing the third component by f we get: x c y c z c /f = r 11 r 12 x o r 21 r 22 y o r 31 /f r 32 /f z o /f x t y t 1 (4) Now perspective projection (eq. 2) gives (if we let k = z c /f ): k(u u o ) r 11 r 12 x o k(v v o ) = r 21 r 22 y o k r 31 /f r 32 /f z o /f x t y t 1 (5) (Odd form makes it easier to match result with projective geometry formulation).

Homogeneous Representation: Homogeneous representation of point in plane uses three numbers [Wylie 70]. (u, v, w) T Actual planar coordinates are obtained by dividing first two elements by third: x = u/w and y = v/w Representation not unique since (ku, kv, kw) T corresponds to same point. A 3 3 matrix T can represent a homogeneous transformation from the object plane to the image plane. ku kv k = t 11 t 12 t 13 t 21 t 22 t 23 t 31 t 32 t 33 x t y t 1 (6)

Matching Homogeneous Transformation: We can match with k(u u o ) k(v v o ) = k ku kv k = r 11 r 12 x o r 21 r 22 y o r 31 /f r 32 /f z o /f t 11 t 12 t 13 t 21 t 22 t 23 t 31 t 32 t 33 x t y t 1 x t y t 1, (5) (6) provided t 11 = r 11, t 12 = r 12, t 13 = x o t 21 = r 21, t 22 = r 22, t 23 = y o t 31 = r 31 /f, t 32 = r 32 /f, t 33 = z o /f (7) (In addition, measurements in the image must be made in a coordinate system with the origin at the principal point, so that u o, v o = 0.)

Constraints on Transform: T same as kt (Scale factor ambiguity). Pick t 33 = 1 (leaves 8 DOF). T must satisfy two non-linear constraints t 11 t 12 + t 21 t 22 + f 2 t 31 t 32 = 0 (8) and t 11 t 11 + t 21 t 21 + f 2 t 31 t 31 = t 12 t 12 + t 22 t 22 + f 2 t 32 t 32 (9) (from the orthonormality of R). Two non-linear constraints reduce DOF from 8 to 6 rotation has 3 DOF and translation has 3 DOF.

Projective Transformation versus Perspective Projection: Given real perspective projection, one can always compute a matrix T. But it is not possible to go in the other direction for most matrices T, there is no corresponding perspective projection. Only a subset of measure of zero of homogeneous transformations T allow physical interpretation as rigid body motion and perspective projection. An arbitrary 3 3 matrix T will not satisfy the two non-linear constraints.

Disallowed Mappings: Mappings allowed by projective geometry but not by perspective projection skewing and anisotropic scaling Distort a normal perspective image by the additional operations ( ) ( ) 1 s 1 0 or 0 1 0 k result is not an image from any position with any camera orientation. Yet such distortions merely transform matrix T thus are allowed in projective geometry.

Disallowed Mappings:

Picture of a Picture: Is there a physical imaging situation that corresponds to homogeneous transformation by an arbitrary matrix T? The transformations of perspective geometry correspond to taking a perspective image of a perspective image. In this case, the overall transformation need not satisfy the non-linear constraints. But, we are interested in direct images, not pictures taken of a picture!

Finding T: Correspondence between (x i,y i, 1) T and (ku i,kv i,k) T (Horn 86): x i t 11 + y i t 12 + t 13 x i u i t 31 y i u i t 32 u i t 33 = 0 x i t 21 + y i t 22 + t 23 x i v i t 31 y i v i t 32 v i t 33 = 0 Four correspondences yield system of 8 homogeneous equations: x 1 y 1 1 0 0 0 x 1 u 1 y 1 u 1 u 1 x 2 y 2 1 0 0 0 x 2 u 2 y 2 u 2 u 2 x 3 y 3 1 0 0 0 x 3 u 3 y 3 u 3 u 3 x 4 y 4 1 0 0 0 x 4 u 4 y 4 u 4 u 4 0 0 0 x 1 y 1 1 x 1 v 1 y 1 v 1 v 1 0 0 0 x 2 y 2 1 x 2 v 2 y 2 v 2 v 2 0 0 0 x 3 y 3 1 x 3 v 3 y 3 v 3 v 3 0 0 0 x 4 y 4 1 x 4 v 4 y 4 v 4 v 4 t 11 t 12 t 13 t 21 t 22 t 23 t 31 t 32 t 33 = 0 0 0 0 0 0 0 0 0 T has 9 elements, but only 8 DOF, since any multiple of T describes same mapping. Arbitrarily pick value for one element of T, say t 33 = 1. Leaves 8 non-homogeneous equations in 8 unknown.

Recovering Orientation using Vanishing Points Consider (αa, αb, 1) T as α. That is, (a, b, 0) T. The vanishing point for line with direction (a, b) T in object plane is u = (t 11 a + t 12 b)/(t 31 a + t 32 b) v = (t 21 a + t 22 b)/(t 31 a + t 32 b) (10) The line from the COP (0, 0, 0) T, to vanishing point in image plane (u, v, f ) T, is parallel (in 3D) to line on object. Apply to x- and y-axes of object, and from the two vanishing points find directions of the two axes in camera coordinate system.

Recovering Orientation using Vanishing Points Vanishing point for x-axis is (1, 0, 0) T in object coordinate system. T(1, 0, 0) T = (t 11,t 21,t 31 ) T Vanishing point for y-axis is (0, 1, 0) T in object coordinate system. T(0, 1, 0) T = (t 12,t 22,t 32 ) T These correspond to (t 11 /t 31,t 21 /t 31 ) T and (t 12 /t 32,t 22 /t 32 ) T in the image. Connect COP (0, 0, 0) T, to the vanishing points in the image plane x = (t 11,t 21,ft 31 ) T, y = (t 12,t 22,ft 32 ) T. (11)

Recovering Rotation Make unit vectors in the direction of the x- and y-axes of the object plane. ˆx = x/ x and ŷ = y/ y So if f is known, directions of object coordinate axes (in camera coordinates) can be found from first two columns of T. Construct triad using This leads to R T = ˆx T ŷ T ẑ T ẑ = ˆx ŷ or R = ( ) ˆx ŷ ẑ (12)

Recovering Direction of Translation Homogeneous coordinates of the origin on the object plane: (0, 0, 1) T Image of origin is at (t 13 /t 33,t 23 /t 33 ) T. T(0, 0, 1) T = (t 13,t 23,t 33 ) T Connecting origin to this point yields vector parallel to t = (t 13,t 23,ft 33 ) T. (13) This is the direction of the translational vector to the object origin obtained directly from last column of T.

Recovering Translation The z component of translation is (f /M) where M is lateral magnification. Need line in object plane that is parallel to the image plane take cross-product of ẑ found above and (0, 0, 1) T : (z 1,z 2,z 3 ) T (0, 0, 1) T = ( z 2,z 1, 0) T This point on object is imaged at T( z 2,z 1, 0) T Find length of line l i from there to image of origin, at (t 13 /t 33,t 23 /t 33 ) T Divide by l t = z1 2 + z2 2 to find M = l i/l t. Finally, t = M(t 13,t 23,ft 33 ) T /(f t 33 )

Analysis In general T does not correspond to rigid body motion and perspective projection. Why were we able to calculate rigid body motion R and t? We selectively neglected inconvenient information in T! Now construct T based on recovered R and t. In general, T T! Even worse, ˆx and ŷ, as constructed, need not be orthogonal. In general R, as constructed, is not a rotation! For R to be orthonormal, (t 11,t 21,ft 31 ) T (t 12,t 22,ft 32 ) T = 0. Least squares estimate of T from data does not ensure this. Can find nearest orthogonal ˆx and ŷ, but... Same story for the other non-linear constraint on elements of T.

Sensitivity to Error in Measurements. With perfect measurements, T has desired properties. What is error in R and t as function of error in measurements? Difficult to find analytically because of complex steps. Error in rotation: Angle of δr = R1 T R 2. Monte Carlo simulation Results: Sensitivity depends on FOV. Sensitivity depends on orientation of planar target. Sensitivity depends on choice of origin!

Two Step Optimization (1) Find linear transformation T that best fits the data. (2) Find rigid body transformation R and t nearest to T. One Step Optimization Adjust R and t directly to minimize sum of squares of errors in image position. The results are different. Least squares adjustment to find a best fit T in two-step method minimizes quantities without any direct physical significance. Can be interpreted as weighted sum of image errors but with arbitrary weights.

Minimize Image Errors: Observed Points Predicted Points based on R and t

Minimizing the Image Projection Error Properly model rigid body motion and perspective projection. Error is image plane distance between predicted and observed features. Non-linear problem. No closed form solution. Find R and t that minimizes sum of squares of image errors. Easy optimization if good representation for rotation is chosen. Unit quaternions (equivalently Pauli Spin matrices, Euler parameters, etc.) Newton-Raphson method is then sufficient. Seven unknowns ( q and t) and one constraint ( q q = 1).

Comparing Error Sensitivity Sample Experiment: Target: Four features in 168 mm by 168 mm pattern Distance: 1600 mm Focal length: 18mm Pixel size: 8.4 µ by 8.4 µ Object Plane Orientation: normal 60 from optical axis Error in image plane measurements: 0.05 pixel (1/20) Monte Carlo simulation. Projective geometry (PG) error is 1.1. Noise gain 22 degrees / pixel. Perspective projection (PP) method error is 0.04. Noise gain 0.8 degrees / pixel. Projective Geometry based method more than an order of magnitude worse. With three points, PP method error is still only 0.06. With more points, errors decrease in proportion to 1/ N Projective Geometry based method bad under variety of conditions tested.

Image Analyst Pattern

Image Analyst Centroiding Method Green dot: True centroid of original disk Red dot: Centroid of thresholded binary image Line Blob Centroiding Method Green dot: True centroid of original disk Red dot: Line blob estimate of centroid

Upper curve: Image Analyst method 0.15 Upper curve: trend 0.310 / sqrt(d) Lower curve: "Line Blob" method Lower curve: trend 0.053 / sqrt(d) 0.1 0.05 0.0 0 5 10 15 20 25 30 35 40 45 50 Diameter of blobs in pixels

Main Algorithm (patent disclosure) Image Points matrix T R and t Section 9.4 Main Algorithm Image Fitting (working implementation) Image Points Ro and to Rn and tn First Guess Iterative Improvement

Patent Disclosure Image Analyst Main Algorithm Image 1/10 pixel X 20 deg/pixel 2 deg Working Implementation Line Blob Image Fitting Image 2/100 pixel X 1 deg/pixel 2/100 deg

Summary Methods based on projective geometry are fundamentally different from methods based on perspective projection; Methods based on projective geometry yield a transformation matrix T that in general does not correspond to a physical imaging situation that is, rotation, translation, and perspective projection; Optimization methods based on actual physical imaging situation produce considerably more accurate results. Additional Notes One can make very accurate measurements using images; One needs to carefully design the target to get high accuracy; One needs to carefully design the target to get reliability.

References: Absolute Orientation: Closed Form Solution of Absolute Orientation using Unit Quaternions, Journal of the Optical Society A, Vol. 4, No. 4, pp. 629 642, Apr. 1987. Relative Orientation: Relative Orientation Revisited, Journal of the Optical Society of America, A, Vol. 8, pp. 1630 1638, Oct. 1991. Interior Orientation / Exterior Orientation: Tsai s Camera Calibration Method Revisited, on http:\\www.ai.mit.edu\people\bkph Detailed Example: Projective Geometry Considered Harmful, on http:\\www.ai.mit.edu\people\bkph

Projective Geometry Considered Harmful Berthold K.P. Horn