Epipolar Geometry and the Essential Matrix

Similar documents
3D Reconstruction with two Calibrated Cameras

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Perspective Projection in Homogeneous Coordinates

Multiple Views Geometry

Two-view geometry Computer Vision Spring 2018, Lecture 10

CS201 Computer Vision Camera Geometry

Stereo Vision. MAN-522 Computer Vision

Computer Vision Projective Geometry and Calibration. Pinhole cameras

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Multiple View Geometry in Computer Vision

The end of affine cameras

C / 35. C18 Computer Vision. David Murray. dwm/courses/4cv.

Comments on Consistent Depth Maps Recovery from a Video Sequence

calibrated coordinates Linear transformation pixel coordinates

Fundamental Matrix & Structure from Motion

CS 664 Slides #9 Multi-Camera Geometry. Prof. Dan Huttenlocher Fall 2003

Multiple View Geometry in Computer Vision

CS231A Course Notes 4: Stereo Systems and Structure from Motion

Multi-View Geometry Part II (Ch7 New book. Ch 10/11 old book)

CSE 252B: Computer Vision II

MAPI Computer Vision. Multiple View Geometry

Lecture 14: Basic Multi-View Geometry

1 (5 max) 2 (10 max) 3 (20 max) 4 (30 max) 5 (10 max) 6 (15 extra max) total (75 max + 15 extra)

Multiple View Geometry

CS-9645 Introduction to Computer Vision Techniques Winter 2019

Lecture 9: Epipolar Geometry

Rectification and Distortion Correction

Rigid Body Motion and Image Formation. Jana Kosecka, CS 482

Unit 3 Multiple View Geometry

Recovering structure from a single view Pinhole perspective projection

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches

Elements of Computer Vision: Multiple View Geometry. 1 Introduction. 2 Elements of Geometry. Andrea Fusiello

Homogeneous Coordinates. Lecture18: Camera Models. Representation of Line and Point in 2D. Cross Product. Overall scaling is NOT important.

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt

Descriptive Geometry Meets Computer Vision The Geometry of Two Images (# 82)

Announcements. Stereo

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

3D Computer Vision. Structure from Motion. Prof. Didier Stricker

Epipolar Geometry class 11

Structure from motion

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

3D Geometry and Camera Calibration

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert

CSE 252B: Computer Vision II

Multiple View Geometry in Computer Vision

The Geometry Behind the Numerical Reconstruction of Two Photos

Camera Calibration Using Line Correspondences

Structure and Motion for Dynamic Scenes - The Case of Points Moving in Planes

Computer Vision Projective Geometry and Calibration. Pinhole cameras

Vision Review: Image Formation. Course web page:

Lecture 3: Camera Calibration, DLT, SVD

Structure from motion

Epipolar Geometry CSE P576. Dr. Matthew Brown

Last lecture. Passive Stereo Spacetime Stereo

Camera Model and Calibration

Projective geometry for Computer Vision

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

CS 231A: Computer Vision (Winter 2018) Problem Set 2

Image Rectification (Stereo) (New book: 7.2.1, old book: 11.1)

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Computer Vision cmput 428/615

Two-View Geometry (Course 23, Lecture D)

Lecture 6 Stereo Systems Multi- view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-24-Jan-15

Vectors and the Geometry of Space

A Stereo Machine Vision System for. displacements when it is subjected to elasticplastic

Linear Multi View Reconstruction and Camera Recovery Using a Reference Plane

1 Projective Geometry

Multiple View Geometry in Computer Vision Second Edition

But First: Multi-View Projective Geometry

CSE 252B: Computer Vision II

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.

Visual Recognition: Image Formation

A Factorization Method for Structure from Planar Motion

Structure from Motion

Fundamental Matrix & Structure from Motion

Announcements. Stereo

Mathematics of a Multiple Omni-Directional System

Constrained Structure and Motion From Multiple Uncalibrated Views of a Piecewise Planar Scene

Computer Vision: Lecture 3

Introduction to Homogeneous coordinates

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

Elements of Geometric Computer Vision. Andrea Fusiello

Rectification and Disparity

Structure from Motion

CS231M Mobile Computer Vision Structure from motion

Augmented Reality II - Camera Calibration - Gudrun Klinker May 11, 2004

Camera model and multiple view geometry

Stereo and Epipolar geometry

MERGING POINT CLOUDS FROM MULTIPLE KINECTS. Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia

Short on camera geometry and camera calibration

CHAPTER 3. Single-view Geometry. 1. Consequences of Projection

Lecture 5 Epipolar Geometry

Vision par ordinateur

CS 231A Computer Vision (Winter 2015) Problem Set 2

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Epipolar geometry. x x

Camera Models and Image Formation. Srikumar Ramalingam School of Computing University of Utah

Alternative interpretation of the Plücker quadric s ambient space and its application

Transcription:

Epipolar Geometry and the Essential Matrix Carlo Tomasi The epipolar geometry of a pair of cameras expresses the fundamental relationship between any two corresponding points in the two image planes, and leads to a key constraint between the coordinates of these points that underlies visual reconstruction. The first Section below describes the epipolar geometry. The Section thereafter expresses its key constraint algebraically. 1 The Epipolar Geometry of a Pair of Cameras Figure 1 shows the main elements of the epipolar geometry for a pair of cameras. P projection ray projection ray epipolar plane center of projection epipolar line of q p epipole e camera C I baseline epipole f q epipolar line of p camera D J center of projection Figure 1: Essential elements of the epipolar geometry of a camera pair. The world point P and the centers of projection of the two cameras identify a plane in space, the epipolar plane of point P. The Figure shows a triangle of this plane, delimited by the two projection rays and by the baseline of the camera pair, that is, the line segment that connects the two centers of projection. 1 If the image planes are thought of extending indefinitely, the baseline intersects the two image planes at two points called the epipoles of the two images. In particular, if the cameras are arranged so that the baseline is parallel to an image plane, then the corresponding epipole is a point at infinity. The epipoles are fixed points for a given camera pair configuration. With cameras somewhat tilted towards each other, and with a sufficiently wide field of view, the epipoles would be image points. Epipole e in the image I taken by camera C would be literally the image of the center of projection of camera D in 1 We use the term baseline for the line segment. However, this term is also often used for the entire line through the two centers of projection. 1

I, and vice versa. Even if the two cameras do not physically see each other, we maintain this description in an abstract sense: each epipole is the image of one camera in the other image. The epipolar plane intersects the two image planes along the two epipolar lines of point P, each of which passes by construction through one of the two projection points p and q and one of the two epipoles. Thus, epipolar lines come in corresponding pairs, and the correspondence is established by the single epipolar plane for the given point P. For a different world point P, the epipolar plane changes, and with it do the image projections of P and the epipolar lines. However, all epipolar planes contain the baseline. Thus, the set of epipolar planes forms a pencil of planes supported by the line through the baseline, and the epipoles are fixed. Suppose now that we are given the two images I and J taken by cameras C and D and a point p in I. We do not know where the corresponding point q is in the other image, nor where the world point P is, except that P must be somewhere along the projection ray of p. However, if we know the relative position and orientation of the two cameras, we know where centers of projection are relative to each other. The two centers of projection and point p identify the epipolar plane, and this in turn determines the epipolar line of point p in image J. The point q must be somewhere on this line. This same construction holds for any other point p on the epipolar line in image I. To understand what the epipolar constraint expresses, consider that the projection rays for two arbitrary points in the two images are generically two skew lines in space. The projection rays of two corresponding points, on the the other hand, are coplanar with each other and with the baseline. The epipolar geometry captures this key constraint, and pairs of point that do not satisfy the constraint cannot possibly correspond to each other. 2 The Essential Matrix This section expresses the epipolar constraint described in the previous note algebraically. To emphasize the distinction between coordinate vectors in two different reference frames, coordinates in the second system have a prime. The standard reference system for a camera C is a right-handed Cartesian coordinate system with its origin at the center of projection of C, its positive Z axis pointing towards the scene along the optical axis of the lens, and its X axis pointing to the right 2 along the rows of the camera sensor. As a consequence, the Y axis points downwards along the columns of the sensor. Coordinates in the standard reference system are measured in units of focal distance. Let p = p 1 p 2 f and q = denote the image coordinates of the projections of the same 3D point P in the standard reference systems of two cameras C and D, and let p = R(p t) (1) be the rigid transformation between the two reference systems. Note that the coordinates of p are in reference system C and those of q are in reference system D: An image point is measured in its own camera s reference system. 2 When the camera is upside-up and viewed from behind it, as when looking through its viewfinder. q 1 q 2 f 2

Solving equation (1) for p yields p = R T p + t = R T (p + Rt) so that the reverse transformation is p = R (p s ) if we let R = R T and s = Rt. (2) When expressed in the reference system of camera C, the directions of the projection rays through corresponding image points with coordinates p and q are along the vectors p and R T q, and the baseline in this reference system is along the translation vector with Euclidean coordinates t. Coplanarity of these three vectors can be expressed by stating that their triple product is zero: (R T q ) T (t p) = 0 that is, (q ) T R (t p) = 0 or (q ) T R [t] p = 0 where t = [ t(1) t(2) t(3)] T ] and [t] = 0 t 3 t 2 t 3 0 t 1 t 2 t 1 0 is the matrix that expresses the cross-product of t with any other vector. In summary, for corresponding points p and q the following equation holds: where (q ) T Ep = 0 (3) E = R [t]. (4) Equation (3) is called the epipolar constraint and the matrix E is called the essential matrix. Equation (3) expresses the coplanarity between any two points p and q on the same epipolar plane for two fixed cameras. If p is fixed in image I, then any point q in image J for which equation (3) holds must be on the epipolar line of p. Formally, the product µ p = Ep (5) for fixed p is a column vector, and equation (3) becomes (q ) T µ p = 0. This is the equation of the epipolar line in image J. Conversely, if q is fixed in the second image, then the product λ T q = (q ) T E (6) is a row vector, and equation (3) becomes which is the equation of the epipolar line of q in image I. λ T q p = 0 (7) 3

Since the epipole in image I belongs to all epipolar lines in I, the vector e of its homogeneous coordinates must satisfy equation (7) regardless of what point q is used in the definition (6) of λ q. This can happen only if e is in the null space of E, which must therefore be degenerate. Algebraically, it is easy to see that the rank of E is two for any nonzero t. To this end, note first that the matrix [t] has rank two if t is nonzero, because [t] t = t t = 0 so the null space of [t] is the line through the origin and along t. Since R is full rank, also the product E = R [t] has rank 2 if t 0, and the null space of E is spanned by t as well. Since [t] is skew-symmetric, [t] T = [t]. So from equations (2) and (4) we obtain E T s = [t] R T ( Rt) = [t] t = 0 so that s spans the null space of E T, that is, the left null space of E. Since the left epipole e is along t and the right epipole f is along s, we can summarize this discussion by stating that the epipole e in image I and the translation vector t are both in the null space of E: Ee = Et = 0 and the epipole f in image J and the translation vector s are both in the left null space of E: (f ) T E = (s ) T E = 0 T. The essential matrix E has special structure, as we now show. For any vector v orthogonal to t, the definition of cross product yields [t] v = t v. Because of this, the matrix [t] maps all unit vectors in its row space into vectors of magnitude t. In other words, the two nonzero singular values of [t] are equal to each other (and to t ). Since multiplication by an orthogonal matrix does not change the matrix s singular values, we conclude that the essential matrix E has two nonzero singular values equal to each other, and a zero singular value. The left and right singular vectors u 3 and v 3 corresponding to the zero singular value of E are unit vectors along the epipoles and the translation vectors, v 3 e t and u 3 f s. (8) In these expressions, the symbol means proportional to, or equal up to a multiplicative constant. Since the other two singular values of E are equal to each other, the corresponding singular vectors are arbitrary, as long as they form orthonormal triples with u 3 and v 3. The preceding discussion is summarized in Table 1. The epipolar constraint (3) is used in two different contexts. In stereo vision, R and t and therefore E are known. Given a point p in I, the epipolar constraint allows restricting the search for a corresponding point q to the epipolar line of p. In visual reconstruction, several pairs (p i, q i ) of corresponding points are given, and equation (3) for each pair yields a linear equation in the entries of E. From this, E and then R and t can be found, as we will see in a later note. 4

For a camera pair (C, D) with nonzero baseline, let p = R(p t) and p = R (p s ) with R = R T and s = Rt be the coordinate transformations between points p in C and points p in D. The essential matrix of the camera pair is the matrix E = R [t] where [t] = 0 t 3 t 2 t 3 0 t 1 t 2 t 1 0 The epipole e in the image I taken by C and that of f in the image J taken by D satisfy Ee = E T f = 0 and also Et = E T s = 0. A point p in image I and its corresponding point q in image J satisfy the epipolar constraint This equation can also be written as follows: (q ) T Ep = 0. λ T q p = (µ p) T q = 0. where λ q = E T q and µ p = Ep are the vectors of coefficients of the epipolar line of q in image I and that of p in image J respectively. The singular value decomposition of E is E UΣV T = [ u 1 u 2 u 3 ] diag(1, 1, 0) [ v1 v 2 v 3 ] T (9) where u 1, u 2, v 1, v 2 are any vectors for which U and V become orthogonal and v 3 e t and u 3 f s Table 1: Definition and properties of the essential matrix. 5