Multiple Views Geometry

Similar documents
calibrated coordinates Linear transformation pixel coordinates

Epipolar Geometry and the Essential Matrix

Structure from motion

Projective geometry for Computer Vision

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz

Unit 3 Multiple View Geometry

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Projective geometry for 3D Computer Vision

3D reconstruction class 11

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Structure from motion

Epipolar Geometry class 11

Epipolar geometry. x x

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction

Stereo Vision. MAN-522 Computer Vision

Stereo and Epipolar geometry

CSE 252B: Computer Vision II

Image Transfer Methods. Satya Prakash Mallick Jan 28 th, 2003

Structure from Motion

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

Epipolar Geometry and Stereo Vision

BIL Computer Vision Apr 16, 2014

Projective geometry, camera models and calibration

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Structure from Motion CSC 767

Structure from Motion. Prof. Marco Marcon

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

1 Projective Geometry

Lecture 9: Epipolar Geometry

EE795: Computer Vision and Intelligent Systems

CS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka

The end of affine cameras

3D Modeling using multiple images Exam January 2008

Structure from Motion

Two-view geometry Computer Vision Spring 2018, Lecture 10

MAPI Computer Vision. Multiple View Geometry

3D Reconstruction from Two Views

CS-9645 Introduction to Computer Vision Techniques Winter 2019

N-Views (1) Homographies and Projection

Two-View Geometry (Course 23, Lecture D)

Epipolar Geometry and Stereo Vision

Last lecture. Passive Stereo Spacetime Stereo

Multiple View Geometry in Computer Vision

Computer Vision Projective Geometry and Calibration. Pinhole cameras

CS201 Computer Vision Camera Geometry

Structure from motion

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

Robust Geometry Estimation from two Images

3D Computer Vision. Structure from Motion. Prof. Didier Stricker

Recovering structure from a single view Pinhole perspective projection

Rectification and Distortion Correction

But First: Multi-View Projective Geometry

Contents. 1 Introduction Background Organization Features... 7

CS 664 Slides #9 Multi-Camera Geometry. Prof. Dan Huttenlocher Fall 2003

Computer Vision cmput 428/615

Structure from Motion

Multiple View Geometry

Step-by-Step Model Buidling

Elements of Computer Vision: Multiple View Geometry. 1 Introduction. 2 Elements of Geometry. Andrea Fusiello

Invariance of l and the Conic Dual to Circular Points C

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert

3D Reconstruction with two Calibrated Cameras

Lecture 6 Stereo Systems Multi- view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-24-Jan-15

Camera Geometry II. COS 429 Princeton University

Epipolar Geometry CSE P576. Dr. Matthew Brown

C / 35. C18 Computer Vision. David Murray. dwm/courses/4cv.

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

A simple method for interactive 3D reconstruction and camera calibration from a single view

Multiple View Geometry in Computer Vision Second Edition

A Summary of Projective Geometry

Augmented Reality II - Camera Calibration - Gudrun Klinker May 11, 2004

C280, Computer Vision

Announcements. Stereo

Visual Recognition: Image Formation

Multiple View Geometry in Computer Vision

Midterm Exam Solutions

Lecture 5 Epipolar Geometry

A Canonical Framework for Sequences of Images

Structure and Motion for Dynamic Scenes - The Case of Points Moving in Planes

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Structure from Motion and Multi- view Geometry. Last lecture

CS 2770: Intro to Computer Vision. Multiple Views. Prof. Adriana Kovashka University of Pittsburgh March 14, 2017

Lecture 6 Stereo Systems Multi-view geometry

Multiple View Geometry in Computer Vision

Multiple View Geometry. Frank Dellaert

CS6670: Computer Vision

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt

CS231A Course Notes 4: Stereo Systems and Structure from Motion

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

CS231M Mobile Computer Vision Structure from motion

Inner Camera Invariants and Their Applications

A Stratified Approach for Camera Calibration Using Spheres

Computer Vision Lecture 17

Fundamental Matrix & Structure from Motion

Epipolar Geometry and Stereo Vision

Perspective Projection in Homogeneous Coordinates

Computer Vision: Lecture 3

A Factorization Method for Structure from Planar Motion

Transcription:

Multiple Views Geometry Subhashis Banerjee Dept. Computer Science and Engineering IIT Delhi email: suban@cse.iitd.ac.in January 2, 28

Epipolar geometry Fundamental geometric relationship between two perspective cameras: epipole: is the point of intersection of the line joining the optical centers - the baseline - with the image plane. The epipole is the image in one camera of the optical center of the other camera. epipolar plane: is the plane defined by a 3D point and the optical centers. Or, equivalently, by an image point and the optical centers. epipolar line: is the line of intersection of the epipolar plane with the image plane. It is the image in one camera of a ray through the optical center and the image point in the other camera. All epipolar lines intersect at the epipole. Epipolar geometry provides a fundamental constraint for the correspondence problem. Epipolar geometry: uncalibrated case Given the two cameras and 2, we have the camera equations: x = P X and x 2 = P 2 X

The optical center projects as P i X = Writing P i = [P i P i t i ] where P i is 3 3 non-singular we have that t i is the optical center. [ ] ti [P i P i t i ] = The epipole e 2 in the second image is the projection of the optical center of the first image: [ ] e 2 = P t 2 The projection of point on infinity along the optical ray < t,x > on to the second image is given by: x 2 = P 2 P x The epipolar line < e 2,x 2 > is given by the cross product e 2 x 2. If [e 2 ] is the 3 3 antisymmetric matrix representing cross product with e 2, then we have that the epipolar line is given by [e 2 ] P 2 P x = Fx Any point x 2 on this epipolar line satisfies x 2 T Fx = F is called the fundamental matrix. It is of rank 2 and can be computed from 8 point correspondences. Clearly Fe = (degenerate epipolar line) and e 2 T F =. The epipoles are obtained as the null spaces of F.

.2 Epipolar geometry: calibrated case There are two camera coordinate systems related by R,T X = RX + T Taking the vector product with T followed by the scalar product with X X (T RX) = which expresses that vectors OX, O X and OO are coplanar. This can be written as where is the Essential matrix. X T EX = E = [T] R Image points and rays in Euclidean 3-space are related by: x X x X y = C Y and y = C Y Z Z Hence, we have x T C T EC x = Thus, the relation between the essential and fundamental matrix is: F = C T EC

.3 Epipolar geometry: examples

.4 How to find correspondences. Given corners: 2. Unguided matching: Obtain a small number of seed matches using crosscorrelation and pessimistic thresholds.

3. Compute epipolar geometry and reject outliers. 4. Guided matching: Search for matches in a band about the epipolar lines.

2 Structure determination 2. Overview Overview Image Pair Calibrate: camera camera 2 epipolar geometry calibrated route uncalibrated route / C C R T Compute F from image correspondences 3D structure (metric) 3D structure (projective) 2

2.2 Projective structure using a five point basis Select the projection center of the first camera as the coordinate origin (,,, ) T and that of the second camera as the unit point (,,, ) T. Complete the 3D basis by choosing 3 other visible 3D points A, A 2 and A 3 such that no four of the five points are coplanar. Let a, a 2, a 3 and a, a 2, a 3 be the image projections of A, A 2 and A 3 in the two images. Make a projective transformation of each image such that these 3 points and the epipoles form a standard basis a = a =, a 2 = a 2 =, a 3 = a 3 =, e = e = Also fix the 3D coordinates as A =, A 2 =, A 3 = It follows that P = and P = Once P s are known finding the projective structure of the other points is straightforward. Localization errors in any of the points affect subsequent computation 2.3 Robust computation of projective structure Choose the first camera as P = [I ] which corresponds to the world coordinate system having its origin at the optical center of the first camera, and its axes aligned with the camera axes. It is always possible to make this choice.

The choice of the second camera consistent with the fundamental matrix is P = [M e ] and the fundamental matrix is given as: F = e M e is the projection of the optical center of the first camera, ( T, ) T, onto the second image [ ] P = e Similarly, the optical center of the second camera is is ((M e ) T, ) T, and its projection onto the first image is e. [ ] M e P = M e = e Given P and P, the 3D point X i corresponding to a pair of image point x i and x i can be computed by intersecting the back-projected rays [ ] [ ] xi X i = + λ [ ] [ ] M e = M x + µ i To summarize:. Compute the fundamental matrix F from x i x i. 2. Decompose F as F = e M. 3. Compute 3D points X i by intersecting back-projected rays using P = [I ] and P = [M e ]

2.4 Projective ambiguity Although the two projection matrices define F the converse is not true. Suppose P and P are two matrices consistent with F, then x = PX and x = P X But, if H is any arbitrary homography of 3-space, then x = (PH )(HX) and x = (P H )(HX ) Thus, we can only recover projective structure modulo an unknown homography. How to recover H PE, the homography that upgrades the projective structure to Euclidean?

2.5 Examples of projective reconstruction Uncalibrated reconstruction example Original images Projective/ane reconstructions Euclidean reconstructions 2

3 Planes 3. Planes and epipolar geometry The homography between two planes is represented by a 3 3 matrix x = Tx The epipoles are image projections of a 3D point lying on all planes, because the line joining the optical centers intersects any plane. e = Te Four coplanar points and two points off the plane determine the epipolar geometry.. Compute the plane projective transformation T, such that x i = Tx i, i =...4. 2. Determine the epipole e in the x image as the intersection of the lines (Tx 5 x 5) and (Tx 6 x 6). 3. The epipolar line in the x image of any other point x is given as (Tx e ). 4. F = e T

Plane Projective Transfer Original images Transfer and superimposed images 3

3.2 Distinguished planes and F Let T be a homography between two images through a plane π. Thus we have (projectively) x = T x Given F we know e and e. Now if F = [e ] T is a valid decomposition, then so is F = [e ] T, where where a is an arbitrary 3 vector. T = T + e a T Thus, the most general projection matrices consistent with F are P = [I ] and P = [T + e a be ] This is four parameter family, b is an arbitrary scalar. The matrix T = T + e a T is a projective transformation of the plane π represented by the four vector (,,, ) T, since for points on this plane 4 =, X and X X x = P Y Z = [I ] Y X Z = Y Z X X x = P Y Z = [T be ] Y X Z = T Y = Tx Z Thus, the choice of a corresponds to which actual plane in 3-space is π. 4 Upgrading projective structure to Euclidean: Self Calibration and stratification 4. From projective to affine Identifying the π is equivalent to updating the projective structure to affine structure.

An affine reconstruction is where the point coordinates {X A } are known up to an affine transformation of their Euclidean values {X E } [ ] A X A = T A X E ta = {A,t A } are unknown and same for all points. Choosing a corresponding to the real plane at infinity updates the projective reconstruction to affine. T then is the infinite homography, H, the homography through the plane at infinity. 4.2 From affine to metric Suppose that P E = C [I ] and P E = C [R t] i.e., we assume that the two camera positions have fixed internals. Suppose, also, that we have recovered affine structure {X A } and a pair of projection matrices P = [I ] and P = [T e ] Then we know that T = H. We also know that where, P E = PH and P E = P H H = [ C k is the affine transformation that upgrades the affine reconstruction to Euclidean. The above gives us ] T = H = CRC Rearranging and using RR T = I, we have where K = T KT T K = CC T Thus, the internal parameter matrix can be computed from the infinite homography using Cholesky decomposition.

5 Some Applications 5. Novel view generation using view morphing [Seitz and Dyer, 996] Parallel views Suppose the camera has moved from the world origin to the position (C x,c y, ) and the two projection matrices are f Π = f

and Π = f f C x f f C y Let p and p are image projections of a scene point P. Linear interpolation of p and p yields ( s)p + sp = ( s) Z Π P + s Z Π P = Z Π sp where Π s = ( s)π + sπ Π s represents a camera with center C s = (sc x,sc y, ) and focal length f s = ( s)f + sf. Linear morph is consistent with a physical pin-hole camera. Non-parallel camera

When two images share the optical center they are related by a planar homography. Thus there exist planar homographies between I Î, I Î and I s Îs. View morphing can be achieved as a 3 step process. The results also hold for uncalibrated cameras. Demo Monalisa

5.2 Video compression Tracking triangulation Sequence Sequence 2 5.3 Panoramas When the camera rotates about its optical center or undergoes pure translation then the views are related by a planar homography. Recovering the homography from tracking allows us to stitch the images in to an image mosaic. Demo