Three-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras

Similar documents
Multiple View Geometry

Lecture 14: Basic Multi-View Geometry

Stereo Vision. MAN-522 Computer Vision

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

Rectification and Distortion Correction

Correspondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]

C / 35. C18 Computer Vision. David Murray. dwm/courses/4cv.

Dense 3D Reconstruction. Christiano Gava

Chaplin, Modern Times, 1936

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Dense 3D Reconstruction. Christiano Gava

Final project bits and pieces

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,

5LSH0 Advanced Topics Video & Analysis

Computer Vision I. Dense Stereo Correspondences. Anita Sellent 1/15/16

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Depth from two cameras: stereopsis

Depth from two cameras: stereopsis

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Computer Vision Lecture 17

Computer Vision Lecture 17

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Epipolar Geometry and Stereo Vision

Outline. ETN-FPI Training School on Plenoptic Sensing

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Epipolar Geometry CSE P576. Dr. Matthew Brown

Depth Camera for Mobile Devices

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.

Epipolar Geometry and the Essential Matrix

Introduction to Homogeneous coordinates

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Module 5: Video Modeling Lecture 28: Illumination model. The Lecture Contains: Diffuse and Specular Reflection. Objectives_template

Stereo and structured light

DD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication

calibrated coordinates Linear transformation pixel coordinates

LUMS Mine Detector Project

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

Rectification. Dr. Gerhard Roth

CS4495/6495 Introduction to Computer Vision. 3B-L3 Stereo correspondence

Stereo Image Rectification for Simple Panoramic Image Generation

CS 664 Slides #9 Multi-Camera Geometry. Prof. Dan Huttenlocher Fall 2003

CHAPTER 3. Single-view Geometry. 1. Consequences of Projection

Epipolar Geometry and Stereo Vision

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?

Information page for written examinations at Linköping University TER2

Epipolar geometry contd.

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.

Stereo imaging ideal geometry

Lecture'9'&'10:'' Stereo'Vision'

Multiple View Geometry

Recap from Previous Lecture

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision

MERGING POINT CLOUDS FROM MULTIPLE KINECTS. Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia

The end of affine cameras

Two-view geometry Computer Vision Spring 2018, Lecture 10

CS5670: Computer Vision

Computer Vision I. Announcement. Stereo Vision Outline. Stereo II. CSE252A Lecture 15

Instance-level recognition I. - Camera geometry and image alignment

Epipolar Geometry in Stereo, Motion and Object Recognition

Stereo Observation Models

12/3/2009. What is Computer Vision? Applications. Application: Assisted driving Pedestrian and car detection. Application: Improving online search

Rectification and Disparity

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR

Model-Based Stereo. Chapter Motivation. The modeling system described in Chapter 5 allows the user to create a basic model of a

Multiple Views Geometry

MAPI Computer Vision. Multiple View Geometry

Computer Vision Projective Geometry and Calibration. Pinhole cameras

N-Views (1) Homographies and Projection

CV: 3D to 2D mathematics. Perspective transformation; camera calibration; stereo computation; and more

An idea which can be used once is a trick. If it can be used more than once it becomes a method

Announcements. Stereo

Vision Review: Image Formation. Course web page:

Binocular Stereo Vision. System 6 Introduction Is there a Wedge in this 3D scene?

Camera Model and Calibration

Depth Sensors Kinect V2 A. Fornaser

CS6670: Computer Vision

Ray Optics. Lecture 23. Chapter 34. Physics II. Course website:

Comments on Consistent Depth Maps Recovery from a Video Sequence

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems

Other approaches to obtaining 3D structure

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

Multiple View Geometry in Computer Vision

Pin Hole Cameras & Warp Functions

EE795: Computer Vision and Intelligent Systems

Computer Graphics. Lecture 13. Global Illumination 1: Ray Tracing and Radiosity. Taku Komura

Pattern Feature Detection for Camera Calibration Using Circular Sample

Local Illumination. CMPT 361 Introduction to Computer Graphics Torsten Möller. Machiraju/Zhang/Möller

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

1 Projective Geometry

Scalable geometric calibration for multi-view camera arrays

Stereo matching. Francesco Isgrò. 3D Reconstruction and Stereo p.1/21

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

EE795: Computer Vision and Intelligent Systems

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

WORCESTER POLYTECHNIC INSTITUTE

Transcription:

Three-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inria.fr http://perception.inrialpes.fr/

Outline The geometry of active stereo. Surface reflectivity. Camera-projector calibration Active stereo matching and reconstruction

Generalities The basic principle is to associate a light projector with a standard camera. The projector and the camera share the same geometric model. Optical (non-linear distorsions) are present with both devices. They will be treated separately. 3D reconstruction is based on the same principle as stereo: sensor calibration and matching. Instead of matching two images, we match a pattern with itself after being backscattered by scene objects.

The Geometry of Active Stereo Perspective effects (foreshortening). Surface discontinuities. Cluttered scenes.

Foreshortening

Discontinuities

Clutter

Surface Reflectivity and External Illumination This is a scientific topic in its own right. Reflectance (or absorption) depends on the material properties and of the wavelength of the incident light. A rough way to think of surface reflectivity is to divide surfaces into diffuse (matte) and specular (mirror). In general a surface is a combination of these effects plus other ones: metal, glossy, rough, transparent, translucent, etc. cause problems in practice. External illumination (daylight, artificial light, etc.) acts as noise added to the infra-red light emitted by the projector.

Reflectivity

Artifacts with the Kinect Camera

Back to Geometry Consider an IR camera and a projector in general positions (before rectification). Parameters to be estimated: The internal parameters of the IR camera The external parameters of the IR camera The projection matrix of the light projector

IR-Camera and Projector Geometry and Calibration

External IR-Camera Calibration The basic equation is m [ R t ] M where M lies on the plane Z = 0, hence we obtain: sx sy = [ r 1 r 2 t ] X Y s 1 Each correspondence between an image point m i and a calibration point M i yields two homogeneous linear equations in the unknown extrinsics: { xi (X i r 13 + Y i r 23 + t 3 ) = X i r 11 + Y i r 12 + t 1 y i (X i r 13 + Y i r 23 + t 3 ) = X i r 21 + Y i r 22 + t 2 In practice it is useful to constrain the parameters of the rotation matrix, i.e., r 1 = r 2 = 1, r 1 r 2 = 0. Various optimization methods can be used in practice.

External IR-Camera Parameters The external camera matrix is: [ R t ] = [ r1 r 2 r 1 r 2 t ] The coordinates of the planar set of points in the IR-camera frame: X Y = [ r 1 r 2 t ] X Y Z 1 Recall the simple relationship between m and M when the latter is in camera frame: X x Y Z = Z y 1

Point/Plane Duality in Projective Space The point M belongs to the plane µ: αx + βy + γz + δh = 0 or µ M = 0 with the notation M = ( X Y Z H ) and H = 1. The 4 4 point-transformation D (a rotation followed by a translation) maps the coordinates of a point from the calibration frame to the camera frame: M = D M Let µ M = 0 with µ = ( 0 0 1 0 ) be the plane of the calibration pattern. Substituting M we obtain ( D µ ) M = 0 with: [ D = R 0 t R 1 ] and µ = D µ The latter is the plane-transformation or the dual of the point-transformation.

Projector Calibration We assume that the planar calibration pattern is in the same position as before. A ray of light through the projector image-point p = (u v 1) is p P M. This ray of light is observed in the IR-camera image at m. Recall that M = Zm and that µ M = 0. Hence the depth Z is: δ Z = αx + βy + γ For each camera-projector point correspondence m i p i we obtain two equations: { ui [Z i (x i p 31 + y i p 32 + p 33 ) + p 34 ] = Z i (x i p 11 + y i p 12 + p 13 ) + p 14 v i [Z i (x i p 31 + y i p 32 + p 33 ) + p 34 ] = Z i (x i p 21 + y i p 22 + p 23 ) + p 24 This requires several such correspondences and several parallel planes.

Projector s Intrinsics and Extrinsics If enough correspondences are available, one can estimate the projection matrix P either linearly or with the help of a non-linear optimization method (bundle adjustment). Intrinsic and extrinsic parameter can be made explicit by identification: P K [ R t ] or [ P3 3 p 4 ] [ KR Kt ] Making use of RR = I one can easily extract the intrinsic and extrinsic parameters from P.

Camera-Projector Calibration Step-by-Step 1 Calibrate the internal parameters of the IR camera (OpenCV). 2 Calibrate the external parameters with respect to a single planar calibration pattern (OpenCV). 3 Transform the calibration plane in the camera coordinate frame. 4 Using the same position of the calibration pattern to obtain m i p i correspondences. 5 Move the calibration pattern in several parallel positions. 6 Estimate the calibration parameters of the projector in the camera frame. 7 Rectify the camera-projector setup to align the camera image with the projector image.

Active Stereo Reconstruction With a calibrated & rectified camera-projector setup, the relationship between the depth Z and the disparity d is: Z = b d with d = x p x c where b is the baseline, x p and x c are the horizontal coordinates of a projector image-point and a camera image-point. The coordinates of a point are: X Y Z = Z x c y c 1

Active Stereo Matching The problem is to find x c x p correspondences along epipolar lines (along the rows). Each pixel of the projector image must be signed such that it can be easily recognized in the IR-camera image: Temporal signature; Spatial signature, or a combination of both Define a K K T window around each pixel. Arrange the pixels in a vector of size N = K K T

Cross-Correlation Let the entries of two N-dimensional vectors v and w correspond to N realizations of two random variables v and w. Expectation: E[v] = 1 N N n=1 p nv n with N n=1 p n = 1. Variance: var[v] = E[v 2 ] E[v] 2 Standard deviation: σ v = var[v] Standardization: ˆv = v E[v] σ v Cross-correlation: ρ vw = corr[ˆv, ŵ] = E ρ vw = N n=1 [ v E[v] σ v p n (v n E[v])(w n E[w]) σ v σ w ] w E[w] σ w

Matching Two Pixels Vector v contains pixel information associated with the projected light: v = (L(n)) N n=1. Vector w contains pixel information associated with the IR image: w = (I(n)) N n=1. The entries of these two vectors are non-null, hence: 0 ρ vw 1. ρ vw = 1 v = w. ρ vw = 0 if v and w are uncorrelated. v is associated with IR image pixel (x c, y) w is associated with projector image pixel (x p, y) A match is accepted if ρ vw > t.

Matching Two Images Based on the cross-correlation coefficient it is possible to implement various global image-matching techniques. Because of the epipolar geometry holds and if the images are rectified, the matching can be done row-wise. The most straightforward way to match two corresponding row is to use a sequence-alignment algorithm. Sequence-alignment algorithms are generally based on dynamic programming (DP).

Conclusions To calibrate a Kinect-like camera, it is necessary to have access to the projector image, but this is not the case with Kinect because the manufacturer (Primesense) does not provide this access. The principles behind calibration allow to understand the weaknesses and the current limitations of this type of cameras: limited range, limited to indoor lighting, extremely sensitive to saturated light, glossy surfaces, occlusions, transparent and translucent objects, scattering effects (wool, hair, etc.). Projected-light depth cameras have good resolution and good depth accuracy, both are better than time-of-flight cameras.