Chaplin, Modern Times, 1936

Similar documents
CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Final project bits and pieces

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS4495/6495 Introduction to Computer Vision. 3B-L3 Stereo correspondence

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Stereo: Disparity and Matching

Stereo vision. Many slides adapted from Steve Seitz

Stereo. Many slides adapted from Steve Seitz

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?

Ninio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29,

BIL Computer Vision Apr 16, 2014

Project 3 code & artifact due Tuesday Final project proposals due noon Wed (by ) Readings Szeliski, Chapter 10 (through 10.5)

Epipolar Geometry and Stereo Vision

Recap from Previous Lecture

Epipolar Geometry and Stereo Vision

CS5670: Computer Vision

Computer Vision Lecture 17

Computer Vision Lecture 17

What have we leaned so far?

Lecture 10: Multi view geometry

Image Based Reconstruction II

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems

Introduction à la vision artificielle X

Miniature faking. In close-up photo, the depth of field is limited.

Project 2 due today Project 3 out today. Readings Szeliski, Chapter 10 (through 10.5)

Multi-stable Perception. Necker Cube

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

Lecture 10: Multi-view geometry

Multiple View Geometry

Epipolar Geometry and Stereo Vision

Computer Vision I. Announcements. Random Dot Stereograms. Stereo III. CSE252A Lecture 16

CEng Computational Vision

Epipolar Geometry and Stereo Vision

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

Dense 3D Reconstruction. Christiano Gava

Lecture 9 & 10: Stereo Vision

Step-by-Step Model Buidling

CS 2770: Intro to Computer Vision. Multiple Views. Prof. Adriana Kovashka University of Pittsburgh March 14, 2017

Multiple View Geometry

Dense 3D Reconstruction. Christiano Gava

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Lecture 14: Computer Vision

Multi-view stereo. Many slides adapted from S. Seitz

Stereo and Epipolar geometry

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview

Stereo Epipolar Geometry for General Cameras. Sanja Fidler CSC420: Intro to Image Understanding 1 / 33

A virtual tour of free viewpoint rendering

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

EE795: Computer Vision and Intelligent Systems

Homographies and RANSAC

6.819 / 6.869: Advances in Computer Vision Antonio Torralba and Bill Freeman. Lecture 11 Geometry, Camera Calibration, and Stereo.

Lecture 14: Basic Multi-View Geometry

Stereo Matching. Stereo Matching. Face modeling. Z-keying: mix live and synthetic

Announcements. Stereo Vision Wrapup & Intro Recognition

Image Rectification (Stereo) (New book: 7.2.1, old book: 11.1)

Computer Vision I. Announcement. Stereo Vision Outline. Stereo II. CSE252A Lecture 15

Introduction to Computer Vision. Week 10, Winter 2010 Instructor: Prof. Ko Nishino

Stereo Matching.

Stereo Wrap + Motion. Computer Vision I. CSE252A Lecture 17

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

CS5670: Computer Vision

Mosaics wrapup & Stereo

Lecture'9'&'10:'' Stereo'Vision'

Cameras and Stereo CSE 455. Linda Shapiro

3D Photography: Active Ranging, Structured Light, ICP

3D Photography: Stereo

Epipolar Geometry CSE P576. Dr. Matthew Brown

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --

Finally: Motion and tracking. Motion 4/20/2011. CS 376 Lecture 24 Motion 1. Video. Uses of motion. Motion parallax. Motion field

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

CS 4495/7495 Computer Vision Frank Dellaert, Fall 07. Dense Stereo Some Slides by Forsyth & Ponce, Jim Rehg, Sing Bing Kang

3D Fusion of Infrared Images with Dense RGB Reconstruction from Multiple Views - with Application to Fire-fighting Robots

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

Last time: Disparity. Lecture 11: Stereo II. Last time: Triangulation. Last time: Multi-view geometry. Last time: Epipolar geometry

Final Exam Study Guide

Advanced Digital Photography and Geometry Capture. Visual Imaging in the Electronic Age Lecture #10 Donald P. Greenberg September 24, 2015

Advanced Digital Photography and Geometry Capture. Visual Imaging in the Electronic Age Lecture #10 Donald P. Greenberg September 24, 2015

Kinect Device. How the Kinect Works. Kinect Device. What the Kinect does 4/27/16. Subhransu Maji Slides credit: Derek Hoiem, University of Illinois

arxiv: v1 [cs.cv] 28 Sep 2018

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

CS 4495 Computer Vision Motion and Optic Flow

Three-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo --

Structure from motion

Computer Vision, Lecture 11

Notes 9: Optical Flow

Peripheral drift illusion

Structure from motion

EE795: Computer Vision and Intelligent Systems

Computer Vision I. Announcement

Instance-level recognition

Photo by Carl Warner

International Journal for Research in Applied Science & Engineering Technology (IJRASET) A Review: 3D Image Reconstruction From Multiple Images

Structure from Motion

Stereo Vision. MAN-522 Computer Vision

On-line and Off-line 3D Reconstruction for Crisis Management Applications

Transcription:

Chaplin, Modern Times, 1936

[A Bucket of Water and a Glass Matte: Special Effects in Modern Times; bonus feature on The Criterion Collection set]

Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates of that point? Camera 1 Camera 2 R 1,t 1 R 2,t 2 Camera 3 R 3,t 3 Slide credit: Noah Snavely

Multi-view geometry problems Motion: Given a set of corresponding points in two or more images, compute the camera parameters? R 1,t 1 R 2,t 2 R 3,t 3 Camera 1 Camera 2 Camera 3?? Slide credit: Noah Snavely

Multi-view geometry problems Stereo correspondence: Given a point in one of the images, where could its corresponding points be in the other images? Camera 1 Camera 2 Camera 3 R 1,t 1 R 2,t 2 R 3,t 3 Slide credit: Noah Snavely

Multi-view geometry problems Optical flow: Given two images, find the location of a world point in a second close-by image with no camera info. Camera 1 Camera 2

Fundamental matrix Let x be a point in left image, x in right image l l x x Epipolar relation x maps to epipolar line l x maps to epipolar line l Epipolar mapping described by a 3x3 matrix F: l = Fx l = F T x It follows that: x Fx = 0

Fundamental matrix This matrix F is called the Essential Matrix when image intrinsic parameters are known the Fundamental Matrix more generally (uncalibrated case) Can solve for F from point correspondences Each (x, x ) pair gives one linear equation in entries of F x Fx = 0 F has 9 entries, but really only 7 degrees of freedom. With 8 points it is simple to solve for F, but it is also possible with 7. See Marc Pollefey s notes for a nice tutorial

Stereo image rectification

Stereo image rectification Reproject image planes onto a common plane parallel to the line between camera centers Pixel motion is horizontal after this transformation Two homographies (3x3 transform), one for each input image reprojection C. Loop and Z. Zhang. Computing Rectifying Homographies for Stereo Vision. IEEE Conf. Computer Vision and Pattern Recognition, 1999.

Rectification example

CS 4495 Computer Vision A. Bobick Correspondence problem Motion and Optic Flow Multiple match hypotheses satisfy epipolar constraint, but which is correct? Figure from Gee & Cipolla 1999

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Dense correspondence search For each epipolar line: Adapted from Li Zhang For each pixel / window in the left image: Compare with every pixel / window on same epipolar line in right image Pick position with minimum match cost (e.g., SSD, normalized correlation)

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Correspondence problem Intensity profiles Source: Andrew Zisserman

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Correspondence problem Neighborhoods of corresponding points are similar in intensity patterns. Source: Andrew Zisserman

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Correlation-based window matching

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Correlation-based window matching

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Correlation-based window matching

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Correlation-based window matching

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Correlation-based window matching??? Textureless regions are non-distinct; high ambiguity for matches.

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Effect of window size W = 3 W = 20 Want window large enough to have sufficient intensity variation, yet small enough to contain only pixels with about the same disparity. Figures from Li Zhang

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Problem: Occlusion Uniqueness says up to match per pixel When is there no match? Occluded pixels

CS 4495 Computer Vision A. Bobick Disparity gradient constraint Motion and Optic Flow Assume piecewise continuous surface, so want disparity estimates to be locally smooth Figure from Gee & Cipolla 1999

CS 4495 Computer Vision A. Bobick Ordering constraint Motion and Optic Flow Points on same surface (opaque object) will be in same order in both views Figure from Gee & Cipolla 1999

CS 4495 Computer Vision A. Bobick Ordering constraint Motion and Optic Flow Won t always hold, e.g. consider transparent object, or an occluding surface Figures from Forsyth & Ponce

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Stereo Tsukuba test scene (now old)

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Results with window search Window-based matching (best window size) Ground truth

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Better solutions Beyond individual correspondences to estimate disparities: Optimize correspondence assignments jointly Scanline at a time (DP) Full 2D grid (graph cuts)

intensity CS 4495 Computer Vision A. Bobick Motion and Optic Flow Scanline stereo Try to coherently match pixels on the entire scanline Different scanlines are still optimized independently Left image Right image

Left occlusion CS 4495 Computer Vision A. Bobick Motion and Optic Flow Shortest paths for scan-line stereo Right occlusion Left image I Right image I S left one-to-one q t Left occlusion Right occlusion s p S right Can be implemented with dynamic programming Ohta & Kanade 85, Cox et al. 96, Intille & Bobick, 01 Slide credit: Y. Boykov

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Coherent stereo on 2D grid Scanline stereo generates streaking artifacts Can t use dynamic programming to find spatially coherent disparities/ correspondences on a 2D grid

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Stereo as energy minimization What defines a good stereo correspondence? 1. Match quality Want each pixel to find a good match in the other image 2. Smoothness If two pixels are adjacent, they should (usually) move about the same amount

CS 4495 Computer Vision A. Bobick Stereo matching as energy minimization I 1 I 2 D Motion and Optic Flow W 1 (i) W 2 (i+d(i)) D(i) E Edata ( I1, I2, D) Esmooth ( D) E W ( i) W ( i D( i 2 data 1 2 )) i Esmooth D( i) D( j) neighborsi, j Energy functions of this form can be minimized using graph cuts. Y. Boykov, O. Veksler, and R. Zabih, Fast Approximate Energy Minimization via Graph Cuts, PAMI 2001 Source: Steve Seitz

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Better results Graph cut method Boykov et al., Fast Approximate Energy Minimization via Graph Cuts, International Conference on Computer Vision, September 1999. Ground truth For the latest and greatest: http://www.middlebury.edu/stereo/

CS 4495 Computer Vision A. Bobick Motion and Optic Flow Challenges Low-contrast textureless image regions Occlusions Violations of brightness constancy Specular reflections Really large baselines Foreshortening and appearance change Camera calibration errors

SIFT + Fundamental Matrix + RANSAC + Sparse correspondence

SIFT + Fundamental Matrix + RANSAC + dense correspondence Building Rome in a Day By Sameer Agarwal, Yasutaka Furukawa, Noah Snavely, Ian Simon, Brian Curless, Steven M. Seitz, Richard Szeliski Communications of the ACM, Vol. 54 No. 10, Pages 105-112 2009

SIFT + Fundamental Matrix + RANSAC + dense correspondence Building Rome in a Day By Sameer Agarwal, Yasutaka Furukawa, Noah Snavely, Ian Simon, Brian Curless, Steven M. Seitz, Richard Szeliski Communications of the ACM, Vol. 54 No. 10, Pages 105-112

SIFT + Fundamental Matrix + RANSAC + dense correspondence

Once I have my depth map, what can I do with it? Measure. Combine! (Reorganize?)

What if we want to align but we have no matched pairs? Hough transform and RANSAC not applicable (t x, t y ) Problem: no initial guesses for correspondence

Important applications Medical imaging: match brain scans or contours Robotics: match point clouds Kwok and Tang

ICP demonstration Bouaziz et al.

Iterative Closest Points (ICP) Algorithm Goal: Estimate transform between two dense point sets S 1 and S 2 1. Initialize transformation Compute difference in mean positions, subtract Compute difference in scales, normalize 2. Assign each point in S 1 to its nearest neighbor in S 2 3. Estimate transformation parameters T Least squares or robust least squares, e.g., rigid transform 4. Transform the points in S 1 using estimated parameters T 5. Repeat steps 2-4 until change is very small (convergence)

Example: solving for translation (t x, t y ) Problem: no initial guesses for correspondence ICP solution 1. Initialize t by mean point translation 2. Find nearest neighbors for each point 3. Compute transform using matches 4. Move points using transform 5. Repeat steps 2-4 until convergence x y B i B i x y A i A i t t x y

Example: aligning boundaries 1. Extract edge pixels p 1.. p n and q 1.. q m 2. Compute initial transformation (e.g., compute translation and scaling by center of mass, variance within each image) 3. Get nearest neighbors: for each point p i find corresponding match(i) = argmin dist(pi, qj) j 4. Compute transformation T based on matches 5. Transform points p according to T 6. Repeat 3-5 until convergence p q

ICP demonstration Bouaziz et al.

Stereo correspondence Let x be a point in left image, x in right image Epipolar relation x maps to epipolar line l x maps to epipolar line l l x x l

How does a depth camera work? Microsoft Kinect v1 Intel laptop depth camera

Active stereo with structured light Project structured light patterns onto the object Simplifies the correspondence problem Allows us to use only one camera camera Stereo system! projector L. Zhang, B. Curless, and S. M. Seitz. Rapid Shape Acquisition Using Color Structured Light and Multi-pass Dynamic Programming. 3DPVT 2002

Kinect: Structured infrared light http://bbzippo.wordpress.com/2010/11/28/kinect-in-infrared/

How does a depth camera work? Stereo in infrared.

Time of Flight (Kinect V2) Depth cameras in HoloLens use time of flight SONAR for light Emit light of a known wavelength, and time how long it takes for it to come back

With either technique I gain depth maps over time. Optex Depth Camera Based on Canesta Solution