Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Similar documents
Project 2 due today Project 3 out today. Readings Szeliski, Chapter 10 (through 10.5)

Recap from Previous Lecture

Project 3 code & artifact due Tuesday Final project proposals due noon Wed (by ) Readings Szeliski, Chapter 10 (through 10.5)

CS5670: Computer Vision

Stereo: Disparity and Matching

What have we leaned so far?

Stereo Matching.

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?

CS4495/6495 Introduction to Computer Vision. 3B-L3 Stereo correspondence

Introduction to Computer Vision. Week 10, Winter 2010 Instructor: Prof. Ko Nishino

Stereo and Epipolar geometry

Stereo vision. Many slides adapted from Steve Seitz

Epipolar Constraint. Epipolar Lines. Epipolar Geometry. Another look (with math).

Multiple View Geometry

Chaplin, Modern Times, 1936

Correspondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]

Computer Vision I. Announcements. Random Dot Stereograms. Stereo III. CSE252A Lecture 16

Lecture 14: Computer Vision

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision

Announcements. Stereo Vision Wrapup & Intro Recognition

Final project bits and pieces

Stereo. Many slides adapted from Steve Seitz

Image Based Reconstruction II

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Epipolar Geometry and Stereo Vision

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

BIL Computer Vision Apr 16, 2014

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Epipolar Geometry and Stereo Vision

Computer Vision I. Announcement. Stereo Vision Outline. Stereo II. CSE252A Lecture 15

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F

Computer Vision Lecture 17

Computer Vision Lecture 17

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

Lecture 9 & 10: Stereo Vision

Multiple View Geometry

Introduction à la vision artificielle X

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

EE795: Computer Vision and Intelligent Systems

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava

Stereo Matching. Stereo Matching. Face modeling. Z-keying: mix live and synthetic

Lecture'9'&'10:'' Stereo'Vision'

3D photography. Digital Visual Effects, Spring 2007 Yung-Yu Chuang 2007/5/15

Computer Vision I. Announcement

Passive 3D Photography

Announcements. Stereo Vision II. Midterm. Example: Helmholtz Stereo Depth + Normals + BRDF. Stereo

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems

6.819 / 6.869: Advances in Computer Vision Antonio Torralba and Bill Freeman. Lecture 11 Geometry, Camera Calibration, and Stereo.

Model-Based Stereo. Chapter Motivation. The modeling system described in Chapter 5 allows the user to create a basic model of a

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview

Cameras and Stereo CSE 455. Linda Shapiro

Image Rectification (Stereo) (New book: 7.2.1, old book: 11.1)

Final Review CMSC 733 Fall 2014

Project 4 Results. Representation. Data. Learning. Zachary, Hung-I, Paul, Emanuel. SIFT and HoG are popular and successful.

CS5670: Computer Vision

A virtual tour of free viewpoint rendering

Lecture 14: Basic Multi-View Geometry

Lecture 10: Multi view geometry

Epipolar Geometry and Stereo Vision

Lecture 6 Stereo Systems Multi-view geometry

Miniature faking. In close-up photo, the depth of field is limited.

Stereo and structured light

Kinect Device. How the Kinect Works. Kinect Device. What the Kinect does 4/27/16. Subhransu Maji Slides credit: Derek Hoiem, University of Illinois

Lecture 10: Multi-view geometry

Multi-view stereo. Many slides adapted from S. Seitz

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Stereo Vision. MAN-522 Computer Vision

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

Passive 3D Photography

7. The Geometry of Multi Views. Computer Engineering, i Sejong University. Dongil Han

Capturing, Modeling, Rendering 3D Structures

Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza

Mosaics wrapup & Stereo

Computer Vision, Lecture 11

Geometric Reconstruction Dense reconstruction of scene geometry

CS6670: Computer Vision

Stereo imaging ideal geometry

Depth from two cameras: stereopsis

Step-by-Step Model Buidling

Stereo Vision A simple system. Dr. Gerhard Roth Winter 2012

Comparison of Graph Cuts with Belief Propagation for Stereo, using Identical MRF Parameters

Finally: Motion and tracking. Motion 4/20/2011. CS 376 Lecture 24 Motion 1. Video. Uses of motion. Motion parallax. Motion field

Epipolar Geometry and Stereo Vision

1 (5 max) 2 (10 max) 3 (20 max) 4 (30 max) 5 (10 max) 6 (15 extra max) total (75 max + 15 extra)

Depth from two cameras: stereopsis

Segmentation Based Stereo. Michael Bleyer LVA Stereo Vision

Stereo Vision II: Dense Stereo Matching

Stereo Wrap + Motion. Computer Vision I. CSE252A Lecture 17

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION

Stereo Vision Computer Vision (Kris Kitani) Carnegie Mellon University

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

Lecture 10 Dense 3D Reconstruction

A Comparative Study of Stereovision Algorithms

EE795: Computer Vision and Intelligent Systems

Transcription:

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Teesta suspension bridge-darjeeling, India

Mark Twain at Pool Table", no date, UCR Museum of Photography

Woman getting eye exam during immigration procedure at Ellis Island, c. 1905-1920, UCR Museum of Phography

Announcements Problem set 5 is on the class web page. (I made minor changes today).

Information on Stereo Trucco and Verri Read Chapter 7 from the start through 7.3.2, also 7.3.7 and 7.4, 7.4.1. The rest is optional. Many slides taken from Octavia Camps and Steve Seitz

Main Points Cameras with known position. Stereo allows depth by triangulation Two parts: Finding corresponding points. Computing depth (easy part). Constraints: Geometry, epipolar constraint. Photometric: Brightness constancy, only partly true. Ordering: only partly true. Smoothness of objects: only partly true.

Matching Cost function: What you compare: points, regions, features. How you compare: eg., SSD, correlation. How you optimize. Local greedy matches. 1D search. 2D search.

Why Stereo Vision? 2D images project 3D points into 2D: P Q P =Q 3D Points on the same viewing line have the same 2D image: 2D imaging results in depth information loss (Camps) O

Stereo Assumes (two) cameras. Known positions. Recover depth.

Recovering Depth Information: P Q P 1 Q 1 P 2 =Q 2 O 2 O 1 Depth can be recovered with two images and triangulation. (Camps)

So Stereo has two steps Finding matching points in the images Then using them to compute depth.

Epipolar Constraint Most powerful correspondence constraint. Simplifies discussion of depth recovery.

Stereo correspondence Determine Pixel Correspondence Pairs of points that correspond to same scene point epipolar line epipolar plane epipolar line Epipolar Constraint Reduces correspondence problem to 1D search along conjugate epipolar lines (Seitz)

Simplest Case Image planes of cameras are parallel. Focal points are at same height. Focal lengths same. Then, epipolar lines are horizontal scan lines. blackboard

We can always achieve this geometry with image rectification Image Reprojection reproject image planes onto common plane parallel to line between optical centers Notice, only focal point of camera really matters (Seitz)

Let s discuss reconstruction with this geometry before correspondence, because it s much easier. blackboard P x l Z x r T + x Z x f r l = T Z Disparity: d T Z = f x l x r = x l x r f p l p r O l O r T Then given Z, we can compute X T is the stereo baseline and Y. d measures the difference in retinal position between corresponding points x l and x r give the x coordinate of corresponding points in the left and right images. (Camps)

Correspondence: What should we match? Objects? Regions? Edges? Pixels? Collections of pixels?

Julesz: had huge impact because it showed that recognition not needed for stereo.

Correspondence: Epipolar constraint.

Correspondence: Photometric constraint Same world point has same intensity in both images. Lambertian fronto-parallel Issues: Noise Specularity Foreshortening

Using these constraints we can use matching for stereo For each epipolar line For each pixel in the left image compare with every pixel on same epipolar line in right image pick pixel with minimum match cost This will never work, so: Improvement: match windows (Seitz)

Comparing Windows: =? f g Most popular For each window, match to closest window on epipolar line in other image. (Camps)

Window size Effect of window size (Seitz) W = 3 W = 20 Better results with adaptive window T. Kanade and M. Okutomi, A Stereo Matching Algorithm with an Adaptive Window: Theory and Experiment,, Proc. International Conference on Robotics and Automation, 1991. D. Scharstein and R. Szeliski. Stereo matching with nonlinear diffusion. International Journal of Computer Vision, 28(2):155-174, July 1998

Stereo results Data from University of Tsukuba Scene Ground truth (Seitz)

Results with window correlation Window-based matching (best window size) (Seitz) Ground truth

Ordering constraint Usually, order of points in two images is same. blackboard Uniqueness One pixel cannot match more than one pixel. Occlusions This means some points must go unmatched

This enables Dynamic Programming If we match pixel i in image 1 to pixel j in image 2, no matches that follow will affect which are the best preceding matches. Example with pixels.

First of all, we can represent a matching with a disparity map. Since disparity is non-negative, we ll use -1 to indicate an occlusion. So a 1D disparity map for the left image could be: [- 1 1 1-1 2 2 0] This means the first pixel is occluded, the second has a disparity of 1, etc. Notice that whenever there is an occlusion, the disparity will generally increase by one because we are advancing one pixel in the left image, without advancing in the right image (unless there s been an occlusion at the same time in the right image). When the disparity decreases, this means there s been an occlusion in the right image.

Next, given two images and a disparity map, we can assign a cost to this hypothesized matching. There are many ways to do this, but let s look at a simple example. When we match two pixels, the cost is the square of the difference in their intensities. For every occluded pixel, we assign a fixed cost. (In the problem set, we scale intensities to range from 0 to 1 and use an occlusion cost of.01. See Problem Set 7 for notes on how to find the disparity map with lowest cost using dynamic programming.

Other constraints Smoothness: disparity usually doesn t change too quickly. Unfortunately, this makes the problem 2D again. Solved with a host of graph algorithms, Markov Random Fields, Belief Propagation,. Occlusion and disparity are connected.

Correspondence with MRF Every pixel is a site. Label of a pixel is its disparity. Disparity implies two pixels match. Prob. depends on similarity of pixels. Disparity at one pixel related to others since nearby pixels have similar disparities. Penalty based on different disparities at neighboring pixels. Finding best labeling is NP-hard, but good local algorithms exist.

α β swap

Min-Cut gives best swap Min-cut requires edge to one label be cut. Cut between neighbors w/ diff. labels. Link to each label is cost of applying that label; cut means label is applied. Link between pixels = neighborhood cost (0 when same label). Keep performing swaps with all pairs of disparities until convergence.

Results with graph cuts Graph Cuts Boykov et al., Fast Approximate Energy Minimization via Graph Cuts, International Conference on Computer Vision, September 1999. (Seitz) Ground truth

Summary First, we understand constraints that make the problem solvable. Some are hard, like epipolar constraint. Ordering isn t a hard constraint, but most useful when treated like one. Some are soft, like pixel intensities are similar, disparities usually change slowly. Then we find optimization method. Which ones we can use depend on which constraints we pick.