Miniature faking. In close-up photo, the depth of field is limited.

Similar documents
Project 4 Results. Representation. Data. Learning. Zachary, Hung-I, Paul, Emanuel. SIFT and HoG are popular and successful.

Think-Pair-Share. What visual or physiological cues help us to perceive 3D shape and depth?

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Stereo: Disparity and Matching

Computer Vision Lecture 17

Computer Vision Lecture 17

Mosaics wrapup & Stereo

Multi-View Geometry (Ch7 New book. Ch 10/11 old book)

Epipolar Geometry and Stereo Vision

Chaplin, Modern Times, 1936

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

Epipolar Geometry and Stereo Vision

Photo by Carl Warner

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision

Multiple View Geometry

Why is computer vision difficult?

Camera Geometry II. COS 429 Princeton University

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?

Basic distinctions. Definitions. Epstein (1965) familiar size experiment. Distance, depth, and 3D shape cues. Distance, depth, and 3D shape cues

6.819 / 6.869: Advances in Computer Vision Antonio Torralba and Bill Freeman. Lecture 11 Geometry, Camera Calibration, and Stereo.

Undergrad HTAs / TAs. Help me make the course better! HTA deadline today (! sorry) TA deadline March 21 st, opens March 15th

Image correspondences and structure from motion

10/03/11. Model Fitting. Computer Vision CS 143, Brown. James Hays. Slides from Silvio Savarese, Svetlana Lazebnik, and Derek Hoiem

Geometric camera models and calibration

Stereo and Epipolar geometry

CS 2770: Intro to Computer Vision. Multiple Views. Prof. Adriana Kovashka University of Pittsburgh March 14, 2017

Homographies and RANSAC

Multi-stable Perception. Necker Cube

Epipolar Geometry and Stereo Vision

Final project bits and pieces

All human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them.

Multi-stable Perception. Necker Cube

Recovering structure from a single view Pinhole perspective projection

Dense 3D Reconstruction. Christiano Gava

Local features: detection and description. Local invariant features

Dense 3D Reconstruction. Christiano Gava

CS5670: Computer Vision

Announcements. Stereo

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Stereo Vision A simple system. Dr. Gerhard Roth Winter 2012

CSE 4392/5369. Dr. Gian Luca Mariottini, Ph.D.

arxiv: v1 [cs.cv] 28 Sep 2018

Epipolar Geometry and Stereo Vision

What have we leaned so far?

Stereo vision. Many slides adapted from Steve Seitz

Announcements. Stereo

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Lecture 10: Multi-view geometry

Image warping and stitching

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Image warping and stitching

Computer Vision I. Announcements. Random Dot Stereograms. Stereo III. CSE252A Lecture 16

Lecture 10: Multi view geometry

Robert Collins CSE486, Penn State Lecture 08: Introduction to Stereo

Local features: detection and description May 12 th, 2015

Step-by-Step Model Buidling

Camera Calibration. COS 429 Princeton University

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

BIL Computer Vision Apr 16, 2014

Final Exam Study Guide

Project 3 code & artifact due Tuesday Final project proposals due noon Wed (by ) Readings Szeliski, Chapter 10 (through 10.5)

Stereo Epipolar Geometry for General Cameras. Sanja Fidler CSC420: Intro to Image Understanding 1 / 33

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Finally: Motion and tracking. Motion 4/20/2011. CS 376 Lecture 24 Motion 1. Video. Uses of motion. Motion parallax. Motion field

EE795: Computer Vision and Intelligent Systems

Project 2 due today Project 3 out today. Readings Szeliski, Chapter 10 (through 10.5)

Image warping and stitching

Image Rectification (Stereo) (New book: 7.2.1, old book: 11.1)

Mosaics. Today s Readings

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability

CS5670: Computer Vision

Lecture 14: Computer Vision

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F

Projective Geometry and Camera Models

calibrated coordinates Linear transformation pixel coordinates

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Structure from Motion

Stereo. Many slides adapted from Steve Seitz

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Cameras and Stereo CSE 455. Linda Shapiro

Depth from two cameras: stereopsis

Multiple View Geometry

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Binocular cues to depth PSY 310 Greg Francis. Lecture 21. Depth perception

Instance-level recognition part 2

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,

Instance-level recognition I. - Camera geometry and image alignment

Instance-level recognition II.

Depth from two cameras: stereopsis

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Structure from Motion

Local features and image matching. Prof. Xin Yang HUST

A Systems View of Large- Scale 3D Reconstruction

Spatial track: range acquisition modeling

Feature Matching and Robust Fitting

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

arxiv: v1 [cs.cv] 28 Sep 2018

Transcription:

Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg

Miniature faking

Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg

Review Previous section: Model fitting and outlier rejection

Review: Hough transform y m x b y m 3 5 3 3 2 2 3 7 11 10 4 3 2 3 1 4 5 2 2 1 0 1 3 3 Slide from S. Savarese x b

Review: RANSAC Algorithm: N I 14 1. Sample (randomly) the number of points required to fit the model (#=2) 2. Solve for model parameters using samples 3. Score by the fraction of inliers within a preset threshold of the model Repeat 1-3 until the best model is found with high confidence

Review: 2D image transformations Szeliski 2.1

What if you want to align but have no prior matched pairs? Hough transform and RANSAC not applicable Important applications Medical imaging: match brain scans or contours Robotics: match point clouds

Iterative Closest Points (ICP) Algorithm Goal: estimate transform between two dense sets of points 1. Initialize transformation (e.g., compute difference in means and scale) 2. Assign each point in {Set 1} to its nearest neighbor in {Set 2} 3. Estimate transformation parameters e.g., least squares or robust least squares 4. Transform the points in {Set 1} using estimated parameters 5. Repeat steps 2-4 until change is very small

Example: aligning boundaries 1. Extract edge pixels p 1.. p n and q 1.. q m 2. Compute initial transformation (e.g., compute translation and scaling by center of mass, variance within each image) 3. Get nearest neighbors: for each point p i find corresponding match(i) = argmin dist(pi, qj) j 4. Compute transformation T based on matches 5. Warp points p according to T 6. Repeat 3-5 until convergence p q

Example: solving for translation (t x, t y ) Problem: no initial guesses for correspondence y x A i A i B i B i t t y x y x ICP solution 1. Find nearest neighbors for each point 2. Compute transform using matches 3. Move points using transform 4. Repeat steps 1-3 until convergence

Algorithm Summaries Least Squares Fit closed form solution robust to noise not robust to outliers Robust Least Squares improves robustness to outliers requires iterative optimization Hough transform robust to noise and outliers can fit multiple models only works for a few parameters (1-4 typically) RANSAC robust to noise and outliers works with a moderate number of parameters (e.g, 1-8) Iterative Closest Point (ICP) For local alignment only: does not require initial correspondences

This section multiple views Today Intro to multiple views and Stereo Monday Camera calibration Wednesday Fundamental Matrix Friday Optical Flow

Oriented and Translated Camera R j w t k w O w i w

Degrees of freedom X t x K R 1 1 0 0 0 1 33 32 31 23 22 21 13 12 11 0 0 z y x t r r r t r r r t r r r v u s v u w z y x 5 6

How to calibrate the camera? 1 * * * * * * * * * * * * Z Y X s sv su X t x K R

Stereo: Intro Computer Vision James Hays Slides by Kristen Grauman

Multiple views Hartley and Zisserman stereo vision Lowe structure from motion optical flow

Why multiple views? Structure and depth are inherently ambiguous from single views. Images from Lana Lazebnik

Why multiple views? Structure and depth are inherently ambiguous from single views. P1 P2 P1 =P2 Optical center

What cues help us to perceive 3d shape and depth?

[Figure from Prados & Faugeras 2006] Shading

Focus/defocus Images from same point of view, different camera parameters 3d shape / depth estimates [figs from H. Jin and P. Favaro, 2002]

Texture [From A.M. Loh. The recovery of 3-D structure using visual texture patterns. PhD thesis]

Perspective effects Image credit: S. Seitz

Motion Figures from L. Zhang http://www.brainconnection.com/teasers/?main=illusion/motion-shape

Occlusion Rene Magritt'e famous painting Le Blanc- Seing (literal translation: "The Blank Signature") roughly translates as "free hand. 1965

If stereo were critical for depth perception, navigation, recognition, etc., then this would be a problem

Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates of that point? Camera 1 Camera 2 R 1,t 1 R 2,t 2 Camera 3 R 3,t 3 Slide credit: Noah Snavely

Multi-view geometry problems Stereo correspondence: Given a point in one of the images, where could its corresponding points be in the other images? Camera 1 Camera 2 Camera 3 R 1,t 1 R 2,t 2 R 3,t 3 Slide credit: Noah Snavely

Multi-view geometry problems Motion: Given a set of corresponding points in two or more images, compute the camera parameters? R 1,t 1 R 2,t 2 R 3,t 3 Camera 1 Camera 2 Camera 3?? Slide credit: Noah Snavely

Human eye Rough analogy with human visual system: Pupil/Iris control amount of light passing through lens Retina - contains sensor cells, where image is formed Fovea highest concentration of cones Fig from Shapiro and Stockman

Human stereopsis: disparity Human eyes fixate on point in space rotate so that corresponding images form in centers of fovea.

Human stereopsis: disparity Disparity occurs when eyes fixate on one object; others appear at different visual angles

Stereo photography and stereo viewers Take two pictures of the same subject from two slightly different viewpoints and display so that each eye sees only one of the images. Invented by Sir Charles Wheatstone, 1838 Image from fisher-price.com

http://www.johnsonshawmuseum.org

http://www.johnsonshawmuseum.org

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

http://www.well.com/~jimg/stereo/stereo_list.html

Autostereograms Exploit disparity as depth cue using single image. (Single image random dot stereogram, Single image stereogram) Images from magiceye.com

Images from magiceye.com Autostereograms

Estimating depth with stereo Stereo: shape from motion between two views We ll need to consider: Info on camera pose ( calibration ) Image point correspondences scene point optical center image plane

Stereo vision Two cameras, simultaneous views Single moving camera and static scene

Camera parameters Camera frame 2 Extrinsic parameters: Camera frame 1 Camera frame 2 Camera frame 1 Intrinsic parameters: Image coordinates relative to camera Pixel coordinates Extrinsic params: rotation matrix and translation vector Intrinsic params: focal length, pixel sizes (mm), image center point, radial distortion parameters We ll assume for now that these parameters are given and fixed.

Geometry for a simple stereo system First, assuming parallel optical axes, known camera parameters (i.e., calibrated cameras):

World point image point (left) Depth of p image point (right) Focal length optical center (left) baseline optical center (right)

Geometry for a simple stereo system Assume parallel optical axes, known camera parameters (i.e., calibrated cameras). What is expression for Z? Similar triangles (p l, P, p r ) and (O l, P, O r ): T x l Z f x r T Z disparity Z f T x r x l

Depth from disparity image I(x,y) Disparity map D(x,y) image I (x,y ) (x,y )=(x+d(x,y), y) So if we could find the corresponding points in two images, we could estimate relative depth

Where do we need to search? To be continued