Assignment 2: Stereo and 3D Reconstruction from Disparity

Similar documents
Image Rectification (Stereo) (New book: 7.2.1, old book: 11.1)

Lecture 14: Basic Multi-View Geometry

Correspondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]

Multiple View Geometry

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Multi-View Geometry Part II (Ch7 New book. Ch 10/11 old book)

Final Exam Study Guide

Stereo Vision. MAN-522 Computer Vision

Rectification and Disparity

Rectification and Distortion Correction

Computer Vision Lecture 17

Computer Vision Lecture 17

Dense 3D Reconstruction. Christiano Gava

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision

CS4495/6495 Introduction to Computer Vision. 3B-L3 Stereo correspondence

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

CS201 Computer Vision Camera Geometry

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz

CS 231A Computer Vision (Winter 2015) Problem Set 2

CS 231A: Computer Vision (Winter 2018) Problem Set 2

Step-by-Step Model Buidling

Image Transfer Methods. Satya Prakash Mallick Jan 28 th, 2003

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches

Geometry of Multiple views

CSCI 5980/8980: Assignment #4. Fundamental Matrix

Stereo imaging ideal geometry

BIL Computer Vision Apr 16, 2014

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Model-Based Stereo. Chapter Motivation. The modeling system described in Chapter 5 allows the user to create a basic model of a

Dense 3D Reconstruction. Christiano Gava

Computer Vision I. Dense Stereo Correspondences. Anita Sellent 1/15/16

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

Depth from two cameras: stereopsis

Stereo matching. Francesco Isgrò. 3D Reconstruction and Stereo p.1/21

Project 4 Results. Representation. Data. Learning. Zachary, Hung-I, Paul, Emanuel. SIFT and HoG are popular and successful.

Chaplin, Modern Times, 1936

1 Projective Geometry

Announcements. Stereo

Epipolar Geometry and Stereo Vision

Final project bits and pieces

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Depth from two cameras: stereopsis

Rectification. Dr. Gerhard Roth

Lecture 9: Epipolar Geometry

Lecture 6 Stereo Systems Multi-view geometry

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems

3D Reconstruction from Two Views

Announcements. Stereo

Epipolar Geometry and Stereo Vision

Epipolar Geometry and Stereo Vision

Recap from Previous Lecture

Multiple Views Geometry

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

C / 35. C18 Computer Vision. David Murray. dwm/courses/4cv.

Epipolar Geometry and Stereo Vision

Stereo: Disparity and Matching

Lecture 14: Computer Vision

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

CSE152 Introduction to Computer Vision Assignment 3 (SP15) Instructor: Ben Ochoa Maximum Points : 85 Deadline : 11:59 p.m., Friday, 29-May-2015

Image Based Reconstruction II

CS5670: Computer Vision

CV: 3D to 2D mathematics. Perspective transformation; camera calibration; stereo computation; and more

Stereo Image Rectification for Simple Panoramic Image Generation

Two-view geometry Computer Vision Spring 2018, Lecture 10

Computer Vision I. Announcement. Stereo Vision Outline. Stereo II. CSE252A Lecture 15

Stereo Wrap + Motion. Computer Vision I. CSE252A Lecture 17

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

calibrated coordinates Linear transformation pixel coordinates

Structure from Motion and Multi- view Geometry. Last lecture

Epipolar Geometry CSE P576. Dr. Matthew Brown

Camera Drones Lecture 3 3D data generation

Stereo Epipolar Geometry for General Cameras. Sanja Fidler CSC420: Intro to Image Understanding 1 / 33

Comments on Consistent Depth Maps Recovery from a Video Sequence

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

Binocular Stereo Vision. System 6 Introduction Is there a Wedge in this 3D scene?

Introduction to 3D Machine Vision

Structure from motion

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Lecture 6 Stereo Systems Multi- view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-24-Jan-15

3D Geometry and Camera Calibration

Camera Model and Calibration

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Lecture 5 Epipolar Geometry

Stereo Vision Computer Vision (Kris Kitani) Carnegie Mellon University

What have we leaned so far?

Multiple View Geometry

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

CS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

Structure from Motion

Practice Exam Sample Solutions

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Project 2 due today Project 3 out today. Readings Szeliski, Chapter 10 (through 10.5)

LUMS Mine Detector Project

3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava

Transcription:

CS 6320, 3D Computer Vision Spring 2013, Prof. Guido Gerig Assignment 2: Stereo and 3D Reconstruction from Disparity Out: Mon Feb-11-2013 Due: Mon Feb-25-2013, midnight (theoretical and practical parts, submission to CANVAS) TA: Anshul Joshi, MEB 3115 Office hours TA: Mo/We 3pm to 5pm, G. Gerig Mo/We 3pm to 5pm Required Readings: Computer Vision, Forsyth and Ponce, Stereopsis: New book Chapter 7, old book Chapter Slides to chapters as provided on course web-page (http://www.sci.utah.edu/~gerig/cs6320-s2013/cs6320_3d_computer_vision.html) Grading Theoretical problems: These serve as your own study of the material using the textbook and all materials provided on our course homepage. Detailed solutions will be provided. Practical problem: Grading will primarily concern your solution strategy and solution of the camera calibration, and the report that describes your project, your development of the methodology, results, and critical assessments of the results. Late submissions: Late submissions result in 10% deduction for each day. The assignment will no longer be accepted 3 days after the deadline. Honor Code, Misconduct: Please read the School of Computing misconduct information. I. Theoretical Problems Write a report on your solutions for the theoretical problems. This report can be handed in on paper during the class lecture on Monday February 13 or can be added to the electronic pdf/word report of the theoretical part (submitted to the CADE handin system, deadline Monday Feb. 13 midnight). Problem 1: Epipolar Geometry with 3 Cameras We discussed epipolar geometry betwen two cameras, which limits a search for correspondence to an epipolar line. The Java demo (http://www.ai.sri.com/ luong/research/meta3dviewer/epipolargeo.html) also illustrates the epipolar geometry for 3 camera views. Given the fact that for two cameras, the constraint for finding a corresponding point in the right camera for each point in the left camera is a line (i.e. still one degree of freedom), discuiss the situation with 3 cameras. Could it be that with 3 cameras, point to point correspondences are uniquely defined? Why or why not?

Figure 1: Epipolar geometry with 3 camera views. Problem 2: Epiploar geometry and disparity forward translating camera In our course lecture on image rectification (CS6320-CV-F2012-Rectification.pdf), we discussed the epipolar geometry and fundamental matrix of a forward translating camera. Given the definition of the epipolar plane as the plane defined by the two optical centers and a world point, explain why the geometry for a forward moving camera results in a set of radially oriented intersecting planes as shown in the following figure. Given geometric considerations of a forward moving camera, develop a framework to calculate depth of world points from disparity as observed in consecutive frames. Make use of the definition of disparity as presented for the case of a stereo camera setup. Figure 2: Epipolar geometry with 3 camera views. Problem 3: Point correspondence (Problem 7.8 new Textbook, slides stereo). Show that the correlation function reaches its maximum value of 1 when the image brightness of the two windows are related by the affine transform I = λi + µ for some constants λ and µ with λ > 0. I. Practical Problems Pleae note that you have to solve problem 4a but that you can then choose 4B or 4c.

Problem 4: 3D from stereo image pairs This objective of this assignment is to take a stereo pair of images, and explore the epiploar gemeotry and the ability to recover depth from a pair of images. For this assignment, you will need two stereo pairs of your choice, one with arbitrary translation and rotation of the two camera views, and one where you simulate a translated camera by shifting one camera horizontally and by recording the distance of the baseline. 4a: Epipolar geometry from F-matrix We discussed the essential and fundamental matrices and its properties to establish a relationship between corresponding points in the left and right cameras, and relationship of a points and their associated epipolar lines. Following the examples discussed in our slides (CS6320-CV-F2013-chap11-multiple-views-part- II-animated.pptx) anc chapter 7.1 in our textbook, you will explore the following steps: Shoot a stereo-pair of a scene of your choice with your camera, where you take a left and a right picture where you translate and rotate the camera. Specify a set of corresponding pixel landmark pairs in the left and right camera views. Calculate the F-matrix using instructions as discussed in the slides. Choose a point in the left image and calculate the epipolar line in the right image. Display the point and the line as overlays in your images. Do the same for a point in the right image and epipolar line in the elft image. Calculate the position of the epipole of the left camera (see last two slides for instructions). Discuss if the calculated position seems reasonable. 4b: Dense distance map from dense disparity Write an algorithm to search for corresponding points along same scan lines (rectified image pairs). Given optimal correspondence based on the match criteria (here we will used normalized cross correlation), we will calculate disparity and associated depth Z for each pixel. The result is a Z-buffer, or distance image. Given the high computational expense, you are strongly encouraged to choose small size images for this experiment. Please note that images obtained by your camera can be easily subsampled using standard image display programs. We will use a rectified image pair to simplify correspondence search along horizontal lines. You can create two images by simply shifting a camera horizontally along its image plane (e.g. placing a camera against a wall and shifting it horizontally along this wall). Remember to get a precise estimate of the baseline length of this shift. Write a Matlab program to scan two images in a TV-scan like fashion, i.e. a line-byline, row-by-row algorithm where at each location, the nxn neighborhood of each pixel is available for calculations. We need an algorithm for simultaneous calculations on two images with different scan locations.

For each nxn window in the left image, calculate normalized cross-correlation (see course slides) for each window in the right image. Remember that for each window in the left image, you need to search the whole horizontal line for the best matching window. Choose the best corresponding location as the location with the highest correlation. It is suggested to write a version for 3x3 and for 5x5 windows. Please note that the image border cannot be processed, i.e. you need to leave a frame of 1 or 2 pixel width, respectively. Calculate disparity for each corresponding pair of pixels. The disparity in pixel units needs to be converted into mm-units using the known pixel size from camera calibration. Using the disparity, calculate the depth Z at each pixel location. Display the stereo pair and the resulting depth image side by side and discuss your result. A nice way to display depth images is the calculation and display of the slope, i.e. the image gradient. This can be obtained by calculating intensity differences of neighboring pixels (horizonally or verticall) and then display those as an image. Please note that differencing results in negative and positive values, so that double-type image matrices need to be shifted into the positive range of integer images to be displayed properly. 4c: 3D Object geometry via triangulation Using a similar setup as in the previous question (cameras with parallel image planes), take a stereo picture of an object with simple geometry, e.g. a cube (see Fig. 3. Such objects are defined by a small set of landmarks (here 3D corners) and can be reconstructed by displaying a set of lines joining these key points. Define a small set of corresponding key locations by either manual definition of landmarks or a correlation-based image processing method (with selection of only the major key points). Calculate horizontal and vertical disparity in pixel units and mm-units, and from those the 3D point coordinates (X,Y,Z). Use Matlab to display the 3D points and edge lines for the reconstructed object (only those visible in your images). Choose display viewpoints different from the camera views to verify the quality of 3D reconstruction. Discuss your results. 4d: Bonus Problem: Do both 4b and 4c Optional bonus question, for those who have capacity to do more, by solving 4b and 4c.

Figure 3: Two pictures of a simple geometric object (here vertical stereo) and its 3D reconstruction with a wireframe illustration.