6.819 / 6.869: Advances in Computer Vision Antonio Torralba and Bill Freeman. Lecture 11 Geometry, Camera Calibration, and Stereo.
|
|
- Jonah Goodwin
- 5 years ago
- Views:
Transcription
1 6.819 / 6.869: Advances in Computer Vision Antonio Torralba and Bill Freeman Lecture 11 Geometry, Camera Calibration, and Stereo.
2 2d from 3d; 3d from multiple 2d measurements? 2d 3d?
3 Perspective projection f f (0,0,0) image plane Virtual image plane
4 Perspective projection Y y? (0,0,0) f Virtual image plane Similar triangles: y / f = Y / Z y = f Y/Z Perspective projection: ( X,Y,Z ) f X Z, f Y Z Z
5 Vanishing points graphics/two_point_perspective.html
6 Other projection models: Orthographic projection ( x, y, z) ( x, y)
7 Three camera projections (1) Perspective: (2) Weak perspective: (3) Orthographic: ), ( ),, (, ),, (, ),, ( 0 0 y x z y x z fy z fx z y x z fy z fx z y x 3-d point 2-d image position
8 Three camera projections Perspective projection Parallel (orthographic) projection Weak perspective?
9 Homogeneous coordinates Is the perspective projection a linear transformation? no division by z is nonlinear Trick: add one more coordinate: homogeneous image coordinates homogeneous world coordinates Converting from homogeneous coordinates Slide by Steve Seitz
10 Perspective Projection Projection is a matrix multiply using homogeneous coordinates: This is known as perspective projection The matrix is the projection matrix Slide by Steve Seitz / f 0 x y z 1 = x y z / f f x z, f y z
11 Perspective Projection How does scaling the projection matrix change the transformation? Slide by Steve Seitz / f 0 x y z 1 = x y z / f f x z, f y z f f x y z 1 = fx fy z f x z, f y z
12 Orthographic Projection Special case of perspective projection Image World Also called parallel projection What s the projection matrix?? Slide by Steve Seitz
13 Orthographic Projection Special case of perspective projection Distance from the COP to the PP is infinite Image World Also called parallel projection What s the projection matrix? Slide by Steve Seitz
14 2D Transformations
15 2D Transformations Example: translation = + tx ty
16 2D Transformations Example: translation = + tx ty = 1 0 tx 0 1 ty. 1
17 2D Transformations Example: translation written 3 ways: non-homog homog in, non-h out, homog in, homog out = + tx ty = 1 0 tx 0 1 ty. 1 = 1 0 tx 0 1 ty Now we can chain transformations
18 Translation and rotation, written in each set of coordinates Non-homogeneous coordinates! B p = B R A! p + B! t A A from Homogeneous coordinates! B p = B C A! p A to where B A C = B A R! B A t
19 Camera calibration Use the camera to tell you things about the world: Relationship between coordinates in the world and coordinates in the image: geometric camera calibration, see Szeliski, section 5.2, 5.3 for references (Relationship between intensities in the world and intensities in the image: photometric image formation, see Szeliski, sect. 2.2.)
20 Camera calibration Intrinsic parameters Image coordinates relative to camera!" Pixel coordinates Extrinsic parameters Camera frame 1!" Camera frame 2
21 Camera calibration Intrinsic parameters Extrinsic parameters
22 Intrinsic parameters: from idealized world coordinates to pixel values Forsyth&Ponce Perspective projection u v = = f f x z y z
23 Intrinsic parameters But pixels are in some arbitrary spatial units u v = α = α x z y z
24 Intrinsic parameters Maybe pixels are not square u v = = α β x z y z
25 Intrinsic parameters The origin of our camera pixel coordinates may be somewhere other than under the camera optical axis. u v = = α β x z y z + u + v 0 0
26 Intrinsic parameters 0 0 ) sin( ) cot( v z y v u z y z x u + = + = θ β θ α α May be skew between camera pixel axes (but usually this angle is 90 deg). v θ u vʹ uʹ v u v u u v v ) cot( ) cos( ) sin( θ θ θ = ʹ = ʹ = ʹ
27 ! p = K C! p Intrinsic parameters, homogeneous coordinates 0 0 ) sin( ) cot( v z y v u z y z x u + = + = θ β θ α α u v 1 = α α cot(θ) u 0 0 β sin(θ) v x y z 1 Using homogenous coordinates, we can write this as: or: In camera-based coords In pixels
28 Camera calibration Intrinsic parameters Extrinsic parameters
29 World and camera coordinate systems In the first lecture, we placed the world coordinates in the center of the scene.
30 Extrinsic parameters: translation and rotation of camera frame C!! p = C W R W p +! C W t Non-homogeneous coordinates C! p C = W R! C W t W! p Homogeneous coordinates
31 pixels Camera coordinates C!!! p = K C p p C = W R! C W t W! p Intrinsic World coordinates Extrinsic Combining extrinsic and intrinsic calibration parameters, in homogeneous coordinates Forsyth&Ponce! p = K C W R! C W t!! p = M W p ( ) W! p
32 Other ways to write the same equation pixel coordinates! p = M W! p world coordinates u v = 1 T. m 1 T. m 2 T. m W p x W p y W p z 1 Conversion back from homogeneous coordinates leads to: u v = = m m m m ! P! P! P! P
33 Summary camera parameters A camera is described by several parameters Translation T of the optical center from the origin of world coords Rotation R of the image plane focal length f, principle point (x c, y c ), pixel size (s x, s y ) blue parameters are called extrinsics, red are intrinsics Projection equation sx * X * Y Z * 1 x = sy = * * * * = ΠX s * The projection matrix models the cumulative effect of all parameters Useful to decompose into a series of operations * * fs x 0 x' c R Π = 0 fs y y' c x x 3 1 * * 0 3x1 I 3x 3 T 3x1 0 1x 3 1 intrinsics projection rotation translation identity matrix The definitions of these parameters are not completely standardized especially intrinsics varies from one book to another
34 do we calculate the camera s calibration matrix, or measure?
35 Calibration target Find the position, u i and v i, in pixels, of each calibration object feature point.
36 Camera calibration From before, we had these equations relating image positions, u,v, to points at 3-d positions P (in homogeneous coordinates): So for each feature point, i, we have:
37 Camera calibration Stack all these measurements of i=1 n points into a big matrix (cluttering vector arrows omitted from P and m):
38 In vector form: Camera calibration Showing all the elements:
39 Camera calibration Q m = 0 We want to solve for the unit vector m (the stacked one) that minimizes The minimum eigenvector of the matrix Q T Q gives us that (see Forsyth&Ponce, 3.1), because it is the unit vector x that minimizes x T Q T Q x.
40 Camera calibration Once you have the M matrix, can recover the intrinsic and extrinsic parameters.
41 Vision systems One camera Two cameras N cameras
42 Stereo vision ~6cm ~50cm
43 Depth without objects Random dot stereograms (Bela Julesz) Julesz, 1971
44 Stereo photography and stereo viewers Take two pictures of the same subject from two slightly different viewpoints and display so that each eye sees only one of the images. Image courtesy of fisher-price.com Slide credit: Kristen Grauman
45 Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923 Slide credit: Kristen Grauman
46 Anaglyph pinhole camera
47 Autostereograms Exploit disparity as depth cue using single image. (Single image random dot stereogram, Single image stereogram) Images from magiceye.com Slide credit: Kristen Grauman
48 My conundrum regarding stereo displays Real 3d scenes often look to me like thin, flat layers, stacked in depth. Why is that? 48
49 Estimating depth with stereo Stereo: shape from disparities between two views We ll need to consider: Info on camera pose ( calibration ) Image point correspondences scene point optical center image plane Slide credit: Kristen Grauman
50 Geometry for a simple stereo system Assume a simple setting: Two identical cameras parallel optical axes known camera parameters (i.e., calibrated cameras). O l O r
51 World point image point (left) Depth of p image point (right) Focal length optical center (left) baseline optical center (right) Slide credit: Kristen Grauman
52 Geometry for a simple stereo system Assume parallel optical axes, known camera parameters (i.e., calibrated cameras). We can triangulate via: Similar triangles (p l, P, p r ) and (O l, P, O r ): disparity Slide credit: Kristen Grauman
53 Depth from disparity image I(x,y) Disparity map D(x,y) image I (x,y ) (x,y )=(x+d(x,y), y) Slide credit: Kristen Grauman
54 General case, with calibrated cameras The two cameras need not have parallel optical axes. Vs.
55 Stereo correspondence constraints Camera 1 Camera 2 p p? O O If we see a point in camera 1, are there any constraints on where we will find it on camera 2?
56 Epipolar constraint p p? O O Geometry of two views constrains where the corresponding pixel for some image point in the first view must occur in the second view: It must be on the line carved out by a plane connecting the world point and optical centers. Why is this useful?
57 Epipolar constraint This is useful because it reduces the correspondence problem to a 1D search along an epipolar line. Image from Andrew Zisserman Slide credit: Kristen Grauman
58 Epipolar geometry Epipolar Line Epipolar Plane Epipolar Line Epipole Baseline Epipole Epipolar plane: plane containing baseline and world point Epipole: point of intersection of baseline with the image plane Epipolar line: intersection of epipolar plane with the image plane Baseline: line joining the camera centers All epipolar lines intersect at the epipole An epipolar plane intersects the left and right image planes in epipolar lines Slide credit: Kristen Grauman
59 Example Slide credit: Kristen Grauman
60 Example: parallel cameras Where are the epipoles? Figure from Hartley & Zisserman Slide credit: Kristen Grauman
61 Example: converging cameras Figure from Hartley & Zisserman Slide credit: Kristen Grauman
62 So far, we have the explanation in terms of geometry. Now, how to express the epipolar constraints algebraically? Slide credit: Kristen Grauman
63 Stereo geometry, with calibrated cameras Main idea Slide credit: Kristen Grauman
64 Stereo geometry, with calibrated cameras If the stereo rig is calibrated, we know : how to rotate and translate camera reference frame 1 to get to camera reference frame 2. Rotation: 3 x 3 matrix R; translation: 3 vector T. Slide credit: Kristen Grauman
65 Stereo geometry, with calibrated cameras If the stereo rig is calibrated, we know : how to rotate and translate camera reference frame 1 to get to camera reference frame 2. Slide credit: Kristen Grauman
66 From geometry to algebra X' = RX + T T Xʹ = T RX + T T Normal to the plane = T RX ( T Xʹ) = Xʹ ( T RX) X ʹ = 0 Slide credit: Kristen Grauman
67 From geometry to algebra X' = RX + T T Xʹ = T RX + T T Normal to the plane = T RX ( T Xʹ) = Xʹ ( T RX) X ʹ = 0 Slide credit: Kristen Grauman
68 Aside: cross product Vector cross product takes two vectors and returns a third vector that s perpendicular to both inputs. So here, c is perpendicular to both a and b, which means the dot product = 0. Slide credit: Kristen Grauman
69 Matrix form of cross product Can be expressed as a matrix multiplication. c b b b a a a a a a b a!!! = = [ ] = a a a a a a a x Slide credit: Kristen Grauman
70 Let Xʹ Xʹ E = ( T RX) = 0 ( T ) x RX = 0 T x R Xʹ T EX = 0 Essential matrix E is called the essential matrix, and it relates corresponding image points between both cameras, given the rotation and translation. If we observe a point in one image, its position in the other image is constrained to lie on line defined by above.
71 x and x are scaled versions of X and X
72 Let pts x and x in the image planes are scaled versions of X and X. E is called the essential matrix, and it relates corresponding image points between both cameras, given the rotation and translation. If we observe a point in one image, its position in the other image is constrained to lie on line defined by above. Note: these points are in camera coordinate systems.
73 Essential matrix example: parallel cameras d 0 d 0 For the parallel cameras, image of any point must lie on same horizontal line in each image plane. Slide credit: Kristen Grauman
74 image I(x,y) Disparity map D(x,y) image I (x,y ) (x,y )=(x+d(x,y),y) What about when cameras optical axes are not parallel? Slide credit: Kristen Grauman
75 Stereo image rectification: example Source: Alyosha Efros
76 Stereo image rectification In practice, it is convenient if image scanlines (rows) are the epipolar lines. Reproject image planes onto a common plane parallel to the line between optical centers Pixel motion is horizontal after this transformation Two homographies (3x3 transforms), one for each input image reprojection See Szeliski book, Sect , Fig. 2.12, and Mapping from one camera to another p. 56: Adapted from Li Zhang Slide credit: Kristen Grauman
77 Your basic stereo algorithm For each epipolar line For each pixel in the left image compare with every pixel on same epipolar line in right image pick pixel with minimum match cost Improvement: match windows CSE 576, Slide Spring credit: 2008 Rick Szeliski Stereo matching 77
78 Image block matching How do we determine correspondences? block matching or SSD (sum squared differences) d is the disparity (horizontal motion) How big should the neighborhood be? CSE 576, Spring 2008 Slide credit: Rick Szeliski Stereo matching 78
79 Neighborhood size Smaller neighborhood: more details Larger neighborhood: fewer isolated mistakes w = 3 w = 20 CSE 576, Slide Spring credit: 2008 Rick Szeliski Stereo matching 79
80 Matching criteria Raw pixel values (correlation) Band-pass filtered images [Jones & Malik 92] Corner like features [Zhang, ] Edges [many people ] Gradients [Seitz 89; Scharstein 94] Rank statistics [Zabih & Woodfill 94] CSE 576, Slide Spring credit: 2008 Rick Szeliski Stereo matching 80
81 Local evidence framework For every disparity, compute raw matching costs Why use a robust function? occlusions, other outliers Can also use alternative match criteria CSE 576, Slide Spring credit: 2008 Rick Szeliski Stereo matching 81
82 Local evidence framework Aggregate costs spatially Here, we are using a box filter (efficient moving average implementation) Can also use weighted average, [non-linear] diffusion CSE 576, Slide Spring credit: 2008 Rick Szeliski Stereo matching 82
83 Local evidence framework Choose winning disparity at each pixel Interpolate to sub-pixel accuracy E(d) d* d CSE 576, Slide Spring credit: 2008 Rick Szeliski Stereo matching 83
84 Active stereo with structured light Li Zhang s one-shot stereo camera 1 camera 1 projector projector camera 2 Project structured light patterns onto the object simplifies the correspondence problem Li Zhang, Brian Curless, and Steven M. Seitz. Rapid Shape Acquisition Using Color Structured Light and Multi-pass Dynamic Programming. In Proceedings of the 1st International Symposium on 3D Data Processing, Visualization, and Transmission (3DPVT), Padova, Italy, June 19-21, 2002, pp CSE 576, Slide Spring credit: 2008 Rick Szeliski Stereo matching 84
85 CSE 576, Spring 2008 Li Zhang, Brian Curless, and Steven M. Seitz Stereo matching 85
86 CSE 576, Spring 2008 Stereo matching 86
87 CSE 576, Spring 2008 Stereo matching 87
88 CSE 576, Spring 2008 Stereo matching 88
89 89
90 90
91 91
92 Bibliography D. Scharstein and R. Szeliski. A taxonomy and evaluation of dense twoframe stereo correspondence algorithms. International Journal of Computer Vision, 47(1):7-42, May R. Szeliski. Stereo algorithms and representations for image-based rendering. In British Machine Vision Conference (BMVC'99), volume 2, pages , Nottingham, England, September G. M. Nielson, Scattered Data Modeling, IEEE Computer Graphics and Applications, 13(1), January 1993, pp S. B. Kang, R. Szeliski, and J. Chai. Handling occlusions in dense multiview stereo. In CVPR'2001, vol. I, pages , December Y. Boykov, O. Veksler, and Ramin Zabih, Fast Approximate Energy Minimization via Graph Cuts, Unpublished manuscript, A.F. Bobick and S.S. Intille. Large occlusion stereo. International Journal of Computer Vision, 33(3), September pp D. Scharstein and R. Szeliski. Stereo matching with nonlinear diffusion. International Journal of Computer Vision, 28(2): , July 1998 CSE 576, Slide Spring credit: 2008 Rick Szeliski Stereo matching 92
93 Bibliography Volume Intersection Martin & Aggarwal, Volumetric description of objects from multiple views, Trans. Pattern Analysis and Machine Intelligence, 5(2), 1991, pp Szeliski, Rapid Octree Construction from Image Sequences, Computer Vision, Graphics, and Image Processing: Image Understanding, 58(1), 1993, pp Voxel Coloring and Space Carving Seitz & Dyer, Photorealistic Scene Reconstruction by Voxel Coloring, Proc. Computer Vision and Pattern Recognition (CVPR), 1997, pp Seitz & Kutulakos, Plenoptic Image Editing, Proc. Int. Conf. on Computer Vision (ICCV), 1998, pp Kutulakos & Seitz, A Theory of Shape by Space Carving, Proc. ICCV, 1998, pp CSE 576, Slide Spring credit: 2008 Rick Szeliski Stereo matching 93
94 Bibliography Related References Bolles, Baker, and Marimont, Epipolar-Plane Image Analysis: An Approach to Determining Structure from Motion, International Journal of Computer Vision, vol 1, no 1, 1987, pp Faugeras & Keriven, Variational principles, surface evolution, PDE's, level set methods and the stereo problem", IEEE Trans. on Image Processing, 7(3), 1998, pp Szeliski & Golland, Stereo Matching with Transparency and Matting, Proc. Int. Conf. on Computer Vision (ICCV), 1998, Roy & Cox, A Maximum-Flow Formulation of the N-camera Stereo Correspondence Problem, Proc. ICCV, 1998, pp Fua & Leclerc, Object-centered surface reconstruction: Combining multi-image stereo and shading", International Journal of Computer Vision, 16, 1995, pp Narayanan, Rander, & Kanade, Constructing Virtual Worlds Using Dense Stereo, Proc. ICCV, 1998, pp CSE 576, Slide Spring credit: 2008 Rick Szeliski Stereo matching 94
Stereo Matching. Stereo Matching. Face modeling. Z-keying: mix live and synthetic
Stereo Matching Stereo Matching Given two or more images of the same scene or object, compute a representation of its shape? Computer Vision CSE576, Spring 2005 Richard Szeliski What are some possible
More informationRecap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?
Recap: Features and filters Epipolar geometry & stereo vision Tuesday, Oct 21 Kristen Grauman UT-Austin Transforming and describing images; textures, colors, edges Recap: Grouping & fitting Now: Multiple
More informationStereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman
Stereo 11/02/2012 CS129, Brown James Hays Slides by Kristen Grauman Multiple views Multi-view geometry, matching, invariant features, stereo vision Lowe Hartley and Zisserman Why multiple views? Structure
More informationComputer Vision Lecture 17
Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week
More informationComputer Vision Lecture 17
Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester
More informationThere are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...
STEREO VISION The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Bill Freeman and Antonio Torralba (MIT), including their own
More informationEpipolar Geometry and Stereo Vision
CS 1674: Intro to Computer Vision Epipolar Geometry and Stereo Vision Prof. Adriana Kovashka University of Pittsburgh October 5, 2016 Announcement Please send me three topics you want me to review next
More informationStereo: Disparity and Matching
CS 4495 Computer Vision Aaron Bobick School of Interactive Computing Administrivia PS2 is out. But I was late. So we pushed the due date to Wed Sept 24 th, 11:55pm. There is still *no* grace period. To
More informationProject 4 Results. Representation. Data. Learning. Zachary, Hung-I, Paul, Emanuel. SIFT and HoG are popular and successful.
Project 4 Results Representation SIFT and HoG are popular and successful. Data Hugely varying results from hard mining. Learning Non-linear classifier usually better. Zachary, Hung-I, Paul, Emanuel Project
More informationCS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching
Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix
More informationImage Rectification (Stereo) (New book: 7.2.1, old book: 11.1)
Image Rectification (Stereo) (New book: 7.2.1, old book: 11.1) Guido Gerig CS 6320 Spring 2013 Credits: Prof. Mubarak Shah, Course notes modified from: http://www.cs.ucf.edu/courses/cap6411/cap5415/, Lecture
More informationEpipolar Geometry and Stereo Vision
CS 1699: Intro to Computer Vision Epipolar Geometry and Stereo Vision Prof. Adriana Kovashka University of Pittsburgh October 8, 2015 Today Review Projective transforms Image stitching (homography) Epipolar
More informationStereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision
Stereo Thurs Mar 30 Kristen Grauman UT Austin Outline Last time: Human stereopsis Epipolar geometry and the epipolar constraint Case example with parallel optical axes General case with calibrated cameras
More informationStereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz
Stereo II CSE 576 Ali Farhadi Several slides from Larry Zitnick and Steve Seitz Camera parameters A camera is described by several parameters Translation T of the optical center from the origin of world
More informationBIL Computer Vision Apr 16, 2014
BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm
More informationMiniature faking. In close-up photo, the depth of field is limited.
Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg
More informationCS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching
Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix
More informationBinocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?
Binocular Stereo Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image Where does the depth information come from? Binocular stereo Given a calibrated binocular stereo
More informationThink-Pair-Share. What visual or physiological cues help us to perceive 3D shape and depth?
Think-Pair-Share What visual or physiological cues help us to perceive 3D shape and depth? [Figure from Prados & Faugeras 2006] Shading Focus/defocus Images from same point of view, different camera parameters
More informationMultiple View Geometry
Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V
More informationChaplin, Modern Times, 1936
Chaplin, Modern Times, 1936 [A Bucket of Water and a Glass Matte: Special Effects in Modern Times; bonus feature on The Criterion Collection set] Multi-view geometry problems Structure: Given projections
More informationCS5670: Computer Vision
CS5670: Computer Vision Noah Snavely, Zhengqi Li Stereo Single image stereogram, by Niklas Een Mark Twain at Pool Table", no date, UCR Museum of Photography Stereo Given two images from different viewpoints
More informationWhat have we leaned so far?
What have we leaned so far? Camera structure Eye structure Project 1: High Dynamic Range Imaging What have we learned so far? Image Filtering Image Warping Camera Projection Model Project 2: Panoramic
More informationStereo vision. Many slides adapted from Steve Seitz
Stereo vision Many slides adapted from Steve Seitz What is stereo vision? Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape What is
More informationRecovering structure from a single view Pinhole perspective projection
EPIPOLAR GEOMETRY The slides are from several sources through James Hays (Brown); Silvio Savarese (U. of Michigan); Svetlana Lazebnik (U. Illinois); Bill Freeman and Antonio Torralba (MIT), including their
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 12 130228 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Panoramas, Mosaics, Stitching Two View Geometry
More informationLecture 9 & 10: Stereo Vision
Lecture 9 & 10: Stereo Vision Professor Fei- Fei Li Stanford Vision Lab 1 What we will learn today? IntroducEon to stereo vision Epipolar geometry: a gentle intro Parallel images Image receficaeon Solving
More informationStereo. Many slides adapted from Steve Seitz
Stereo Many slides adapted from Steve Seitz Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image image 1 image 2 Dense depth map Binocular stereo Given a calibrated
More informationProject 3 code & artifact due Tuesday Final project proposals due noon Wed (by ) Readings Szeliski, Chapter 10 (through 10.5)
Announcements Project 3 code & artifact due Tuesday Final project proposals due noon Wed (by email) One-page writeup (from project web page), specifying:» Your team members» Project goals. Be specific.
More informationMosaics wrapup & Stereo
Mosaics wrapup & Stereo Tues Oct 20 Last time: How to stitch a panorama? Basic Procedure Take a sequence of images from the same position Rotate the camera about its optical center Compute transformation
More informationFinal project bits and pieces
Final project bits and pieces The project is expected to take four weeks of time for up to four people. At 12 hours per week per person that comes out to: ~192 hours of work for a four person team. Capstone:
More informationPublic Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923
Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923 Teesta suspension bridge-darjeeling, India Mark Twain at Pool Table", no date, UCR Museum of Photography Woman getting eye exam during
More informationCS 2770: Intro to Computer Vision. Multiple Views. Prof. Adriana Kovashka University of Pittsburgh March 14, 2017
CS 277: Intro to Computer Vision Multiple Views Prof. Adriana Kovashka Universit of Pittsburgh March 4, 27 Plan for toda Affine and projective image transformations Homographies and image mosaics Stereo
More informationToday. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography
Computational Photography Matthias Zwicker University of Bern Fall 2009 Today From 2D to 3D using multiple views Introduction Geometry of two views Stereo matching Other applications Multiview geometry
More informationEpipolar Geometry and Stereo Vision
Epipolar Geometry and Stereo Vision Computer Vision Jia-Bin Huang, Virginia Tech Many slides from S. Seitz and D. Hoiem Last class: Image Stitching Two images with rotation/zoom but no translation. X x
More informationEpipolar Geometry and Stereo Vision
Epipolar Geometry and Stereo Vision Computer Vision Shiv Ram Dubey, IIIT Sri City Many slides from S. Seitz and D. Hoiem Last class: Image Stitching Two images with rotation/zoom but no translation. X
More informationImage Based Reconstruction II
Image Based Reconstruction II Qixing Huang Feb. 2 th 2017 Slide Credit: Yasutaka Furukawa Image-Based Geometry Reconstruction Pipeline Last Lecture: Multi-View SFM Multi-View SFM This Lecture: Multi-View
More informationCamera Geometry II. COS 429 Princeton University
Camera Geometry II COS 429 Princeton University Outline Projective geometry Vanishing points Application: camera calibration Application: single-view metrology Epipolar geometry Application: stereo correspondence
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem
More informationCameras and Stereo CSE 455. Linda Shapiro
Cameras and Stereo CSE 455 Linda Shapiro 1 Müller-Lyer Illusion http://www.michaelbach.de/ot/sze_muelue/index.html What do you know about perspective projection? Vertical lines? Other lines? 2 Image formation
More informationRecap from Previous Lecture
Recap from Previous Lecture Tone Mapping Preserve local contrast or detail at the expense of large scale contrast. Changing the brightness within objects or surfaces unequally leads to halos. We are now
More informationSome books on linear algebra
Some books on linear algebra Finite Dimensional Vector Spaces, Paul R. Halmos, 1947 Linear Algebra, Serge Lang, 2004 Linear Algebra and its Applications, Gilbert Strang, 1988 Matrix Computation, Gene H.
More informationEECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems
EECS 442 Computer vision Stereo systems Stereo vision Rectification Correspondence problem Active stereo vision systems Reading: [HZ] Chapter: 11 [FP] Chapter: 11 Stereo vision P p p O 1 O 2 Goal: estimate
More informationLecture 10: Multi-view geometry
Lecture 10: Multi-view geometry Professor Stanford Vision Lab 1 What we will learn today? Review for stereo vision Correspondence problem (Problem Set 2 (Q3)) Active stereo vision systems Structure from
More informationMultiple View Geometry
Multiple View Geometry CS 6320, Spring 2013 Guest Lecture Marcel Prastawa adapted from Pollefeys, Shah, and Zisserman Single view computer vision Projective actions of cameras Camera callibration Photometric
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction
More informationStereo and Epipolar geometry
Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka
More informationFundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F
Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix F Fundamental
More informationMulti-View Geometry (Ch7 New book. Ch 10/11 old book)
Multi-View Geometry (Ch7 New book. Ch 10/11 old book) Guido Gerig CS-GY 6643, Spring 2016 gerig@nyu.edu Credits: M. Shah, UCF CAP5415, lecture 23 http://www.cs.ucf.edu/courses/cap6411/cap5415/, Trevor
More informationCOSC579: Scene Geometry. Jeremy Bolton, PhD Assistant Teaching Professor
COSC579: Scene Geometry Jeremy Bolton, PhD Assistant Teaching Professor Overview Linear Algebra Review Homogeneous vs non-homogeneous representations Projections and Transformations Scene Geometry The
More informationCamera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration
Camera Calibration Jesus J Caban Schedule! Today:! Camera calibration! Wednesday:! Lecture: Motion & Optical Flow! Monday:! Lecture: Medical Imaging! Final presentations:! Nov 29 th : W. Griffin! Dec 1
More informationMulti-view geometry problems
Multi-view geometry Multi-view geometry problems Structure: Given projections o the same 3D point in two or more images, compute the 3D coordinates o that point? Camera 1 Camera 2 R 1,t 1 R 2,t 2 Camera
More informationImage Formation I Chapter 2 (R. Szelisky)
Image Formation I Chapter 2 (R. Selisky) Guido Gerig CS 632 Spring 22 cknowledgements: Slides used from Prof. Trevor Darrell, (http://www.eecs.berkeley.edu/~trevor/cs28.html) Some slides modified from
More informationAnnouncements. Stereo
Announcements Stereo Homework 2 is due today, 11:59 PM Homework 3 will be assigned today Reading: Chapter 7: Stereopsis CSE 152 Lecture 8 Binocular Stereopsis: Mars Given two images of a scene where relative
More informationCS4495/6495 Introduction to Computer Vision. 3B-L3 Stereo correspondence
CS4495/6495 Introduction to Computer Vision 3B-L3 Stereo correspondence For now assume parallel image planes Assume parallel (co-planar) image planes Assume same focal lengths Assume epipolar lines are
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationLecture 9: Epipolar Geometry
Lecture 9: Epipolar Geometry Professor Fei Fei Li Stanford Vision Lab 1 What we will learn today? Why is stereo useful? Epipolar constraints Essential and fundamental matrix Estimating F (Problem Set 2
More informationLecture 11 MRF s (conbnued), cameras and lenses.
6.869 Advances in Computer Vision Bill Freeman and Antonio Torralba Spring 2011 Lecture 11 MRF s (conbnued), cameras and lenses. remember correction on Gibbs sampling Motion application image patches image
More informationImage Formation I Chapter 1 (Forsyth&Ponce) Cameras
Image Formation I Chapter 1 (Forsyth&Ponce) Cameras Guido Gerig CS 632 Spring 215 cknowledgements: Slides used from Prof. Trevor Darrell, (http://www.eecs.berkeley.edu/~trevor/cs28.html) Some slides modified
More informationImage Formation I Chapter 1 (Forsyth&Ponce) Cameras
Image Formation I Chapter 1 (Forsyth&Ponce) Cameras Guido Gerig CS 632 Spring 213 cknowledgements: Slides used from Prof. Trevor Darrell, (http://www.eecs.berkeley.edu/~trevor/cs28.html) Some slides modified
More informationStereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz
Stereo CSE 576 Ali Farhadi Several slides from Larry Zitnick and Steve Seitz Why do we perceive depth? What do humans use as depth cues? Motion Convergence When watching an object close to us, our eyes
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 5: Projection Reading: Szeliski 2.1 Projection Reading: Szeliski 2.1 Projection Müller Lyer Illusion http://www.michaelbach.de/ot/sze_muelue/index.html Modeling
More informationLecture 10: Multi view geometry
Lecture 10: Multi view geometry Professor Fei Fei Li Stanford Vision Lab 1 What we will learn today? Stereo vision Correspondence problem (Problem Set 2 (Q3)) Active stereo vision systems Structure from
More informationLecture 14: Computer Vision
CS/b: Artificial Intelligence II Prof. Olga Veksler Lecture : Computer Vision D shape from Images Stereo Reconstruction Many Slides are from Steve Seitz (UW), S. Narasimhan Outline Cues for D shape perception
More informationComputer Vision I. Announcements. Random Dot Stereograms. Stereo III. CSE252A Lecture 16
Announcements Stereo III CSE252A Lecture 16 HW1 being returned HW3 assigned and due date extended until 11/27/12 No office hours today No class on Thursday 12/6 Extra class on Tuesday 12/4 at 6:30PM in
More informationComputer Vision I. Announcement. Stereo Vision Outline. Stereo II. CSE252A Lecture 15
Announcement Stereo II CSE252A Lecture 15 HW3 assigned No class on Thursday 12/6 Extra class on Tuesday 12/4 at 6:30PM in WLH Room 2112 Mars Exploratory Rovers: Spirit and Opportunity Stereo Vision Outline
More informationSome books on linear algebra
Some books on linear algebra Finite Dimensional Vector Spaces, Paul R. Halmos, 1947 Linear Algebra, Serge Lang, 2004 Linear Algebra and its Applications, Gilbert Strang, 1988 Matrix Computation, Gene H.
More informationAnnouncements. Stereo
Announcements Stereo Homework 1 is due today, 11:59 PM Homework 2 will be assigned on Thursday Reading: Chapter 7: Stereopsis CSE 252A Lecture 8 Binocular Stereopsis: Mars Given two images of a scene where
More informationEpipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz
Epipolar Geometry Prof. D. Stricker With slides from A. Zisserman, S. Lazebnik, Seitz 1 Outline 1. Short introduction: points and lines 2. Two views geometry: Epipolar geometry Relation point/line in two
More informationStructure from motion
Multi-view geometry Structure rom motion Camera 1 Camera 2 R 1,t 1 R 2,t 2 Camera 3 R 3,t 3 Figure credit: Noah Snavely Structure rom motion? Camera 1 Camera 2 R 1,t 1 R 2,t 2 Camera 3 R 3,t 3 Structure:
More informationBut First: Multi-View Projective Geometry
View Morphing (Seitz & Dyer, SIGGRAPH 96) Virtual Camera Photograph Morphed View View interpolation (ala McMillan) but no depth no camera information Photograph But First: Multi-View Projective Geometry
More informationStereo Epipolar Geometry for General Cameras. Sanja Fidler CSC420: Intro to Image Understanding 1 / 33
Stereo Epipolar Geometry for General Cameras Sanja Fidler CSC420: Intro to Image Understanding 1 / 33 Stereo Epipolar geometry Case with two cameras with parallel optical axes General case Now this Sanja
More informationLecture 14: Basic Multi-View Geometry
Lecture 14: Basic Multi-View Geometry Stereo If I needed to find out how far point is away from me, I could use triangulation and two views scene point image plane optical center (Graphic from Khurram
More informationMachine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy
1 Machine vision Summary # 11: Stereo vision and epipolar geometry STEREO VISION The goal of stereo vision is to use two cameras to capture 3D scenes. There are two important problems in stereo vision:
More informationProject 2 due today Project 3 out today. Readings Szeliski, Chapter 10 (through 10.5)
Announcements Stereo Project 2 due today Project 3 out today Single image stereogram, by Niklas Een Readings Szeliski, Chapter 10 (through 10.5) Public Library, Stereoscopic Looking Room, Chicago, by Phillips,
More informationHomographies and RANSAC
Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position
More informationStereo Vision. MAN-522 Computer Vision
Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in
More informationMulti-view stereo. Many slides adapted from S. Seitz
Multi-view stereo Many slides adapted from S. Seitz Beyond two-view stereo The third eye can be used for verification Multiple-baseline stereo Pick a reference image, and slide the corresponding window
More informationStereo Image Rectification for Simple Panoramic Image Generation
Stereo Image Rectification for Simple Panoramic Image Generation Yun-Suk Kang and Yo-Sung Ho Gwangju Institute of Science and Technology (GIST) 261 Cheomdan-gwagiro, Buk-gu, Gwangju 500-712 Korea Email:{yunsuk,
More informationA virtual tour of free viewpoint rendering
A virtual tour of free viewpoint rendering Cédric Verleysen ICTEAM institute, Université catholique de Louvain, Belgium cedric.verleysen@uclouvain.be Organization of the presentation Context Acquisition
More informationColorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.
Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Stereo Vision 2 Inferring 3D from 2D Model based pose estimation single (calibrated) camera > Can
More informationAnnouncements. Stereo Vision Wrapup & Intro Recognition
Announcements Stereo Vision Wrapup & Intro Introduction to Computer Vision CSE 152 Lecture 17 HW3 due date postpone to Thursday HW4 to posted by Thursday, due next Friday. Order of material we ll first
More information3D Photography: Stereo Matching
3D Photography: Stereo Matching Kevin Köser, Marc Pollefeys Spring 2012 http://cvg.ethz.ch/teaching/2012spring/3dphoto/ Stereo & Multi-View Stereo Tsukuba dataset http://cat.middlebury.edu/stereo/ Stereo
More informationCamera Calibration. COS 429 Princeton University
Camera Calibration COS 429 Princeton University Point Correspondences What can you figure out from point correspondences? Noah Snavely Point Correspondences X 1 X 4 X 3 X 2 X 5 X 6 X 7 p 1,1 p 1,2 p 1,3
More informationMulti-View Geometry Part II (Ch7 New book. Ch 10/11 old book)
Multi-View Geometry Part II (Ch7 New book. Ch 10/11 old book) Guido Gerig CS-GY 6643, Spring 2016 gerig@nyu.edu Credits: M. Shah, UCF CAP5415, lecture 23 http://www.cs.ucf.edu/courses/cap6411/cap5415/,
More informationReminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches
Reminder: Lecture 20: The Eight-Point Algorithm F = -0.00310695-0.0025646 2.96584-0.028094-0.00771621 56.3813 13.1905-29.2007-9999.79 Readings T&V 7.3 and 7.4 Essential/Fundamental Matrix E/F Matrix Summary
More informationUnit 3 Multiple View Geometry
Unit 3 Multiple View Geometry Relations between images of a scene Recovering the cameras Recovering the scene structure http://www.robots.ox.ac.uk/~vgg/hzbook/hzbook1.html 3D structure from images Recover
More informationEpipolar Geometry CSE P576. Dr. Matthew Brown
Epipolar Geometry CSE P576 Dr. Matthew Brown Epipolar Geometry Epipolar Lines, Plane Constraint Fundamental Matrix, Linear solution + RANSAC Applications: Structure from Motion, Stereo [ Szeliski 11] 2
More informationComputer Vision cmput 428/615
Computer Vision cmput 428/615 Basic 2D and 3D geometry and Camera models Martin Jagersand The equation of projection Intuitively: How do we develop a consistent mathematical framework for projection calculations?
More informationCS 664 Slides #9 Multi-Camera Geometry. Prof. Dan Huttenlocher Fall 2003
CS 664 Slides #9 Multi-Camera Geometry Prof. Dan Huttenlocher Fall 2003 Pinhole Camera Geometric model of camera projection Image plane I, which rays intersect Camera center C, through which all rays pass
More information1 Projective Geometry
CIS8, Machine Perception Review Problem - SPRING 26 Instructions. All coordinate systems are right handed. Projective Geometry Figure : Facade rectification. I took an image of a rectangular object, and
More informationCorrespondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri] Introduction Disparity: Informally: difference between two pictures Allows us to gain a strong
More information3D Geometry and Camera Calibration
3D Geometry and Camera Calibration 3D Coordinate Systems Right-handed vs. left-handed x x y z z y 2D Coordinate Systems 3D Geometry Basics y axis up vs. y axis down Origin at center vs. corner Will often
More informationCS4670: Computer Vision
CS467: Computer Vision Noah Snavely Lecture 13: Projection, Part 2 Perspective study of a vase by Paolo Uccello Szeliski 2.1.3-2.1.6 Reading Announcements Project 2a due Friday, 8:59pm Project 2b out Friday
More informationCEng Computational Vision
CEng 583 - Computational Vision 2011-2012 Spring Week 4 18 th of March, 2011 Today 3D Vision Binocular (Multi-view) cues: Stereopsis Motion Monocular cues Shading Texture Familiar size etc. "God must
More informationGeometric camera models and calibration
Geometric camera models and calibration http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 13 Course announcements Homework 3 is out. - Due October
More informationStructure from Motion and Multi- view Geometry. Last lecture
Structure from Motion and Multi- view Geometry Topics in Image-Based Modeling and Rendering CSE291 J00 Lecture 5 Last lecture S. J. Gortler, R. Grzeszczuk, R. Szeliski,M. F. Cohen The Lumigraph, SIGGRAPH,
More informationN-Views (1) Homographies and Projection
CS 4495 Computer Vision N-Views (1) Homographies and Projection Aaron Bobick School of Interactive Computing Administrivia PS 2: Get SDD and Normalized Correlation working for a given windows size say
More informationCamera Model and Calibration
Camera Model and Calibration Lecture-10 Camera Calibration Determine extrinsic and intrinsic parameters of camera Extrinsic 3D location and orientation of camera Intrinsic Focal length The size of the
More informationTwo-view geometry Computer Vision Spring 2018, Lecture 10
Two-view geometry http://www.cs.cmu.edu/~16385/ 16-385 Computer Vision Spring 2018, Lecture 10 Course announcements Homework 2 is due on February 23 rd. - Any questions about the homework? - How many of
More information