Computer Vision I. Dense Stereo Correspondences. Anita Sellent 1/15/16
|
|
- Camron Clarke
- 5 years ago
- Views:
Transcription
1 Computer Vision I Dense Stereo Correspondences Anita Sellent
2 Stereo Two Cameras Overlapping field of view Known transformation between cameras From disparity compute depth [ Bradski, Kaehler: Learning OpenCV, O'Reilly 2008] 2
3 Triangulation: Top View fb fb Z= = xl x r d d =x l x r disparity B B x l +x r = Z Z f Z xl xr f f C1 B C2 3
4 Triangulation: Top View C1 B C2 4
5 Epipolar Lines
6 Epipolar Lines Epipolar Lines ( see CV 1)
7 Epipolar Lines Epipolar Lines ( see CV 1)
8 Epipolar Lines Epipolar Lines ( see CV 1)
9 Practically Cameras are never perfectly aligned. The lenses of the cameras are never perfect.
10 Practically Cameras are never perfectly aligned. The lenses of the cameras are never perfect. e om s ide v o Pr atic m the a m p! l e al h
11 Practically Image undistortion Remove radial and tangential lens distortion Stereo calibration Compute the geometrical relationships between cameras in space Stereo rectification Transform individual images so that they appear as if the had seem taken by two row aligned image planes 16
12 Practically Image undistortion Remove radial and tangential lens distortion x=g 1( ^x, ^y )= ^x (1+K 1 r +K 2 r + )+(P 2 (r +2 x^ )+2 P1 x^ ^y )(1+K 1 r +K 2 r + ) y =g 2 ( x^, ^y ) 17
13 Practically Image undistortion Remove radial and tangential lens distortion s=g 1 (x, y)=x (1+K 1 r 2 +K 2 r 4 + )+(P2 (r 2+2 x 2 )+2 P1 x y )(1+K 1 r 2 +K 2 r 4 + ) t=g 2( x, y ) Undistorted position Distorted position 18
14 Geometric Image Transformations T Coordinates: T (x, y ) (x, y) I (s, t) Image: 19
15 Image Interpolation Schemes Nearest Neighbor Approach ( x, y ) Bilinear Interpolation (x, y) 20
16 Practically Image undistortion Stereo calibration Stereo rectification Planar Strong features 21
17 A Stereo Vision System 22
18 Calibration (offline) Rectification Stereo pair Stereo Correspondences Rectified Stereo pair Disparity map Triangulation Depth map Intrinsic and extrinsic parameter A Stereo Vision System PC, FPGA 23
19 Disparity Estimation: Assumption I The color of a pixel does not change with the viewpoint 24
20 Quiz: Which Points Satisfy Color Constancy? : None of the above 25
21 Disparity Estimation: Assumption I The color of a pixel does not change with the viewpoint Noise Reflective Material Transparent Material Scattering Effects Pixels that are visible only in one camera 26
22 Disparity Estimation: Assumption I The color of a pixel does not change with the viewpoint Noise Reflective Material Transparent Material Scattering Effects Pixels that are visible only in one camera 27
23 Disparity Estimation: Assumption I The color of a pixel does not change with the viewpoint I l (x, y) I r ( x d, y ) 28
24 Disparity Estimation: Assumption I There are 256 different gray values in an image. This image has 450 pixels per line. 29
25 Disparity Estimation: Assumption II Neighboring pixels belong to the same surface d (x, y ) d (x +ϵx, y +ϵ y ) 30
26 Disparity Estimation: Assumption II Neighboring pixels belong to the same surface Neighboring pixels belong to a fronto-paralell plane d (x, y )=d (x +Δ x, y +Δ y ) 31
27 A Basic Stereo Algorithm: Blockmatching For each pixel For each disparity value Compare patch similarity Assign disparity with highest similarity 32
28 A Basic Stereo Algorithm: Blockmatching For each pixel For each disparity value Compare patch similarity Assign disparity with highest similarity 33
29 Sparse vs. Dense Depth Estimation II Sparse Feature Matches (e.g. Harris corners ) Highly distinguishable texture Region descriptor robust to scale, rotation,... Dense Correspondences For every pixel Search along epipolar line 34
30 A Basic Stereo Algorithm: Blockmatching For each pixel For each disparity value Compare patch similarity Assign disparity with highest similarity 35
31 Discretization of the Solution Space Pros: Can use fast optimization algorithms Restrict need for image interpolation More flexibility in similarity measures Cons: Only certain distances become measurable 36
32 Discretization of the Solution Space Assume limited range of discrete disparities 37
33 Disparity fb fb Z= = xl x r d d =x l x r disparity Z C1 C2 disparity 38
34 Discretization of the Solution Space Assume limited range of discrete disparities Discrete estimates Ground truth 39
35 A Basic Stereo Algorithm: Blockmatching For each pixel For each disparity value Compare patch similarity Assign disparity with highest similarity 40
36 What are good Similarity Measures? Sum of Squared differences ( on a window ) Sum of Absolute differences ( on a window ) Any robust function ( on a window ) Any robust function on a brightness dependent window Cross Correlation (Normalized ) Census Transform with Hamming distance... [ ] 41
37 Image Noise Photon/ Shot noise Dark noise Photons arrive at the sensor at random times Thermal activity leads to random signals Read noise Charge conversion introduces random error ^I (x, y)=i (x, y)+ n(x, y) n N (0, σ) 42
38 Gaussian Image Noise Noise for each pixel in each camera is independent I^l (x, y)=i l (x, y)+nl (x, y ) I^r (x, y )=I r (x, y)+nr ( x, y ) Difference of images with disparity I^l (x, y) I^r ( x d, y )=I l (x, y ) I r (x d, y )+n (x, y) Probability Pr (I l, I r d, θ)= ( x, y) N 1 exp ( σ 2 π 2 (I l (x ) I r (x d )) 2σ 2 ) Negative logarithm E= log( Pr (I l, I r d, θ))= x N 1 2 ( I (x) I (x d )) +C l r 2 2σ 43
39 Example: Line Fitting Given: a collection of data-points Task: fit a line to the data-points {(x i, y i)}i=1... N [ Forsyth, Ponce(2003): Computer Vision, Ch. 15] 44
40 Example: Line Fitting Given: a collection of data-points Task: fit a line to the data-points {(x i, y i)}i=1... N x = u +λ n x y v ny ()() ( ) λ N (0, σ) [ Forsyth, Ponce(2003): Computer Vision, Ch. 15] 45
41 Example: Line Fitting {(x i, y i)}i=1... N Given: a collection of data-points Task: fit a line to the data-points Approach: Gaussian noise in normal direction x = u +λ n x y v ny ()() ( ) λ N (0, σ) [ Forsyth, Ponce(2003): Computer Vision, Ch. 15] 46
42 Example: Line Fitting Reminder: Express lines as 2 L={(x, y) ℝ x n x + y n y =d } nx ny () d [ Forsyth, Ponce(2003): Computer Vision, Ch. 15] 47
43 Example: Line Fitting Reminder: Express lines as Likelihood function 2 L={(x, y) ℝ x n x + y n y =d } Pr (x nx, n y, d )= Pr (x i, y i n x, n y,d ) 2 i (nx x i +n y y i d ) 1 = exp( ) 2 i σ 2 π 2σ nx ny () d [ Forsyth, Ponce(2003): Computer Vision, Ch. 15] 48
44 Example: Line Fitting Reminder: Express lines as Likelihood function 2 L={(x, y) ℝ x n x + y n y =d } Pr (x nx, n y, d )= Pr (x i, y i n x, n y,d ) i Max likelihood nx ny () E= (n x x i +n y y i d) 2 i 2 2 n x +n y =1 d [ Forsyth, Ponce(2003): Computer Vision, Ch. 15] 49
45 Example: Line Fitting [ Forsyth, Ponce(2003): Computer Vision, Ch. 15] 50
46 Similarity Measures Do brightness differences have a normal distribution? (Is noise our only problem? ) 51
47 Similarity Measures Do brightness differences have a normal distribution? Gaussian image noise ^I (x, y)=i (x, y)+ n(x, y) Lecture : Optical Flow 52
48 Similarity Measures Do brightness differences have a normal distribution? Gaussian image noise Color constancy assumption Lecture : Optical Flow 53
49 Similarity Measures Do brightness differences have a normal distribution? Gaussian image noise Color constancy assumption Inadequate window sizes & occlusion Lecture : Optical Flow 54
50 Example: Line Fitting [ Forsyth, Ponce(2003): Computer Vision, Ch. 15] Lecture : Optical Flow 55
51 Example: Line Fitting Zoom in from left image [ Forsyth, Ponce(2003): Computer Vision, Ch. 15] Lecture : Optical Flow 56
52 Example: Line Fitting How can we de-emphasize the influence of distant outliers? Assume non-gaussian distributions of error use non-quadratic error functions for the log-likelihood Pr (x nx, n y, d )= Pr (x i, y i n x, n y,d ) i 1 = exp( i σ 2 π 2 (nx x i +n y y i d ) 2σ 2 ) E= log( Pr (x n x, n y, d ))=C 1 (n x xi + n y y i d) +C 2 2 i Lecture : Optical Flow 57
53 Example: Line Fitting How can we de-emphasize the influence of distant outliers? Assume non-gaussian distributions of error use non-quadratic error functions for the log-likelihood Pr (x nx, n y, d )= Pr (x i, y i n x, n y,d ) i 1 = exp( i σ 2 π 2 (nx x i +n y y i d ) 2σ 2 ) E= log( Pr (x n x, n y, d ))=C 1 (n x xi + n y y i d) +C 2 2 i =C1 ρ(n x xi +n y y i d)+ C2 i Lecture : Optical Flow 58
54 Robust Error Function Desirable properties Inliers should be punished approximately quadratically they follow a Gaussian distribution Outliers should have small/ no influence they are more probable than under Gaussian distribution Lecture : Optical Flow 59
55 Robust Error Function Influence function Assume here differentiable error function How much differs the energy with the location of datapoint? ^ x)=ρ(nx x +n y y d ) E( ρ ϕ= x Lecture : Optical Flow 60
56 Error Functions Quadratic Error function 2 x ρ(x σ)= 2 2σ Influence function x ϕ( x σ )= 2 σ Probability distribution 2 1 x Pr (x σ)= exp ( ) 2 Z 2σ Lecture : Optical Flow 61
57 Error Functions Linear Error function ρ(x σ)= x Influence function Probability distribution ϕ( x σ )= sign( x) 1 Pr (x σ)= exp ( ρ) Z Lecture : Optical Flow 62
58 Error Functions Huber Error function Influence function x ϵ 2 ρ(x σ)= x ϵ + x ϵ 2ϵ 2 x else ϕ( x σ )= Probability distribution x ϵ sign( x) else Lecture : Optical Flow 1 Pr (x σ)= exp ( ρ) Z 63
59 Error Functions Lorentian Student's t distribution Error function 1 2 ρ(x σ)=log(1+ 2 x ) 2σ Influence function 2x ϕ( x σ )= 2 2 2σ + x Lecture : Optical Flow Probability distribution x 2 α Pr (x σ)=(1+ 2 ) 2σ 64
60 Error Functions: Comparisons quadratic Huber linear Lorentian 2 x ρ(x σ)= 2 2 σ +x Geman McClure Lecture : Optical Flow 65
61 Error Distributions: Comparisons quadratic linear Huber Lorentian Lecture : Optical Flow 66
62 Example: Line Fitting Correct choice of scale parameter σ Lecture : Optical Flow 67
63 Example: Line Fitting σ too small all data points are weighted down Lecture : Optical Flow 68
64 Example: Line Fitting σ too large all data points are considered Lecture : Optical Flow 69
65 Convex Functions Any local minimum of a convex function is a global minimum Lecture : Optical Flow 70
66 Quiz: Which Error Function is Convex? 2 1 quadratic Huber linear 3 ρ(x σ)=log(1+ Lorentian x) 2 2σ x2 ρ(x σ)= 2 2 σ +x Geman McClure 5 : None 6 : all Lecture : Optical Flow 71
67 Disparity Space Image C (x, y, d ) d y x 72
68 Disparity Space Image d y x 73
69 A Basic Stereo Algorithm: Blockmatching For each pixel For each disparity value Compare patch similarity Assign disparity with highest similarity Result: Noisy, no smooth surfaces 74
Dense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem
More informationCorrespondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri] Introduction Disparity: Informally: difference between two pictures Allows us to gain a strong
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction
More informationComputer Vision I - Filtering and Feature detection
Computer Vision I - Filtering and Feature detection Carsten Rother 30/10/2015 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image
More informationLecture 14: Basic Multi-View Geometry
Lecture 14: Basic Multi-View Geometry Stereo If I needed to find out how far point is away from me, I could use triangulation and two views scene point image plane optical center (Graphic from Khurram
More informationLUMS Mine Detector Project
LUMS Mine Detector Project Using visual information to control a robot (Hutchinson et al. 1996). Vision may or may not be used in the feedback loop. Visual (image based) features such as points, lines
More informationStereo Vision. MAN-522 Computer Vision
Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in
More informationMultiple View Geometry
Multiple View Geometry CS 6320, Spring 2013 Guest Lecture Marcel Prastawa adapted from Pollefeys, Shah, and Zisserman Single view computer vision Projective actions of cameras Camera callibration Photometric
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationAssignment 2: Stereo and 3D Reconstruction from Disparity
CS 6320, 3D Computer Vision Spring 2013, Prof. Guido Gerig Assignment 2: Stereo and 3D Reconstruction from Disparity Out: Mon Feb-11-2013 Due: Mon Feb-25-2013, midnight (theoretical and practical parts,
More informationComputer Vision Lecture 17
Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester
More informationComputer Vision Lecture 17
Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week
More informationThree-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras
Three-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inria.fr http://perception.inrialpes.fr/ Outline The geometry of active stereo.
More informationCS5670: Computer Vision
CS5670: Computer Vision Noah Snavely, Zhengqi Li Stereo Single image stereogram, by Niklas Een Mark Twain at Pool Table", no date, UCR Museum of Photography Stereo Given two images from different viewpoints
More informationComputer Vision I - Basics of Image Processing Part 1
Computer Vision I - Basics of Image Processing Part 1 Carsten Rother 28/10/2014 Computer Vision I: Basics of Image Processing Link to lectures Computer Vision I: Basics of Image Processing 28/10/2014 2
More informationCamera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration
Camera Calibration Jesus J Caban Schedule! Today:! Camera calibration! Wednesday:! Lecture: Motion & Optical Flow! Monday:! Lecture: Medical Imaging! Final presentations:! Nov 29 th : W. Griffin! Dec 1
More informationRectification and Disparity
Rectification and Disparity Nassir Navab Slides prepared by Christian Unger What is Stereo Vision? Introduction A technique aimed at inferring dense depth measurements efficiently using two cameras. Wide
More information3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,
3D Sensing and Reconstruction Readings: Ch 12: 12.5-6, Ch 13: 13.1-3, 13.9.4 Perspective Geometry Camera Model Stereo Triangulation 3D Reconstruction by Space Carving 3D Shape from X means getting 3D coordinates
More informationHartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne
Hartley - Zisserman reading club Part I: Hartley and Zisserman Appendix 6: Iterative estimation methods Part II: Zhengyou Zhang: A Flexible New Technique for Camera Calibration Presented by Daniel Fontijne
More informationFinal Exam Study Guide
Final Exam Study Guide Exam Window: 28th April, 12:00am EST to 30th April, 11:59pm EST Description As indicated in class the goal of the exam is to encourage you to review the material from the course.
More informationStereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision
Stereo Thurs Mar 30 Kristen Grauman UT Austin Outline Last time: Human stereopsis Epipolar geometry and the epipolar constraint Case example with parallel optical axes General case with calibrated cameras
More informationLecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza
Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time, ICRA 14, by Pizzoli, Forster, Scaramuzza [M. Pizzoli, C. Forster,
More informationData Term. Michael Bleyer LVA Stereo Vision
Data Term Michael Bleyer LVA Stereo Vision What happened last time? We have looked at our energy function: E ( D) = m( p, dp) + p I < p, q > N s( p, q) We have learned about an optimization algorithm that
More informationMidterm Examination CS 534: Computational Photography
Midterm Examination CS 534: Computational Photography November 3, 2016 NAME: Problem Score Max Score 1 6 2 8 3 9 4 12 5 4 6 13 7 7 8 6 9 9 10 6 11 14 12 6 Total 100 1 of 8 1. [6] (a) [3] What camera setting(s)
More informationarxiv: v1 [cs.cv] 28 Sep 2018
Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,
More informationRecap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?
Recap: Features and filters Epipolar geometry & stereo vision Tuesday, Oct 21 Kristen Grauman UT-Austin Transforming and describing images; textures, colors, edges Recap: Grouping & fitting Now: Multiple
More informationCapturing, Modeling, Rendering 3D Structures
Computer Vision Approach Capturing, Modeling, Rendering 3D Structures Calculate pixel correspondences and extract geometry Not robust Difficult to acquire illumination effects, e.g. specular highlights
More informationCS4495/6495 Introduction to Computer Vision. 3B-L3 Stereo correspondence
CS4495/6495 Introduction to Computer Vision 3B-L3 Stereo correspondence For now assume parallel image planes Assume parallel (co-planar) image planes Assume same focal lengths Assume epipolar lines are
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 12 130228 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Panoramas, Mosaics, Stitching Two View Geometry
More informationMotion and Tracking. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)
Motion and Tracking Andrea Torsello DAIS Università Ca Foscari via Torino 155, 30172 Mestre (VE) Motion Segmentation Segment the video into multiple coherently moving objects Motion and Perceptual Organization
More informationApplication questions. Theoretical questions
The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list
More informationComputer Vision, Lecture 11
Computer Vision, Lecture 11 Professor Hager http://www.cs.jhu.edu/~hager Computational Stereo Much of geometric vision is based on information from (or more) camera locations hard to recover 3D information
More informationGeometric Reconstruction Dense reconstruction of scene geometry
Lecture 5. Dense Reconstruction and Tracking with Real-Time Applications Part 2: Geometric Reconstruction Dr Richard Newcombe and Dr Steven Lovegrove Slide content developed from: [Newcombe, Dense Visual
More informationSIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014
SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image
More informationCHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION
CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION In this chapter we will discuss the process of disparity computation. It plays an important role in our caricature system because all 3D coordinates of nodes
More informationROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW
ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW Thorsten Thormählen, Hellward Broszio, Ingolf Wassermann thormae@tnt.uni-hannover.de University of Hannover, Information Technology Laboratory,
More informationBinocular Stereo Vision. System 6 Introduction Is there a Wedge in this 3D scene?
System 6 Introduction Is there a Wedge in this 3D scene? Binocular Stereo Vision Data a stereo pair of images! Given two 2D images of an object, how can we reconstruct 3D awareness of it? AV: 3D recognition
More informationStereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz
Stereo II CSE 576 Ali Farhadi Several slides from Larry Zitnick and Steve Seitz Camera parameters A camera is described by several parameters Translation T of the optical center from the origin of world
More informationRectification and Distortion Correction
Rectification and Distortion Correction Hagen Spies March 12, 2003 Computer Vision Laboratory Department of Electrical Engineering Linköping University, Sweden Contents Distortion Correction Rectification
More informationMulti-View Stereo for Static and Dynamic Scenes
Multi-View Stereo for Static and Dynamic Scenes Wolfgang Burgard Jan 6, 2010 Main references Yasutaka Furukawa and Jean Ponce, Accurate, Dense and Robust Multi-View Stereopsis, 2007 C.L. Zitnick, S.B.
More informationIntroduction to 3D Imaging: Perceiving 3D from 2D Images
Introduction to 3D Imaging: Perceiving 3D from 2D Images How can we derive 3D information from one or more 2D images? There have been 2 approaches: 1. intrinsic images: a 2D representation that stores
More informationLaser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR
Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and
More informationDirect Methods in Visual Odometry
Direct Methods in Visual Odometry July 24, 2017 Direct Methods in Visual Odometry July 24, 2017 1 / 47 Motivation for using Visual Odometry Wheel odometry is affected by wheel slip More accurate compared
More informationImage Transformations & Camera Calibration. Mašinska vizija, 2018.
Image Transformations & Camera Calibration Mašinska vizija, 2018. Image transformations What ve we learnt so far? Example 1 resize and rotate Open warp_affine_template.cpp Perform simple resize
More informationYihao Qian Team A: Aware Teammates: Amit Agarwal Harry Golash Menghan Zhang Zihao (Theo) Zhang ILR07 February.16, 2016
Yihao Qian Team A: Aware Teammates: Amit Agarwal Harry Golash Menghan Zhang Zihao (Theo) Zhang ILR07 February.16, 2016 Individual Progress During the last week, I was in charge of transplanting stereo
More informationComputer Vision I - Basics of Image Processing Part 2
Computer Vision I - Basics of Image Processing Part 2 Carsten Rother 07/11/2014 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image
More informationLecture 14: Computer Vision
CS/b: Artificial Intelligence II Prof. Olga Veksler Lecture : Computer Vision D shape from Images Stereo Reconstruction Many Slides are from Steve Seitz (UW), S. Narasimhan Outline Cues for D shape perception
More informationRobotics Programming Laboratory
Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car
More informationFinal Exam Study Guide CSE/EE 486 Fall 2007
Final Exam Study Guide CSE/EE 486 Fall 2007 Lecture 2 Intensity Sufaces and Gradients Image visualized as surface. Terrain concepts. Gradient of functions in 1D and 2D Numerical derivatives. Taylor series.
More informationThere are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...
STEREO VISION The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Bill Freeman and Antonio Torralba (MIT), including their own
More informationMultiple View Geometry
Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V
More informationFinal project bits and pieces
Final project bits and pieces The project is expected to take four weeks of time for up to four people. At 12 hours per week per person that comes out to: ~192 hours of work for a four person team. Capstone:
More informationMultiple Model Estimation : The EM Algorithm & Applications
Multiple Model Estimation : The EM Algorithm & Applications Princeton University COS 429 Lecture Dec. 4, 2008 Harpreet S. Sawhney hsawhney@sarnoff.com Plan IBR / Rendering applications of motion / pose
More informationStereo Vision A simple system. Dr. Gerhard Roth Winter 2012
Stereo Vision A simple system Dr. Gerhard Roth Winter 2012 Stereo Stereo Ability to infer information on the 3-D structure and distance of a scene from two or more images taken from different viewpoints
More informationLecture 10 Dense 3D Reconstruction
Institute of Informatics Institute of Neuroinformatics Lecture 10 Dense 3D Reconstruction Davide Scaramuzza 1 REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time M. Pizzoli, C. Forster,
More informationSUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS
SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS Cognitive Robotics Original: David G. Lowe, 004 Summary: Coen van Leeuwen, s1460919 Abstract: This article presents a method to extract
More informationDepth Estimation of Light Field Images by Optical Flow
Depth Estimation of Light Field Images by Optical Flow Marcus an mpanj@stanford.edu Abstract Light field images present a rich, compact, dataset of multiple images of a from slightly shifted viewpoints.
More informationStereo Vision II: Dense Stereo Matching
Stereo Vision II: Dense Stereo Matching Nassir Navab Slides prepared by Christian Unger Outline. Hardware. Challenges. Taxonomy of Stereo Matching. Analysis of Different Problems. Practical Considerations.
More informationFundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision
Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision What Happened Last Time? Human 3D perception (3D cinema) Computational stereo Intuitive explanation of what is meant by disparity Stereo matching
More informationPin Hole Cameras & Warp Functions
Pin Hole Cameras & Warp Functions Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Pinhole Camera. Homogenous Coordinates. Planar Warp Functions. Motivation Taken from: http://img.gawkerassets.com/img/18w7i1umpzoa9jpg/original.jpg
More informationINFO - H Pattern recognition and image analysis. Vision
INFO - H - 501 Pattern recognition and image analysis Vision Stereovision digital elevation model obstacle avoidance 3D model scanner human machine interface (HMI)... Stereovision image of the same point
More informationStereo vision. Many slides adapted from Steve Seitz
Stereo vision Many slides adapted from Steve Seitz What is stereo vision? Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape What is
More informationOmni-directional Multi-baseline Stereo without Similarity Measures
Omni-directional Multi-baseline Stereo without Similarity Measures Tomokazu Sato and Naokazu Yokoya Graduate School of Information Science, Nara Institute of Science and Technology 8916-5 Takayama, Ikoma,
More informationWhat have we leaned so far?
What have we leaned so far? Camera structure Eye structure Project 1: High Dynamic Range Imaging What have we learned so far? Image Filtering Image Warping Camera Projection Model Project 2: Panoramic
More information3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava
3D Computer Vision Dense 3D Reconstruction II Prof. Didier Stricker Christiano Gava Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
More informationME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies"
ME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies" lhm@jpl.nasa.gov, 818-354-3722" Announcements" First homework grading is done! Second homework is due
More informationLecture 16: Computer Vision
CS4442/9542b: Artificial Intelligence II Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field
More informationStructured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov
Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter
More informationToday. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography
Computational Photography Matthias Zwicker University of Bern Fall 2009 Today From 2D to 3D using multiple views Introduction Geometry of two views Stereo matching Other applications Multiview geometry
More informationDepth Measurement. 2/27/12 ECEn 631
Correspondence Depth Measurement Visual areas must be be seen by both cameras If the physical coordinates of the cameras or sizes of objects are known, depth can be measured using triangulated disparity
More informationStereo imaging ideal geometry
Stereo imaging ideal geometry (X,Y,Z) Z f (x L,y L ) f (x R,y R ) Optical axes are parallel Optical axes separated by baseline, b. Line connecting lens centers is perpendicular to the optical axis, and
More informationDense Image-based Motion Estimation Algorithms & Optical Flow
Dense mage-based Motion Estimation Algorithms & Optical Flow Video A video is a sequence of frames captured at different times The video data is a function of v time (t) v space (x,y) ntroduction to motion
More informationFeature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking
Feature descriptors Alain Pagani Prof. Didier Stricker Computer Vision: Object and People Tracking 1 Overview Previous lectures: Feature extraction Today: Gradiant/edge Points (Kanade-Tomasi + Harris)
More informationOutline. ETN-FPI Training School on Plenoptic Sensing
Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration
More informationChaplin, Modern Times, 1936
Chaplin, Modern Times, 1936 [A Bucket of Water and a Glass Matte: Special Effects in Modern Times; bonus feature on The Criterion Collection set] Multi-view geometry problems Structure: Given projections
More informationPin Hole Cameras & Warp Functions
Pin Hole Cameras & Warp Functions Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Pinhole Camera. Homogenous Coordinates. Planar Warp Functions. Example of SLAM for AR Taken from:
More informationScalable geometric calibration for multi-view camera arrays
Scalable geometric calibration for multi-view camera arrays Bernhard Blaschitz, Doris Antensteiner, Svorad Štolc Bernhard.Blaschitz@ait.ac.at AIT Austrian Institute of Technology GmbH Intelligent Vision
More informationMultiview Stereo COSC450. Lecture 8
Multiview Stereo COSC450 Lecture 8 Stereo Vision So Far Stereo and epipolar geometry Fundamental matrix captures geometry 8-point algorithm Essential matrix with calibrated cameras 5-point algorithm Intersect
More informationObject Recognition with Invariant Features
Object Recognition with Invariant Features Definition: Identify objects or scenes and determine their pose and model parameters Applications Industrial automation and inspection Mobile robots, toys, user
More informationIndex. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263
Index 3D reconstruction, 125 5+1-point algorithm, 284 5-point algorithm, 270 7-point algorithm, 265 8-point algorithm, 263 affine point, 45 affine transformation, 57 affine transformation group, 57 affine
More informationMulti-stable Perception. Necker Cube
Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuyuki Kayahara Multiple view geometry Stereo vision Epipolar geometry Lowe Hartley and Zisserman Depth map extraction Essential matrix
More informationImage Transfer Methods. Satya Prakash Mallick Jan 28 th, 2003
Image Transfer Methods Satya Prakash Mallick Jan 28 th, 2003 Objective Given two or more images of the same scene, the objective is to synthesize a novel view of the scene from a view point where there
More informationLecture'9'&'10:'' Stereo'Vision'
Lecture'9'&'10:'' Stereo'Vision' Dr.'Juan'Carlos'Niebles' Stanford'AI'Lab' ' Professor'FeiAFei'Li' Stanford'Vision'Lab' 1' Dimensionality'ReducIon'Machine'(3D'to'2D)' 3D world 2D image Point of observation
More informationProject Updates Short lecture Volumetric Modeling +2 papers
Volumetric Modeling Schedule (tentative) Feb 20 Feb 27 Mar 5 Introduction Lecture: Geometry, Camera Model, Calibration Lecture: Features, Tracking/Matching Mar 12 Mar 19 Mar 26 Apr 2 Apr 9 Apr 16 Apr 23
More informationGOPRO CAMERAS MATRIX AND DEPTH MAP IN COMPUTER VISION
Tutors : Mr. Yannick Berthoumieu Mrs. Mireille El Gheche GOPRO CAMERAS MATRIX AND DEPTH MAP IN COMPUTER VISION Delmi Elias Kangou Ngoma Joseph Le Goff Baptiste Naji Mohammed Hamza Maamri Kenza Randriamanga
More informationProject 4 Results. Representation. Data. Learning. Zachary, Hung-I, Paul, Emanuel. SIFT and HoG are popular and successful.
Project 4 Results Representation SIFT and HoG are popular and successful. Data Hugely varying results from hard mining. Learning Non-linear classifier usually better. Zachary, Hung-I, Paul, Emanuel Project
More informationAccurate and Dense Wide-Baseline Stereo Matching Using SW-POC
Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Shuji Sakai, Koichi Ito, Takafumi Aoki Graduate School of Information Sciences, Tohoku University, Sendai, 980 8579, Japan Email: sakai@aoki.ecei.tohoku.ac.jp
More information2D Image Processing Feature Descriptors
2D Image Processing Feature Descriptors Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Overview
More informationComputer Vision Projective Geometry and Calibration. Pinhole cameras
Computer Vision Projective Geometry and Calibration Professor Hager http://www.cs.jhu.edu/~hager Jason Corso http://www.cs.jhu.edu/~jcorso. Pinhole cameras Abstract camera model - box with a small hole
More informationComputer Vision. Geometric Camera Calibration. Samer M Abdallah, PhD
Computer Vision Samer M Abdallah, PhD Faculty of Engineering and Architecture American University of Beirut Beirut, Lebanon Geometric Camera Calibration September 2, 2004 1 Computer Vision Geometric Camera
More informationStereo matching. Francesco Isgrò. 3D Reconstruction and Stereo p.1/21
Stereo matching Francesco Isgrò 3D Reconstruction and Stereo p.1/21 Structure of a stereo vision system Extract interesting point from each image Determine a set of matching points Compute the fundamental
More informationFast Stereo Matching of Feature Links
Fast Stereo Matching of Feature Links 011.05.19 Chang-il, Kim Introduction Stereo matching? interesting topics of computer vision researches To determine a disparity between stereo images A fundamental
More informationRange Sensors (time of flight) (1)
Range Sensors (time of flight) (1) Large range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic sensors, infra-red sensors
More informationA Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation
A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation Alexander Andreopoulos, Hirak J. Kashyap, Tapan K. Nayak, Arnon Amir, Myron D. Flickner IBM Research March 25,
More information3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.
3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction
More informationAn investigation into stereo algorithms: An emphasis on local-matching. Thulani Ndhlovu
An investigation into stereo algorithms: An emphasis on local-matching Thulani Ndhlovu Submitted to the Department of Electrical Engineering, University of Cape Town, in fullfillment of the requirements
More informationFeature Based Registration - Image Alignment
Feature Based Registration - Image Alignment Image Registration Image registration is the process of estimating an optimal transformation between two or more images. Many slides from Alexei Efros http://graphics.cs.cmu.edu/courses/15-463/2007_fall/463.html
More informationComputer Vision I. Announcement. Stereo Vision Outline. Stereo II. CSE252A Lecture 15
Announcement Stereo II CSE252A Lecture 15 HW3 assigned No class on Thursday 12/6 Extra class on Tuesday 12/4 at 6:30PM in WLH Room 2112 Mars Exploratory Rovers: Spirit and Opportunity Stereo Vision Outline
More informationCameras and Stereo CSE 455. Linda Shapiro
Cameras and Stereo CSE 455 Linda Shapiro 1 Müller-Lyer Illusion http://www.michaelbach.de/ot/sze_muelue/index.html What do you know about perspective projection? Vertical lines? Other lines? 2 Image formation
More informationCamera Drones Lecture 3 3D data generation
Camera Drones Lecture 3 3D data generation Ass.Prof. Friedrich Fraundorfer WS 2017 Outline SfM introduction SfM concept Feature matching Camera pose estimation Bundle adjustment Dense matching Data products
More information