Subpixel accurate refinement of disparity maps using stereo correspondences

Similar documents
Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

EE795: Computer Vision and Intelligent Systems

Correspondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]

Segmentation Based Stereo. Michael Bleyer LVA Stereo Vision

Segmentation and Tracking of Partial Planar Templates

Data Term. Michael Bleyer LVA Stereo Vision

Multiple View Geometry

CONTENTS. High-Accuracy Stereo Depth Maps Using Structured Light. Yeojin Yoon

Geometric Reconstruction Dense reconstruction of scene geometry

Accurate Image Registration from Local Phase Information

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Capturing, Modeling, Rendering 3D Structures

Using temporal seeding to constrain the disparity search range in stereo matching

Step-by-Step Model Buidling

Stereo Vision II: Dense Stereo Matching

Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza

Supplemental Material: A Dataset and Evaluation Methodology for Depth Estimation on 4D Light Fields

Depth Estimation with a Plenoptic Camera

What have we leaned so far?

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

Learning Two-View Stereo Matching

Epipolar Geometry and Stereo Vision

Robert Collins CSE486, Penn State. Lecture 09: Stereo Algorithms

Filter Flow: Supplemental Material

CS4495/6495 Introduction to Computer Vision. 3B-L3 Stereo correspondence

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Towards a visual perception system for LNG pipe inspection

Epipolar Geometry and Stereo Vision

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Image processing and features

1 (5 max) 2 (10 max) 3 (20 max) 4 (30 max) 5 (10 max) 6 (15 extra max) total (75 max + 15 extra)

Lecture 10 Dense 3D Reconstruction

An investigation into stereo algorithms: An emphasis on local-matching. Thulani Ndhlovu

Dense 3D Reconstruction. Christiano Gava

COS Lecture 10 Autonomous Robot Navigation

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR

CS5670: Computer Vision

Hybrid Rendering for Collaborative, Immersive Virtual Environments

CS 4495/7495 Computer Vision Frank Dellaert, Fall 07. Dense Stereo Some Slides by Forsyth & Ponce, Jim Rehg, Sing Bing Kang

Computer Vision I - Appearance-based Matching and Projective Geometry

Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images

3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava

3D RECONSTRUCTION FROM STEREO/ RANGE IMAGES

Symmetric Sub-Pixel Stereo Matching

Stereo Vision A simple system. Dr. Gerhard Roth Winter 2012

Introduction à la vision artificielle X

Stereo imaging ideal geometry

MAPI Computer Vision. Multiple View Geometry

Stereo vision. Many slides adapted from Steve Seitz

Scaled representations

Comparison between Motion Analysis and Stereo

Dense 3D Reconstruction. Christiano Gava

CS 787: Assignment 4, Stereo Vision: Block Matching and Dynamic Programming Due: 12:00noon, Fri. Mar. 30, 2007.

Limited view X-ray CT for dimensional analysis

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Stereo and Epipolar geometry

Multi-View Stereo for Static and Dynamic Scenes

Local Feature Detectors

Edge Detection. Today s reading. Cipolla & Gee on edge detection (available online) From Sandlot Science

Experiments with Edge Detection using One-dimensional Surface Fitting

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

Motion Estimation. There are three main types (or applications) of motion estimation:

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Segmentation and Grouping

A MULTI-RESOLUTION APPROACH TO DEPTH FIELD ESTIMATION IN DENSE IMAGE ARRAYS F. Battisti, M. Brizzi, M. Carli, A. Neri

Real-Time Disparity Map Computation Based On Disparity Space Image

Range Sensors (time of flight) (1)

Midterm Examination CS 534: Computational Photography

3D reconstruction class 11

Final Exam Study Guide CSE/EE 486 Fall 2007

BIL Computer Vision Apr 16, 2014

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1. Stereo Processing by Semi-Global Matching and Mutual Information. Heiko Hirschmüller

Reliability Based Cross Trilateral Filtering for Depth Map Refinement

Fundamentals of Digital Image Processing

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

Advanced Computer Graphics. Aliasing. Matthias Teschner. Computer Science Department University of Freiburg

Final Review CMSC 733 Fall 2014

Multiple View Geometry

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Edge Detection Techniques in Processing Digital Images: Investigation of Canny Algorithm and Gabor Method

Homographies and RANSAC

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Supplementary Material for A Locally Linear Regression Model for Boundary Preserving Regularization in Stereo Matching

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation

Outline 7/2/201011/6/

Multi-View 3D-Reconstruction

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

ME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies"

Digital Image Processing

AIT Inline Computational Imaging: Geometric calibration and image rectification

Lecture 10: Multi view geometry

Edge and corner detection

Static Scene Reconstruction

Stereo Wrap + Motion. Computer Vision I. CSE252A Lecture 17

School of Computing University of Utah

Multi-view stereo. Many slides adapted from S. Seitz

segments. The geometrical relationship of adjacent planes such as parallelism and intersection is employed for determination of whether two planes sha

Vision 3D articielle Disparity maps, correlation

Project 2 due today Project 3 out today. Readings Szeliski, Chapter 10 (through 10.5)

Transcription:

Subpixel accurate refinement of disparity maps using stereo correspondences Matthias Demant Lehrstuhl für Mustererkennung, Universität Freiburg

Outline 1 Introduction and Overview 2 Refining the Cost Volume 3 Patch Correlation Algorithm 4 Experimental Validation Test Datasets Evaluation Methodology 5 Experimental Results Parabolic Fitting Patch Correlation 6 Conclusion and Outlook 2 von 43

Outline 1 Introduction and Overview 2 Refining the Cost Volume 3 Patch Correlation Algorithm 4 Experimental Validation Test Datasets Evaluation Methodology 5 Experimental Results Parabolic Fitting Patch Correlation 6 Conclusion and Outlook

Introduction and Overview Problem statement (a) Original image (b) Reference image (c) Integer map disparity (d) Subpixel accurate disparity map Figure: A typical example for integer disparities. The disparity map is traversed by stripes, the staircase effect. 4 von 43

Introduction and Overview The standard stereo configuration Figure: The standard stereo geometry system. Camera centers differ by a pure translation. Standard stereo case - Corresponding points have identical y-coordinates. - Disparity = offset between corresponding points along scanline - Depth is inverse proportional to disparity. 5 von 43

Introduction and Overview General approach with two steps 1 Pixel accurate disparity estimation Required: Cost Volume Search disparity with minimum costs Disparity map (+- 2 pixel) Occlusion handling 2 Subpixel refinement 6 von 43

Outline 1 Introduction and Overview 2 Refining the Cost Volume 3 Patch Correlation Algorithm 4 Experimental Validation Test Datasets Evaluation Methodology 5 Experimental Results Parabolic Fitting Patch Correlation 6 Conclusion and Outlook

Refining the Cost Volume Disparity Space Image Wanted: Disparity map d(x, y) with I L (x, y) I R (x d(x, y), y) Cost Volume: DSI(x, y, d) = (I L I R (x d, y)) 2 Aggregation: with block, Gaussian,.. in 2D or 3D reduce noise + environment informations Start correspondence search Assumptions: ordering constraint smoothness constraint... 8 von 43

Refining the Cost Volume Continuous Disparity Space Image Image sensor Integrates intensities over 1 Pixel Sampling insensitive Compare Intervals [x 0.5, x, x + 0.5] measure instead of Pixels Wish: Parabolic fitting Sampling theorem Continuous DSI Continuous values after correspondence search Samples of 0.5 pixel in x- and d-direction necessary to properly reconstruct the DSI 9 von 43

Refining the Cost Volume Parabolic Fitting Algorithm 10 von 43

Refining the Cost Volume Parabolic Fitting Algorithm 10 von 43

Refining the Cost Volume Parabolic Fitting Algorithm Figure: Fitting a parabola to 5 or 11 points around a value in the estimated cost volume (DSI). The red point specifies the calculated minimum. The images at the bottom are examples for unrefined values. The values are out of range of the fitted points or maxima. 11 von 43

Outline 1 Introduction and Overview 2 Refining the Cost Volume 3 Patch Correlation Algorithm 4 Experimental Validation Test Datasets Evaluation Methodology 5 Experimental Results Parabolic Fitting Patch Correlation 6 Conclusion and Outlook

Patch Correlation Algorithm Patch Correlation Algorithm (a) 1D patch (b) 2D patch Figure: Shifting the original patch over the reference patch. Pixel accurate disparity map initial correspondence. Compare patch in left image with corresponding patch in right image. Shift and compare. The signals stay constant. 13 von 43

Patch Correlation Algorithm The Gabor Wavelet Gaussian multiplied with complex exponential Frequency domain: Gaussian at modulation frequency of complex exponential Acts like a bandpass filter Change modulation frequency change position in frequency domain Windowing + expansion in frequency domain 14 von 43

Patch Correlation Algorithm The Gabor Wavelet Gaussian multiplied with complex exponential Frequency domain: Gaussian at modulation frequency of complex exponential Acts like a bandpass filter Change modulation frequency change position in frequency domain Windowing + expansion in frequency domain 14 von 43

Patch Correlation Algorithm t t corresponding points P t Patch in Image I at position t, Q t Patch in Image J at position t 1 Convolve images with a limited number of gabor filters Patches expanded in frequency domain a t (k) = P t (x)e ikx dx = I (x t) g(x)e ikx }{{} gabor wavelet dx 2 Correlate and shift patches according to shift theorem P t δt, Q t k a t (k)e ikδt b t (k) = k c t (k)e ikδt 3 Find maximum of the correlation function 4 Confidence estimation with normalized correlation 15 von 43

Patch Correlation Algorithm Maximum Detection Correlation Function P t δt, Q t k a t (k)e ikδt b t (k) = k c t (k)e ikδt 1 Discrete Fourier Transform + Parabolic Refinement 2 Iterative Techniques Gradient Ascent Newton Rhapson Modified Newton Rhapson 16 von 43

Patch Correlation Algorithm Maximum Detection Correlation Function P t δt, Q t k a t (k)e ikδt b t (k) = k c t (k)e ikδt 1 Discrete Fourier Transform + Parabolic Refinement 2 Iterative Techniques Gradient Ascent Newton Rhapson Modified Newton Rhapson 16 von 43

Patch Correlation Algorithm Overview Figure: The Patch Correlation Algorithm 17 von 43

Outline 1 Introduction and Overview 2 Refining the Cost Volume 3 Patch Correlation Algorithm 4 Experimental Validation Test Datasets Evaluation Methodology 5 Experimental Results Parabolic Fitting Patch Correlation 6 Conclusion and Outlook

Experimental Validation Test Datasets Real Data (a) Original image (b) Reference image (c) Ground truth (d) Original image (e) Reference image (f) Ground truth Figure: The Cones and Teddy images with disparity map. 19 von 43

Experimental Validation Test Datasets Starting Point Figure: The integer disparity map (left) with added Gaussian noise (right). Our starting point of the refinement process. Integer disparity map + Gaussian noise σ = 0.7: Values in [ 2, +2] Range [+0.5, +1.5] with little Gaussian noise Middlebury stereo benchmark: Far behind last place. 20 von 43

Experimental Validation Test Datasets PovRay Scenes (a) Original image (b) Reference image (c) Depth map with 3 256 values (d) Synthesized View Figure: A povray scene with 3 256 depth values. Due to unpredictable discontinuities in the depth map the synthesized view in figure 8(d) is not correct (red box). 21 von 43

Experimental Validation Evaluation Methodology Quality Metrics 1 RMS (root-mean-square) error between the computed disparity map and the ground truth map 2 BMP Percentage of bad matching pixels 3 RMS error between synthetic views (a) Original image (b) Forward warped image (c) Reference image Figure: Using the disparity map to create different views. 9(b) The reference image in coordinates of the original image. 22 von 43

Experimental Validation Evaluation Methodology (a) Ground truth disparity map (b) Occluded regions (black) (c) Textureless regions (white) (d) Discontinuity regions (white) 23 von 43

Outline 1 Introduction and Overview 2 Refining the Cost Volume 3 Patch Correlation Algorithm 4 Experimental Validation Test Datasets Evaluation Methodology 5 Experimental Results Parabolic Fitting Patch Correlation 6 Conclusion and Outlook

Experimental Results Parabolic Fitting Parabolic Fitting Best results with Interval Matching Gaussian 2D aggregation (size 3,5 Pixel) Parabola fitted through 9 points Figure: A complex scene, the refined line was achieved with a parabolic fitting to 9 points. The DSI was aggregated with a Gaussian 2D window using interval differences ID. 25 von 43

Experimental Results Patch Correlation Results in different image regions (a) Initial disp.map (b) Result (c) Result (Gradient) 26 von 43

Experimental Results Patch Correlation Results in different image regions (f) Initial disp.map (g) Result (h) Result (Gradient) 26 von 43

Experimental Results Patch Correlation Results - Cones line profile Figure: The refined line with the patch correlation algorithm. 27 von 43

Experimental Results Patch Correlation Results - Cones line profile Figure: The refined line with the patch correlation algorithm. 27 von 43

Experimental Results Patch Correlation Results in different image regions (a) O Not occluded regions (b) T Textureless regions (c) D Discontinuity regions (d) D Non-discontinuity regions Figure: Percentage of bad matching pixels with an error threshold of 0.7 pixel. 28 von 43

Experimental Results Patch Correlation Patch Correlation - confidence and window size (a) RMS (b) %BMP, Threshold 0.5 Pixel Figure: Changed Confidence and window size. Initial disparity map with gaussian noise: σ = 0.7 29 von 43

Experimental Results Patch Correlation Iterative Techniques Figure: Bad matching pixels and iterations. The convergence of the different iterative techniques. The gradient ascent method converges in most cases in the global maximum. 30 von 43

Experimental Results Patch Correlation Middlebury Ranking Figure: A screenshot of the Middlebury ranking with an error threshold of 0.75. 31 von 43

Outline 1 Introduction and Overview 2 Refining the Cost Volume 3 Patch Correlation Algorithm 4 Experimental Validation Test Datasets Evaluation Methodology 5 Experimental Results Parabolic Fitting Patch Correlation 6 Conclusion and Outlook

Conclusion and Outlook Conclusion and Outlook Initial estimation should be replaced by real results. Synthetic data showed unpredictable errors. Applicable with small disparity ranges Parabolic fitting to DSI is appropriated for small errors < 1 pixel. Intervall Matching better Patch correlation: 75% closer than 0.5 Pixels Errors in discontinuity regions. Scene information: Add semi-global character using informations about sharp areas or scene geometry. 33 von 43

Conclusion and Outlook Blurring Assumption: Well-known scene geometry Question: How can we emphasize small structures? Figure: With a well known scene setup blurred image regions can be detected (blue). In sharp areas the high frequencies can be emphasized applying the Laplacian of Gaussian. The Patch Correlation algorithm (green) can be extended with little additional costs. 34 von 43

Dilation between patches Points on planes + camera shift + plane orientation dilation between pairs of corresponding points (blue) The Patch Correlation algorithm (green) extracts both patches. One of the extracted patches can be scaled according to the dilation.

Thank you for your attention.

Conclusion and Outlook Model of the Point-spread function Assumption: Well-known scene geometry Question: How and when can we emphasize small structures? Point spread function: The brightness of the blurred area can be approximated by a two dimensional Gaussian distribution. Brightness distribution depends on distance. A high σ value affects like a low-pass filter. Emphasize small stuctures in sharp areas: LoG Easy extandable to patch correlation algorithm 37 von 43

Conclusion and Outlook Figure: With a well known scene setup blurred image regions can be detected (blue). In sharp areas the high frequencies can be emphasized applying the Laplacian of Gaussian. The Patch Correlation algorithm (green) can be extended with little additional costs. 38 von 43

Conclusion and Outlook Homography induced by a plane Assumption: Standard stereo configuration Question: How can we classify transformations between corresponding points on same world plane? x J = H JI (π)x I Back-trace ray C x via P + I to π Project x π via P J to J ( H JI (π) = K J I + c p ) T K 1 I. with c = ( ) T c x c y c z and π = ( p 1 ) T The calibration matrices K I, K 1 I and K J are affine transformations. The center part ( I + cp ) T is affine, if c z = 0. In the standard stereo case H JI (π) is an affine transformation (c y = c z = 0). 39 von 43

Conclusion and Outlook Homography induced by a plane Assumption: Standard stereo configuration Question: How can we classify transformations between corresponding points on same world plane? x J = H JI (π)x I Back-trace ray C x via P + I to π Project x π via P J to J ( H JI (π) = K J I + c p ) T K 1 I. with c = ( ) T c x c y c z and π = ( p 1 ) T The calibration matrices K I, K 1 I and K J are affine transformations. The center part ( I + cp ) T is affine, if c z = 0. In the standard stereo case H JI (π) is an affine transformation (c y = c z = 0). 39 von 43

Conclusion and Outlook Homography induced by a plane Assumption: Standard stereo configuration Question: How can we classify transformations between corresponding points on same world plane? x J = H JI (π)x I Back-trace ray C x via P + I to π Project x π via P J to J ( H JI (π) = K J I + c p ) T K 1 I. with c = ( ) T c x c y c z and π = ( p 1 ) T The calibration matrices K I, K 1 I and K J are affine transformations. The center part ( I + cp ) T is affine, if c z = 0. In the standard stereo case H JI (π) is an affine transformation (c y = c z = 0). 39 von 43

Conclusion and Outlook Homography induced by a plane Assumption: Standard stereo configuration Question: How can we classify transformations between corresponding points on same world plane? x J = H JI (π)x I Back-trace ray C x via P + I to π Project x π via P J to J ( H JI (π) = K J I + c p ) T K 1 I. with c = ( ) T c x c y c z and π = ( p 1 ) T The calibration matrices K I, K 1 I and K J are affine transformations. The center part ( I + cp ) T is affine, if c z = 0. In the standard stereo case H JI (π) is an affine transformation (c y = c z = 0). 39 von 43

Conclusion and Outlook Homography induced by a plane Assumption: Standard stereo configuration Question: How can we classify transformations between corresponding points on same world plane? x J = H JI (π)x I Back-trace ray C x via P + I to π Project x π via P J to J ( H JI (π) = K J I + c p ) T K 1 I. with c = ( ) T c x c y c z and π = ( p 1 ) T The calibration matrices K I, K 1 I and K J are affine transformations. The center part ( I + cp ) T is affine, if c z = 0. In the standard stereo case H JI (π) is an affine transformation (c y = c z = 0). 39 von 43

Conclusion and Outlook Homography induced by a plane Assumption: Standard stereo configuration Question: How can we classify transformations between corresponding points on same world plane? x J = H JI (π)x I Back-trace ray C x via P + I to π Project x π via P J to J ( H JI (π) = K J I + c p ) T K 1 I. with c = ( ) T c x c y c z and π = ( p 1 ) T The calibration matrices K I, K 1 I and K J are affine transformations. The center part ( I + cp ) T is affine, if c z = 0. In the standard stereo case H JI (π) is an affine transformation (c y = c z = 0). 39 von 43

Conclusion and Outlook Scene planes, disparities and patches Definition: Iso-disparity surfaces have identical size in both images. Assumption: Points on same world plane, simplified calibration matrices Question: How does texture change, depending on plane orientation? x J = H JI (π)x I x J x J = H JI (π) (x I x I ) For corresponding points: x J x J = (1 + c x p x )(x I x I ) Iso-disparity surfaces are fronto parallel planes. All patches on these surfaces keep the same patch size in both images. 40 von 43

Conclusion and Outlook Scene planes, disparities and patches Definition: Iso-disparity surfaces have identical size in both images. Assumption: Points on same world plane, simplified calibration matrices Question: How does texture change, depending on plane orientation? x J = H JI (π)x I x J x J = H JI (π) (x I x I ) For corresponding points: x J x J = (1 + c x p x )(x I x I ) Iso-disparity surfaces are fronto parallel planes. All patches on these surfaces keep the same patch size in both images. 40 von 43

Conclusion and Outlook The viewable window on fronto-parallel planes (a) Two cameras and isodisparity surfaces (b) Relation between disparity and the size of the viewable window. Figure: The viewable window decreases with increasing disparity. 41 von 43

Conclusion and Outlook Isodisparity surfaces and focal length Figure: Iso-disparity surfaces in the standard stereo configuration with different focal length. The center figure represents equal focal length. 42 von 43

Conclusion and Outlook Applicability in a well-known camera setup Figure: With a known camera shift and the initial disparity map a dilation between patches can be corrected (blue). The Patch Correlation algorithm (green) extracts both patches. One of the extracted patches can be scaled according to the dilation. 43 von 43