PART IV: RS & the Kinect

Similar documents
Forward interpolation on the GPU

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

Information page for written examinations at Linköping University TER2

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Overview. Augmented reality and applications Marker-based augmented reality. Camera model. Binary markers Textured planar markers

Lecture 9: Epipolar Geometry

Depth Sensors Kinect V2 A. Fornaser

calibrated coordinates Linear transformation pixel coordinates

Active Stereo Vision. COMP 4900D Winter 2012 Gerhard Roth

CS 231A: Computer Vision (Winter 2018) Problem Set 2

Epipolar Geometry and Stereo Vision

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

3D Environment Reconstruction

BIL Computer Vision Apr 16, 2014

Handling of Rolling Shutter Effects in Monocular Semi-Dense SLAM Algorithms

High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond

Multiview Stereo COSC450. Lecture 8

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

Scene Reconstruction from Uncontrolled Motion using a Low Cost 3D Sensor

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

MERGING POINT CLOUDS FROM MULTIPLE KINECTS. Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia

Rectification and Distortion Correction

Monocular Tracking and Reconstruction in Non-Rigid Environments

CSCI 5980: Assignment #3 Homography

Project: Camera Rectification and Structure from Motion

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Epipolar Geometry CSE P576. Dr. Matthew Brown

Lecture'9'&'10:'' Stereo'Vision'

Epipolar Geometry and Stereo Vision

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

An idea which can be used once is a trick. If it can be used more than once it becomes a method

EE795: Computer Vision and Intelligent Systems

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

Homographies and RANSAC

1 CSE 252A Computer Vision I Fall 2017

Structure from motion

CSCI 5980/8980: Assignment #4. Fundamental Matrix

Step-by-Step Model Buidling

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches

High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Structure from Motion and Multi- view Geometry. Last lecture

3D Computer Vision. Structure from Motion. Prof. Didier Stricker

Lecture 8 Active stereo & Volumetric stereo

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.

Dense 3D Reconstruction. Christiano Gava

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Live Metric 3D Reconstruction on Mobile Phones ICCV 2013

Advanced Imaging Applications on Smart-phones Convergence of General-purpose computing, Graphics acceleration, and Sensors

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

LUMS Mine Detector Project

Ceilbot vision and mapping system

Outline. 1 Why we re interested in Real-Time tracking and mapping. 3 Kinect Fusion System Overview. 4 Real-time Surface Mapping

EE795: Computer Vision and Intelligent Systems

Lecture 14: Basic Multi-View Geometry

Lecture 9 & 10: Stereo Vision

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Last lecture. Passive Stereo Spacetime Stereo

Lecture 8 Active stereo & Volumetric stereo

Geometry for Computer Vision

Dense 3D Reconstruction. Christiano Gava

Rigid ICP registration with Kinect

A Summary of Projective Geometry

THREE PRE-PROCESSING STEPS TO INCREASE THE QUALITY OF KINECT RANGE DATA

Handheld Augmented Reality. Reto Lindegger

Binocular Stereo Vision. System 6 Introduction Is there a Wedge in this 3D scene?

Structure from Motion

3D Reconstruction of a Hopkins Landmark

Application questions. Theoretical questions

TEXTURE OVERLAY ONTO NON-RIGID SURFACE USING COMMODITY DEPTH CAMERA

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?

A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation

Efficient Video Rectification and Stabilisation for Cell-Phones

Lecture 6 Stereo Systems Multi-view geometry

Fundamental Matrix & Structure from Motion

The main problem of photogrammetry

Object Reconstruction

Mobile Point Fusion. Real-time 3d surface reconstruction out of depth images on a mobile platform

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo --

3D Time-of-Flight Image Sensor Solutions for Mobile Devices

Augmented Reality II - Camera Calibration - Gudrun Klinker May 11, 2004

CS 395T Numerical Optimization for Graphics and AI (3D Vision) Qixing Huang August 29 th 2018

KinectFusion: Real-Time Dense Surface Mapping and Tracking

Lecture 10: Multi view geometry

Camera Model and Calibration

CVPR 2014 Visual SLAM Tutorial Kintinuous

Agenda. DLP 3D scanning Introduction DLP 3D scanning SDK Introduction Advance features for existing SDK

Fast Motion Deblurring for Feature Detection and Matching Using Inertial Measurements

Monocular Visual-Inertial SLAM. Shaojie Shen Assistant Professor, HKUST Director, HKUST-DJI Joint Innovation Laboratory

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --

Stereo and Epipolar geometry

Introduction to Computer Vision

Registration of Moving Surfaces by Means of One-Shot Laser Projection

Stereo Vision. MAN-522 Computer Vision

Transcription:

Computer Vision on Rolling Shutter Cameras PART IV: RS & the Kinect Per-Erik Forssén, Erik Ringaby, Johan Hedborg Computer Vision Laboratory Dept. of Electrical Engineering Linköping University

Tutorial overview 1:30-2:00pm Introduction 2:00-2:15pm Rolling Shutter Geometry 2:15-3:00pm Rectification and Stabilisation Per-Erik Per-Erik Erik 3:00-3:30pm Break 3:30-3:45pm Rolling Shutter and the Kinect Erik 3:45-4:30pm Structure from Motion Johan 2

The Kinect sensor Designed for player interaction with the Xbox 360 You are the Controller 3D scanner Accelerometer sensor Skeletal tracking Stationary in your living room 3

The Kinect sensor The H/W gained popularity in the research community Quasi-dense depth maps in 30 Hz Cheap (~100 USD) 4

The Kinect sensor Range estimation by triangulation A - structured NIR laser projector B - CMOS Colour camera C - CMOS NIR camera 5

RS on Kinect Both Kinect cameras make use of rolling shutters Not a big problem when it is used for gaming: Stationary in your living room People are moving quite slowly far away from the sensor It is however also popular to use the Kinect sensor on mobile platforms RS problems 6

Kinect footage 7

Kinect footage Augmented reality by KinectFusion [Izadi et al. SIGGRAPH 11] 8

RS on Kinect A similar approach as for the video case, but now we also have the depth 3D point correspondences enables us to estimate the full 3D motion (rotation and translation) The 3D point cloud can be rectified [Ringaby & Forssén, ICCV 11] 9

Synchronization problem RGB and depth map: - Different readout time - Different fields-of-view - Correspondence problematic under camera motion 10

Synchronization problem RGB and depth map: - Different readout time - Different fields-of-view - Correspondence problematic under camera motion Solution: Use NIR images 11

NIR images NIR and depth map from the same sensor NIR camera uses shorter shutter speeds, less motion blur Drawback, we need to suppress the structured light pattern (SLP) 12

Suppressing the SLP Source code available at: http://users.isy.liu.se/cvl/perfo/software/ 13

Outlier rejection Optimisation sensitive to point corr. outliers Three rejection steps: - KLT cross-checking - Depth map edge detection - Procrustes RANSAC 14

Depth map edge detection 15

Procrustes RANSAC Kinect depth map is noisy Estimate global translation and rotation between point clouds with Procrustes alg. [Viklands06] (RANSAC) When finished, reject those point correspondences which are above a threshold 16

Point correspondences 17

Sensor Geometry A 3D point X relates to its corresponding homogenous image point x as: x = KX, and X = z(x)k 1 x where K is the intrinsic camera matrix and z(x) is the point s value in the depth image We model the camera motion as a seq. of rotation matrices R(t) and translation vectors d(t) 18

Sensor motion estimation A point in image 1, at row N x corr. to 3D point X 1 and A point in image 2, at row N y corr. to 3D point X 2. They can be transformed to X 0 : X 0 = R(N 1 )X 1 + d(n 1 ) X 0 = R(N 2 )X 2 + d(n 2 ) X 0 is the position the point should have, if it was imaged the same time as the first row in image 1 19

Sensor motion estimation We use this cost function to solve for the rotation and translation: K J = R(N 1,k )X 1,k + d(n 1,k ) k=1 R(N 2,k )X 2,k d(n 2,k ) 2 where K is the number of point corr. 12 unknowns, 3 equations per point corr. R key-rotations as before, and d key-translations 20

Sensor motion estimation SLERP for rotations and Key-rotations and key-translations Linear interpolation for translations 21

Rectification When the camera motion has been estimated the 3D point clouds can be rectified with X = R ref (R(N 1 )X 1 + d(n 1 )) + d ref By projecting the points through the camera, depth map and video frames can also be rectified: x = K[R ref (R(N 1 )X 1 + d(n 1 )) + d ref ] 22

Rectification Original Rectified 23

Rectification 24

Summary When the Kinect is used on a mobile platform rolling-shutter distortions will be present Using correspondences in the NIR images avoids depth-to-image registration problem Depth map noisy, optimisation in 3D more sensitive than video approach Full 6-DOF motion can be estimated and corrected for 25

References Viklands, Algorithms for the Weighted Orthogonal Procrustes Problem and Other Least Squares Problems, Umeå University 2006 Ringaby, Forssén, Scan Rectification for Structured Light Range Sensors with Rolling Shutters, ICCV 11 Izadi et-al KinectFusion: Real-Time Dynamic 3D Surface Reconstruction and Interaction, SIGGRAPH 11 26