Overview. Augmented reality and applications Marker-based augmented reality. Camera model. Binary markers Textured planar markers

Similar documents
Outline. Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion. Media IC & System Lab Po-Chen Wu 2

Step-by-Step Model Buidling

Vision Review: Image Formation. Course web page:

Image correspondences and structure from motion

Today s lecture. Image Alignment and Stitching. Readings. Motion models

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Application questions. Theoretical questions

Pin Hole Cameras & Warp Functions

3D Visualization through Planar Pattern Based Augmented Reality

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo --

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

Image Stitching. Slides from Rick Szeliski, Steve Seitz, Derek Hoiem, Ira Kemelmacher, Ali Farhadi

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Hartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne

Homographies and RANSAC

calibrated coordinates Linear transformation pixel coordinates

Topics to be Covered in the Rest of the Semester. CSci 4968 and 6270 Computational Vision Lecture 15 Overview of Remainder of the Semester

UNIVERSITY OF TORONTO Faculty of Applied Science and Engineering. ROB501H1 F: Computer Vision for Robotics. Midterm Examination.

Epipolar Geometry and Stereo Vision

Planar homographies. Can we reconstruct another view from one image? vgg/projects/singleview/

EECS 442: Final Project

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006

CS6670: Computer Vision

1 Projective Geometry

Rigid Body Motion and Image Formation. Jana Kosecka, CS 482

Epipolar Geometry CSE P576. Dr. Matthew Brown

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

3D Computer Vision. Structure from Motion. Prof. Didier Stricker

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Parameter estimation. Christiano Gava Gabriele Bleser

Camera Geometry II. COS 429 Princeton University

Camera model and multiple view geometry

What have we leaned so far?

Cameras and Stereo CSE 455. Linda Shapiro

Lecture 14: Basic Multi-View Geometry

CSE 252B: Computer Vision II

Stereo and Epipolar geometry

3D Reconstruction of a Hopkins Landmark

Outline. ETN-FPI Training School on Plenoptic Sensing

COSC579: Scene Geometry. Jeremy Bolton, PhD Assistant Teaching Professor

CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry

Segmentation and Tracking of Partial Planar Templates

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,

Mosaics. Today s Readings

Augmented Reality, Advanced SLAM, Applications

Stereo Vision. MAN-522 Computer Vision

Chapter 3 Image Registration. Chapter 3 Image Registration

An introduction to 3D image reconstruction and understanding concepts and ideas

PART IV: RS & the Kinect

Flexible Calibration of a Portable Structured Light System through Surface Plane

Robot Vision: Camera calibration

Structure from Motion. Prof. Marco Marcon

Structure from motion

AUGMENTED REALITY. Antonino Furnari

arxiv: v1 [cs.cv] 28 Sep 2018

Epipolar Geometry and Stereo Vision

Multi-Projector Display with Continuous Self-Calibration

CS 6320 Computer Vision Homework 2 (Due Date February 15 th )

MERGING POINT CLOUDS FROM MULTIPLE KINECTS. Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia

Master Automática y Robótica. Técnicas Avanzadas de Vision: Visual Odometry. by Pascual Campoy Computer Vision Group

Camera Models and Image Formation. Srikumar Ramalingam School of Computing University of Utah

Two-view geometry Computer Vision Spring 2018, Lecture 10

Correcting Radial Distortion of Cameras With Wide Angle Lens Using Point Correspondences

Multiple View Geometry

3D Vision Real Objects, Real Cameras. Chapter 11 (parts of), 12 (parts of) Computerized Image Analysis MN2 Anders Brun,

Image-based Modeling and Rendering: 8. Image Transformation and Panorama

Chaplin, Modern Times, 1936

3D Reconstruction from Two Views

Pin Hole Cameras & Warp Functions

Lecture 9: Epipolar Geometry

Stitching and Blending

Video Analysis for Augmented and Mixed Reality. Kiyoshi Kiyokawa Osaka University

Augmenting Reality, Naturally:

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

BIL Computer Vision Apr 16, 2014

CS 231A Computer Vision (Winter 2018) Problem Set 3

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Computer Vision. Exercise 3 Panorama Stitching 09/12/2013. Compute Vision : Exercise 3 Panorama Stitching

Object Recognition with Invariant Features

Fundamental Matrix & Structure from Motion

CS 231A Computer Vision (Winter 2014) Problem Set 3

Multiple View Geometry

Geometric camera models and calibration

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac

WATERMARKING FOR LIGHT FIELD RENDERING 1

Camera models and calibration

Video Mosaics for Virtual Environments, R. Szeliski. Review by: Christopher Rasmussen

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

Experiments with Edge Detection using One-dimensional Surface Fitting

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Camera Model and Calibration

COMPUTER AND ROBOT VISION

Computer Vision Lecture 17

Dense 3D Reconstruction. Christiano Gava

Computer Vision Lecture 17

Transcription:

Augmented reality

Overview Augmented reality and applications Marker-based augmented reality Binary markers Textured planar markers Camera model Homography Direct Linear Transformation

What is augmented reality? Reality is subjective Sight Hearing Smell Haptics Balance Augmenting sensory information

Augmenting visual information Superpositioning digital information on top of real imagery Who Framed Roger Rabbit (1988) Tom Caudell - Boeing (1990) Xerox & University of Toronto (1990) Virtual Fixtures (1992)

Classical approach Visual information acquisition: camera Camera localization Image: camera Depth: depth camera Other: GPS, WiFi, IMU Displaying augmented information: monitor, mobile phone, projector, smart glasses

Application examples (TV) Olympic games 2004 Monitor/TV Robotic camera USA elections 2008 CNN Hologram conference 35 cameras 20 computers

Application examples (Mobile) Mobile devices Pokemon Go Vuforia, ARKit Wearables BMW Hololens

AR using depth information Depth cameras Active (IR light) Passive (Stereo systems) Automatic scene reconstruction Easier interaction with objects Izadi et al. "KinectFusion: Real-Time Dynamic 3D Surface Reconstruction and Interaction", SIGGRAPH, 2011

AR with visual anchor Localization with visual information Detect key object in image Determine relative position of the object to camera Draw information with this relation

From point to pixel Transform 3D point to camera coordinate system (pixels) Required data K... camera calibration matrix (intrinsic parameters) R,t... rotation and translation matrices (extrinsic parameters)

AR with binary marker Detect markers that are easy to detect and identify Detect marker from edges Identify marker with correlation Known marker size Compute relation to camera Use corners of marker to compute relative position

AR with arbitrary planar marker Match an arbitrary surface Describe local texture Robust matching Less constrained can use existing textures from real world Posters Building facades

Applications of marker AR Catalogs Books Tourism Gaming

Detecting binary marker Machine vision approach Finding possible candidates Adaptive threshold Trace contours Estimate contours as polygons Contours with 4 corners

Recognizing binary marker Threshold on similarity Project region to reference position Normalized cross correlation Orientation test all four options

Generalizing transformation Determine camera extrinsics using a detected marker Transformation between planes (homography) Marker plane (reference) Camera plane (observed)

Required linear algebra Homogeneous coordinates Can describe point in infinity Same point if multiplied with scalar Transformation matrix contains translation Matrix form of vector (cross) product Express product as matrix multiplication of skew-symmetric matrix and vector

Systems of linear equations Get a set of variables from a set of equality relations Known pairs of vectors correspondences Linear transformation contains unknowns Formulate problem in matrix form: Sufficient data - unique solution: Not enough data underdetermined system Too much data overdetermined system

Direct Linear Transform Similarity relations Unknown multiplicative factor Rewritten as homogeneous equations Get rid of scaling factor

DLT for Homography Homography is 3x3 matrix 8 unknowns Assume Need 4 correspondences

The top two equations are independent, the last one is not. We can reorganize the system

Solving homogeneous system N correspondences give us 2*N equations We can get solution for the system using SVD The solution is the last column of V Minimal mean square error Reorganize vector into matrix

Optimizing extrinsics Homography projects points between two planes Determine camera position Inaccurate R and t Objective function is MSR, not reprojection error

Minimizing reprojection error Projection of 3D point to image coordinates is written as When R and t are optimal the difference between image point and projection of 3D point should be zero Cost function is sum of all distances, it is non-linear

Iterative optimization of R and t Initial R ant t from homography Fix t and minimize R Use quaternions Optimize with non-linear methods (e.g. Levenberg-Marquardt, Simplex) Fix R and minimize t Using least-squares optimization Repeat steps until convergence

Binary marker example Simple (no normalization) With normalization

Augmented reality with planar marker Natural surfaces Unable to detect corners robustly Partial occlusions Can detect feature-points Over-sample reference points Not all points will match correctly Robustly estimate homography

Matching keypoints Detector of keypoints Descriptor of regions SIFT, SURF BRIEF, ORB Matching descriptors Distance function Symmetric matches

Robust estimation of homography Many correspondences Over-determined system Not all correspondences are correct Robust matching Exclude outliers from calculation Find sub-set of correspondences that agree on a model

RANSAC Random Sample Consensus Meta algorithm (used for many tasks) Probabilistic interpretation Repeat k times Select random set of 4 correspondences Estimate model homography (DLT) Look which other pairs agree with the model (projection from one plane to the other is small enough) Take the model with largest support (inliers) RANSAC for line fitting (source: F. Moreno)

Reference plane example Disney AR coloring book Urban AR

Beyond planar markers Reference objects Deformable surfaces No reference

Realistic rendering Acquired images are degraded by various factors Augmented reality will be more immersive if these factors are replicated on virtual objects

Motion blur Exposure time + rapid motion Simulated by smoothing image with directional filter

Chromatic aberration Different wavelenghts bend under different angles Simulated by distorting individual color channels Image center Image border

Spherical aberration Spherical lenses do not focus light perfectly Rays on the edge are focused closer Compensated by using multiple lenses No abberation With abberation

Vignetting Optical vignetting multiple lenses Pixel vignetting Digital sensors Angle dependence Software compensation

Radial distortion Imperfect lenses Deviations are apparent near the edge of image

Realistic rendering pipeline Gerhard Reitmayr, Axel Pinz, Visual Coherence in Augmented Reality

Diminished reality Use AR to determine which part of the image to erase using inpainting Herling, J. and Broll, W., Advanced Self-contained Object Removal for Realizing Real-time Diminished Reality in Unconstrained Environments. ISMAR 2010