3D from Images - Assisted Modeling, Photogrammetry. Marco Callieri ISTI-CNR, Pisa, Italy

Similar documents
Tecnologie per la ricostruzione di modelli 3D da immagini. Marco Callieri ISTI-CNR, Pisa, Italy

Tecnologie per la ricostruzione di modelli 3D da immagini. Marco Callieri ISTI-CNR, Pisa, Italy

Digital 3D technologies

Stereo Vision. MAN-522 Computer Vision

Multiple View Geometry

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Computer Vision Lecture 17

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Computer Vision Lecture 17

Dense 3D Reconstruction. Christiano Gava

Miniature faking. In close-up photo, the depth of field is limited.

Il colore: acquisizione e visualizzazione. Lezione 17: 11 Maggio 2012

calibrated coordinates Linear transformation pixel coordinates

Structure from motion

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

Project 4 Results. Representation. Data. Learning. Zachary, Hung-I, Paul, Emanuel. SIFT and HoG are popular and successful.

All human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them.

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

CSE 4392/5369. Dr. Gian Luca Mariottini, Ph.D.

Dense 3D Reconstruction. Christiano Gava

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.

Image Transformations & Camera Calibration. Mašinska vizija, 2018.

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Il colore: acquisizione e visualizzazione. Lezione 20: 11 Maggio 2011

3D scanning. 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects.

BIL Computer Vision Apr 16, 2014

Augmenting Reality, Naturally:

Structured light 3D reconstruction

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Correspondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]

Multiple View Geometry

1 Projective Geometry

Stereo: Disparity and Matching

Final Exam Study Guide

Structure from motion

Stereo and Epipolar geometry

Depth from two cameras: stereopsis

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION

Lecture 9: Epipolar Geometry

CHAPTER 3. Single-view Geometry. 1. Consequences of Projection

Depth from two cameras: stereopsis

CS231A Midterm Review. Friday 5/6/2016

Multiview Reconstruction

Application questions. Theoretical questions

Camera Drones Lecture 3 3D data generation

arxiv: v1 [cs.cv] 28 Sep 2018

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision

Rasterization Overview

Model-Based Stereo. Chapter Motivation. The modeling system described in Chapter 5 allows the user to create a basic model of a

SL A Tordivel - Thor Vollset -Stereo Vision and structured illumination creates dense 3D Images Page 1

Archeoviz: Improving the Camera Calibration Process. Jonathan Goulet Advisor: Dr. Kostas Daniilidis

A Qualitative Analysis of 3D Display Technology

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Stereo Epipolar Geometry for General Cameras. Sanja Fidler CSC420: Intro to Image Understanding 1 / 33

Geometric Reconstruction Dense reconstruction of scene geometry

Step-by-Step Model Buidling

An Overview of Matchmoving using Structure from Motion Methods

Computer Vision I. Announcement. Stereo Vision Outline. Stereo II. CSE252A Lecture 15

CS201 Computer Vision Camera Geometry

Computer Vision. Introduction

What is Computer Vision?

Project Title: Welding Machine Monitoring System Phase II. Name of PI: Prof. Kenneth K.M. LAM (EIE) Progress / Achievement: (with photos, if any)

Chaplin, Modern Times, 1936

Stereo Vision A simple system. Dr. Gerhard Roth Winter 2012

Last update: May 4, Vision. CMSC 421: Chapter 24. CMSC 421: Chapter 24 1

Computer Vision cmput 428/615

LUMS Mine Detector Project

Virtual Testbeds for Planetary Exploration: The Self Localization Aspect

1 (5 max) 2 (10 max) 3 (20 max) 4 (30 max) 5 (10 max) 6 (15 extra max) total (75 max + 15 extra)

: Easy 3D Calibration of laser triangulation systems. Fredrik Nilsson Product Manager, SICK, BU Vision

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Camera Model and Calibration

COMPARISON OF LASER SCANNING, PHOTOGRAMMETRY AND SfM-MVS PIPELINE APPLIED IN STRUCTURES AND ARTIFICIAL SURFACES

Neue Verfahren der Bildverarbeitung auch zur Erfassung von Schäden in Abwasserkanälen?

IMAGE-BASED 3D ACQUISITION TOOL FOR ARCHITECTURAL CONSERVATION

P1: OTA/XYZ P2: ABC c01 JWBK288-Cyganek December 5, :11 Printer Name: Yet to Come. Part I COPYRIGHTED MATERIAL

Binocular cues to depth PSY 310 Greg Francis. Lecture 21. Depth perception

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

CS595:Introduction to Computer Vision

CS4495/6495 Introduction to Computer Vision. 3B-L3 Stereo correspondence

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Final Review CMSC 733 Fall 2014

From Structure-from-Motion Point Clouds to Fast Location Recognition

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction

Project report Augmented reality with ARToolKit

Centre for Digital Image Measurement and Analysis, School of Engineering, City University, Northampton Square, London, ECIV OHB

Detecting motion by means of 2D and 3D information

3D Modeling using multiple images Exam January 2008

12/3/2009. What is Computer Vision? Applications. Application: Assisted driving Pedestrian and car detection. Application: Improving online search

Lecture 14: Computer Vision

INFO - H Pattern recognition and image analysis. Vision

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Multi-view stereo. Many slides adapted from S. Seitz

Parallel algorithms for stereo vision shape from stereo

Transcription:

3D from Images - Assisted Modeling, Photogrammetry Marco Callieri ISTI-CNR, Pisa, Italy

3D from Photos Our not-so-secret dream: obtain a reliable and precise 3D from simple photos Why? Easier, cheaper and more versatile than a 3D scanner We can perceive 3D shapes from images, why should a computer not be able to do the same?

Perspective and Stereo We do perceive the environment three-dimensionality thanks to our stereo vision AND the perspective projection that occurs in our eyes.. Using the geometric laws of this two phenomena we can build the 3D geometry of the scene

Perspective Projection Relationship #1 Each point in the 3D scene corresponds to exactly ONE point on the image plane. If I know the XYZ poisition of some point in the scene, I know which pixel in the photo will generate

Perspective Projection Relationship #2 Each point on the image plane corresponds to exactly ONE line in the 3D scene.? I know the entity that generated that pixel lies somewhere on that line, but I do not know where...

More than one photo Even though relationship #2 is not as strict as #1, combinig more photos, if I choose the SAME feature in the images, the lines will met somewhere in the 3D space, removing the ambiguity...

Perspective Camera But to establish these correspondences, what do I need? All starts here... a camera is defined by parameters: - Position - Orientation - Focal Length - Viewport - Distortion extrinsics intrinsics

Ouroboros But, in this way, It is a dog biting its own tail! In order to know the scene geometry, you do need to know something about the scene geometry We are lucky: by providing semantic information (non metric), it is possible to determine the camera Calibration and Orientation the info needed is: - (hand picked) image-to-image correspondences of points - (hand picked) correspondences between image points and a known geometry

Camera Calibration From some initial data (correspondences), determine all camera parameters Completely mathematical process... however, not an easy one... there are lots of numerical problems.

2 paths... Two possible directions - Assisted Modeling - Automatic Stereo Matching The underlying principles are the same... but they sprouted different kind of tools... As usual, this distinction is becoming blurry......things do converge, after all

Assisted Modeling

Perspective & vanishing point If in an image, the depicted object presents lines along two axis, we can easily detect vanishing point(s) and axis orientation Not really much, but much better than nothing

Sketchup A very strange modeling tool. Follows more the way a technical drawing is done on paper (reporting/referencing) than the usual 3D modeling metaphore. Easier for people with a technical drawing/sketching background. Easier for people with no experience in 3D modeling. Focused towards modeling of buildings and mechanical entities... Acquired by Google some years ago, and then sold again... distributed now as a semi-free tool http://www.sketchup.com/

Sketchup the Luni Temple An ideal tool for very regular buildings... like this one (regular does not equal new)

Sketchup Photo Match Assited modeling from a SINGLE photo Calibration: axis and vanishing points markup Modeling: 3D drawing by axis/reference reporting Partial calibration, with only a single photo, only the axis can be recovered. SketchUp can be used to model by reporting/referencing http://sketchup.google.com/

Photogrammetry Perspective & stereo Common reference points are marked on multiple images From these correspondences it is possible to calculate camera position/parameters and 3D location of the marked points

Photogrammetry At first seems impossible, but with some care, the obtained precision is way below the projected size of a pixel. For a building, it may be few millimeters (yes, it does depend on the size of the object) Photogrammetry is the name of the principle; many different tools and approaches. Often rely on calibrated cameras and markers... Often used in conjunction with other measuring methods...

ImageModeler Photogrammetry commercial tool Points are marked on input images, camera are fully calibrated using these points, camera are calibrated and modeling can be done using the recovered 3D points Acquired by Autodesk... now just a module for other Autodesk tools

PhotoModeler Photogrammetry commercial tool The tool for the professionals... Two steps: camera calibration (with markers and references) and camera pose estimation. Modeling and measuring with lots of different tools Very, very, very complex to use...

XSIGNO Shareware photogrammetry tool Oriented to measurement (and not modeling) Intensive use of markers for camera calibration (intrinsics) and orientation (extrinsics) Looks promising for a small software... But the necessity to use markers may reduce its applicability

The Campanile Movie A short movie from 1997, but with a large impact on the movie industry... A photogrammetry-based model used for a video where real-world sequences are interleaved with digital renderings Seems quite naïve... but it was a demonstration of the maturity of image-based graphics...

Façade Component-based photogrammetry... Correspondence points are used to place parametric blocks (cubes, pyramids, archs) The basic principle is still photogrammetry, but it is focused on buildings (which represent ¾ of the market) View-dependent texture mapping: color information is generated in realtime from the best photo resulting in a much more realistic animation

Photogrammetry Photo-matching Beside 3D model CREATIOn, Photogrammetry is often used for camera matching i.e. determine camera parameters and scene space, to insert virtual elements in the scene Used A LOT in engineering/architecture. This is not the only nor the easier way to do this, but still is a popular choice And for videos? We ll see it later

The sad news NO free photogrammetry tools out there... sorry IMAGEMODELER: acquired by Autodesk... Now component for other tools PHOTOMODEL: very professional tool, good support, quite costly XSIGNO: http://www.xsigno.com/ CANOMA: a very interesting tool acquired by Adobe and disappeared... - try using demo/trial versions - hope for the best...

Recap 2 phases: CALIBRATION & GEOMETRY EXTRACTION - both are based on correspondences - first calibration step returns also some 3D points in the scene - correspondences are made by user or recovered from markers in the scene Results: a series of points in 3D - very high precision - only the marked points are recovered - modeling is done on the recovered points

Automatic 3D extraction

Automatic Idea: use some kind of automatic feature matching to perform a DENSE reconstruction... Not just rely on user-picked points, but try to match the entire surface, pixel by pixel.. Basic situation: - camera calibration and position are known. - automatic matching between image pixels for DENSE reconstruction

Automatic Difficult? Yes and no - NO given registered cameras, a pixel P1 on image 1 will have its corresponding pixel P2 on image 2 laying on a known line. This is called EPIPOLAR line(thanks to the laws of perspective) - YES changes in illumination, different exposition, flat-colored areas, noise and occlusion makes difficult to always pick the right candidate

Shape from Stereo Based on the same principle of human stereo vision: two sensors that perceive the world from slightly different position. From parallax it is possible to obtain a depth for each visible point Our brain does this automatically A machine can be programmed to do the same Same position => background High variance => close Mid variance => mid distance

And i mean automatically The brain does stereo match continuously... even when the input is random junk... this is the base of stereograms. The patterns are constructed such that the matched points follows the disparity of a given depth map depth SX DX SX DX

Dedicated Devices It is available on the market, some hw that is able to perform stereo matching (entirely or partially) and gives back a depth map. In practice it is formed by two (or more) synchronized cameras, plus a DSP. Normally used in robotics, because they are fast but not much precise but hardware is getting better

PhotoModeler - dense Photogrammetry commercial tool, 2 nd version The tool is the same as before... after doing the camera calibration, instead of using only user-picked points, the system does a dense-matching. The result is quite similar to a range map... you need to do standard processing in order to obtain a 3D model...

MenciSoft Commercial solution italian product (quite unusual) Three photos from a calibrated camera sliding over a very precise railing. Easy to use, versatile (multi-scale), fast acquisition Good results, but very long processing time to obtain a final, complete model

MenciSoft

MenciSoft - camera is calibrated (intrinsics) - camera positions are known (extrinsics) What remains is just DENSE matching... The very regular camera placement helps the stereo matching Result: fast, precise and reliable extraction of 3D data The real pain is data processing :)

Fully automatic A step further... also the camera position is unknown 1- automatic feature match between photos to obtain initial calibration 2- automatic dense matching to recover 3D data The additional step is much more difficult than the dense step... finding GOOD features for image calibration is hard because, in this case, there is no epipolar geometry to help. Modern image analysis techniques make it possible (local descriptors like SIFT)

Thanks for your attention Question Time?!???!??!