3D scanning. 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects.

Similar documents
A 3D scanning primer. Marco Callieri Matteo Dellepiane Visual Computig Lab ISTI-CNR

Tecnologie per la ricostruzione di modelli 3D da immagini. Marco Callieri ISTI-CNR, Pisa, Italy

3D Scanning. Qixing Huang Feb. 9 th Slide Credit: Yasutaka Furukawa

Visual Perception Sensors

A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect

Overview of Active Vision Techniques

Tecnologie per la ricostruzione di modelli 3D da immagini. Marco Callieri ISTI-CNR, Pisa, Italy

3D from Images - Assisted Modeling, Photogrammetry. Marco Callieri ISTI-CNR, Pisa, Italy

3D Photography: Active Ranging, Structured Light, ICP

Outline. ETN-FPI Training School on Plenoptic Sensing

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

Optical Active 3D Scanning. Gianpaolo Palma

DIGITAL SURFACE MODELS OF CITY AREAS BY VERY HIGH RESOLUTION SPACE IMAGERY

3D Modeling of Objects Using Laser Scanning

Structured light 3D reconstruction

Range Sensors (time of flight) (1)

Depth Sensors Kinect V2 A. Fornaser

Multi-View Stereo for Community Photo Collections Michael Goesele, et al, ICCV Venus de Milo

3D Photography: Stereo

Stereo Vision A simple system. Dr. Gerhard Roth Winter 2012

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

3D Object Representations. COS 526, Fall 2016 Princeton University

Outline Sensors. EE Sensors. H.I. Bozma. Electric Electronic Engineering Bogazici University. December 13, 2017

10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1.

Active Stereo Vision. COMP 4900D Winter 2012 Gerhard Roth

Computer Vision. 3D acquisition

L2 Data Acquisition. Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods

The main problem of photogrammetry

Sensor Modalities. Sensor modality: Different modalities:

Other Reconstruction Techniques

Processing 3D Surface Data

Scene Segmentation by Color and Depth Information and its Applications

Computer Science Applications to Cultural Heritage. 3D Reconstruction

Calibration of a rotating multi-beam Lidar

ToF Camera for high resolution 3D images with affordable pricing

CHAPTER 2: THREE DIMENSIONAL TOPOGRAPHICAL MAPPING SYSTEM. Target Object

Processing 3D Surface Data

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Stereo vision. Many slides adapted from Steve Seitz

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

ENY-C2005 Geoinformation in Environmental Modeling Lecture 4b: Laser scanning

BIL Computer Vision Apr 16, 2014

Traditional pipeline for shape capture. Active range scanning. New pipeline for shape capture. Draw contours on model. Digitize lines with touch probe

EE565:Mobile Robotics Lecture 3

Multiple View Geometry

Gregory Walsh, Ph.D. San Ramon, CA January 25, 2011

All human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them.

Introduction to 3D Machine Vision

3DPIXA Stereo line scan cameras for ultra-high resolution 3D applications

Measurements using three-dimensional product imaging

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?

Stereo. Many slides adapted from Steve Seitz

Three-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras

Computational Imaging for Self-Driving Vehicles

CSE 4392/5369. Dr. Gian Luca Mariottini, Ph.D.

New Sony DepthSense TM ToF Technology

GABRIELE GUIDI, PHD POLITECNICO DI MILANO, ITALY VISITING SCHOLAR AT INDIANA UNIVERSITY NOV OCT D IMAGE FUSION

Applications of Piezo Actuators for Space Instrument Optical Alignment

3D Computer Vision 1

Modeling. Michael Kazhdan ( /657) HB FvDFH Modeling Seashells, Fowler et al. 1992

CS4758: Rovio Augmented Vision Mapping Project

Computer Graphics. Bing-Yu Chen National Taiwan University The University of Tokyo

Design and Calibration of a Network of RGB-D Sensors for Robotic Applications over Large Workspaces

SANGAM PROJECT BROCHURE:

Chaplin, Modern Times, 1936

Image Based Reconstruction II

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

CS4495/6495 Introduction to Computer Vision

Depth Camera for Mobile Devices

Multiple View Geometry

Information page for written examinations at Linköping University TER2

High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond

Bumblebee2 Stereo Vision Camera

Photoneo's brand new PhoXi 3D Camera is the highest resolution and highest accuracy area based 3D

STEREO-VISION SYSTEM PERFORMANCE ANALYSIS

Sensor technology for mobile robots

Probabilistic Robotics

AICON MoveInspect XR8

Using infrared proximity sensors for close 2D localization and object size recognition. Richard Berglind Neonode

HIGH-PERFORMANCE LONG-RANGE LASER SCANNER WITH CALIBRATED DIGITAL CAMERA: A HYBRID 3D LASER SENSOR SYSTEM

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe

Stereo and structured light

3D Time-of-Flight Image Sensor Solutions for Mobile Devices

CIS 580, Machine Perception, Spring 2015 Homework 1 Due: :59AM

Neue Verfahren der Bildverarbeitung auch zur Erfassung von Schäden in Abwasserkanälen?

Agenda. DLP 3D scanning Introduction DLP 3D scanning SDK Introduction Advance features for existing SDK

Reverse Engineering: Mechanical. Dr. Tarek A. Tutunji

Time-of-flight basics

Il colore: acquisizione e visualizzazione. Lezione 17: 11 Maggio 2012

Solid-State Hybrid LiDAR for Autonomous Driving Product Description

Non-line-of-sight imaging

Structured Light II. Guido Gerig CS 6320, Spring (thanks: slides Prof. S. Narasimhan, CMU, Marc Pollefeys, UNC)

Electronic Travel Aids for Blind Guidance an Industry Landscape Study

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems

An Introduction to Lidar & Forestry May 2013

Laser Eye a new 3D sensor for active vision

: Easy 3D Calibration of laser triangulation systems. Fredrik Nilsson Product Manager, SICK, BU Vision

TERRESTRIAL LASER SCANNER DATA PROCESSING

Transcription:

Acquiring 3D shape

3D scanning 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects. The produced digital model is formed by geometric information that have been measured and have a metric quality. It may be imprecise and incomplete, but still has all the characteristics of a measurement result

3D scanning: a taxonomy Contact Articulated arms Distance Slicing Optical Active Non optical Passive

Articulated arm The probing point has a known position in every moment, thanks to the angle sensor at joints Object is probed in various points, generating a grid that wll use as a guide for modeling Very manual method, still with lot of artistic influence Industrial sensors: The arm is autonomous and touches the surface using a predefined, regular, grid. Precisions in the order of microns

3D scanning: a taxonomy Contact Articulated arms Distance Slicing Optical Active Non optical Passive

Slicing CAT scan, magnetic resonance... Slightly more invasive: The Visible Human Project

3D scanning: a taxonomy Contact Articulated arms Distance Slicing Optical Active Non optical Passive

Shape from Stereo Based on the same principle of human stereo vision: two sensors that perceive the world from slightly different position. From parallax it is possible to obtain a depth for each visible point Our brain does this in automatic A machine can be programmed to do the same Same position => background High variance => close Mid variance => mid distance

Aerial / Satellite Same principle, features are isolated and matched to generate a depth map Disparity comes from different view directions and/or the movement of the plane. Same strategy is used also from satellites

3D scanning: a taxonomy Contact Articulated arms Distance Slicing Optical Active Non optical Passive

Triangulation A light is projected on the surface and its reflection is read back by a sensor Using trigonometric calculation it is simple to recover the 3d position of the illuminated points The geometric principle is the simplest possible! The real problem, to obtain precision, is to carefully calibrate each component CCD Knowing the emission and reception angle, and using the distance between the emitter and the sensor, the distance of each sampled point is calculated α d β Laser z

Minolta Vivid 910 A commercial scanner, with high precision (0.20.3mm), but high cost (>30K euros). A laser line is swept over the object: 300K points are measured in 2.5 seconds.

A cheap scanner: NextEngine Entry-level 3D scanner, simple and cheap. Good quality/price ratio. Ideal to investigate the possible use of this technology in a laboratory/museum/superintendence Pro: - Small price (<2k euros) - Good resolution and result coherence. - Highly portable (small and lightweight) Cons: - Fixed working distance - Need parameter tweaking - Does not work well on some materials (dark & shiny) http://www.nextengine.com

What cannot be scanned Unfortunately, many different surfaces remains difficult or impossible to scan accurately Objects too dark, absorbing colors Transparent or Translucent surfaces Shiny/mirror finish Non-solid surfaces: hairs, furs, feathers Moving objects...

Other technologies Beside lasers, a different light impulse may be used to measure geometry. Instead of a single (or multiple) laser line, project a large, structured pattern. These scanners uses the principle generally called: structured light

Structured Light The principle is still Triangulation, but different patterns are projected on the object. Can be more precise than laser-line triangulation, but require additional hardware (and has some problem on larger objects). Different companies are offering software able to control a camera and a projector. There are also free/open projects which do so...

Breuckmann GmBh A german company that produce and sell extremely precise (and extremely costly) structured light scanners (sampling resolution 0.05mm)

Microsoft KINECT It is basically a very fast (30fps) structured light that uses infrared light (in this way the pattern will be invisible). Resolution is not great, and the RGB color is not completely aligned witht he geometry. However, its cost and performances have shaken the community of 3D hobbysts but also of professionals

Microsoft KINECT It is possible (in theory) to use the Kinect to do a 3D scanning, however: The kinect has been built for speed, not precision: you need a stable position of both the device and the subject. You may need to get more than a shot from the same position and combine it to reduce noise The depth information is compressed, especially in the far area: the subject should stay as close as possible to the device

What about larger Objects? This is a very common question... The answer is, you do need a different instrument Triangulation cannot work on very large objects, it would require an extremely large baseline... Always remaining in the kingdom of light signals and optical properties, a different strategy is used

Time of Flight (TOF) The distance of sampled point is obtained by measuring the time between the laser impulse and the sensor read-back, divided by (two times) the speed of light The measurements is repeated on a regular grid on the object surface WARNING: playing with the speed of light reduce the measurement precision Average error on distance: 5-7 mm

Phase interference The direct and the reflected beam arrive on the crystal, frequencies are no longer aligned, producing interference interference bands are used to determine the distance of the sample The same principle is used on two different scales: CONOSCOPY: coins, paintings, small relief INTERFERENCE TOF: buildings

Time of Flight + interference More advanced systems also rely on interference patterns produced by different wavelenghts, making the devices faster and more accurate. They can also see through vegetation...

LIDAR / SLR Elevation data measured by satellites. But not only geometry Used in combination with analysis of multiple returning signals, can see through vegetation Using different frequencies and analyzing the returned signal, it is possible to distinguish the nature (building, road, water, cultures ) of the probed area Average error on distance: a few meters SLR satellite laser ranging LIDAR light detection and ranging

3D scanning devices Sensor is no longer the main problem... The gamut of measurable object is increasing, in terms of both size and material New hardware is made available as we speak... Phase interference/shift Triangulation Struct. Light Conoscopy Time of flight LIDAR / SLR multispectral

3D scanning technology: key terms It s not easy to compare 3D scanners. There are some key terms that need to be clear, at least: - Accuracy (precision): error/distance. Difference between triangulation and other technologies. - Resolution (essentially, no sense) - Speed (points/min) - Working distance (min/max)

Range map What is recovered from most of the scanning equipment after a single shot is a depth value for each pixel in its sensor These depths are converted in a point grid that is promptly triangulated using this intrinsic regularity. The result of a single scan is called a RANGE MAP

Credits: These slides are an excerpt from the course taught by Matteo Dellepiane http://vcg.isti.cnr.it/~dellepiane/corso_2012/corso_2012.html