Computational Optical Imaging - Optique Numerique. -- Active Light 3D--

Size: px
Start display at page:

Download "Computational Optical Imaging - Optique Numerique. -- Active Light 3D--"

Transcription

1 Computational Optical Imaging - Optique Numerique -- Active Light 3D-- Winter 2013 Ivo Ihrke with slides by Thorsten Thormaehlen

2 Overview 3D scanning techniques Laser triangulation Structured light Photometric stereo Calibration Time-of-Flight Transient Imaging

3 Active Light - 3D Scanning Laser Scanning

4 Range scanning Generate range from stereo information Scales from desktop to whole body scanners

5 Range scanning systems Passive: n 2 cameras; match images

6 Optical Triangulation passive setup: match corresponding images left sensor object right sensor

7 Active Techniques project some additional signal into the scene requires signal source, and additional calibration often more or less direct measurements typically higher precision Examples: range scanning matting applications measurement of reflection properties

8 Range scanning systems Active: Laser light source Sweep & track reflection (triangulation) Project a line instead of a single point

9 Optical triangulation active setup: track light spot laser source z b tan( a) tan( b ) a object base length b sensor b (x,z) ray intersection z

10 Laser triangulation Source: The Digital Michelangelo Project, Stanford University

11 Laser triangulation light plane object camera laser line of sight Depth from ray-plane triangulation: Intersect line of sight with light plane (canonical camera)

12 Laser triangulation Advantages simple approach works quite reliable accuracy proportional to working volume (typical is ~1000:1) scales down to small working volume (e.g. 5 cm. working volume, 50 m. accuracy) Disadvantages long scanning time (no real-time capture of moving objects) does not scale up (baseline too large ) two-lines-of-sight problem (shadowing from either camera or laser) triangulation angle: lower accuracy if too small, shadowing if too big (useful range: 15-30) needs stable reliable calibration with good mechanics, thus often expensive to buy

13 Digital Michelangelo Project, motorized axes extensions for tall statues cyberware scanner (laser triangulation)

14 Digital Michelangelo Project, height of gantry: weight of gantry: 7.5 meters 800 kilograms Source: The Digital Michelangelo Project, Stanford University

15 Digital Michelangelo Project, working in the museum scanning geometry scanning color Source: The Digital Michelangelo Project, Stanford University

16 Digital Michelangelo Project, individually aimed scans 2 billion polygons 7,000 color images 32 gigabytes 30 nights of scanning 22 people Source: The Digital Michelangelo Project, Stanford University

17 Digital Michelangelo Project, photograph 1.0 mm computer model Source: The Digital Michelangelo Project, Stanford University

18 Digital Michelangelo Project, St. Matthew David Forma Urbis Romae Source: The Digital Michelangelo Project, Stanford University

19 Range Scanners Optical Considerations Laser Range Scanners focal plane selection Scheimpflug principle tilt-shift lenses focal plane Scheimpflug principle application in range scanning extend depth of field

20 Tilt-Shift Lens Photographic Examples

21 Range Scanning Problems - Shadowing sensor light source sensor can t see lit point point in shadow longer baseline: more shadowing shorter baseline: less precision

22 Range Scanning Problems Acquired Range Image a single range image is acquired for each view Artifacts: ribbed lots of holes due to surface properties due to occlusion Solution: Merging of multiple scans -> (ICP, Gaël Guennebaud)

23 Range Scanning Problems - Beam location error Assumption: Reflection has Gaussian shape ideal reflection partial occlusion partial illumination Corrupt due to occlusion or texture Biased depth estimate

24 Beam location error Possible solution: Spacetime analysis [Curless, Levoy 95] Estimate when beam centered on spot

25 DIY Range Scanners Laser Range Scanning Cheapo Version [Winkelbach06] hand-held line laser known background geometry need two planes that are not co-planar known camera calibration compute laser plane from lines on the background planes triangulate by ray-plane intersection

26 Active Light - 3D Scanning Structured Light

27 Structured Light Idea: increase speed by projecting multiple stripes But which stripe is which? Answer: Coded stripe patterns Binary code Gray code Color code

28 Range scanning systems Active: Structured light source Analyze image of stripe pattern

29 Structured Lighting: Swept-Planes Revisited Point Grey Flea2 ( x 768) Mitsubishi XD300U ( x 768) Swept-plane scanning recovers 3D depth using ray-plane intersection Use a data projector to replace manually-swept laser/shadow planes How to assign correspondene from projector planes to camera pixels? Solution: Project a spatially- and temporally-encoded image sequence What is the optimal image sequence to project?

30 Structured Lighting: Binary Codes Point Grey Flea2 ( x 768) Mitsubishi XD300U ( x 768) Binary Image Sequence [Posdamer and Altschuler 1982] Each image is a bit-plane of the binary code for projector row/column Minimum of 10 images to encode 1024 columns or 768 rows In practice, 20 images are used to encode 1024 columns or 768 rows Projector and camera(s) must be synchronized

31 Structured Light - Binary Coding Example:3 binary-encoded patterns which allows the measuring surface to be divided in 8 sub-regions projected over time pattern 3 pattern 2 pattern 1

32 Structured Light - Binary Coding Assign each stripe a unique illumination code over time [Posdamer 82] Time Space

33 Structured Light - Binary Coding projected over time pattern 3 Codeword of this píxel: identifies the corresponding pattern stripe pattern 2 pattern 1

34 Structured Lighting: Gray Codes Point Grey Flea2 ( x 768) Mitsubishi XD300U ( x 768) Gray Code Image Sequence Each image is a bit-plane of the Gray code for each projector row/column Requires same number of images as a binary image sequence, but has better performance in practice [Inokuchi 1984] Bin2Gray(B,G) 1 G B 2 for i n-1 downto 0 3 G[i] B[i+1] xor B[i]

35 Binary vs Gray Codes Decimal Binary Gray Code

36 Binary vs Gray Codes Binary Gray Code pattern 3 pattern 2 pattern 1

37 Gray Codes: Decoding Performance Recovered Rows Recovered Columns 3D Reconstruction using Structured Light [Inokuchi 1984] Implemented using a total of 42 images (2 to measure dynamic range, 20 to encode rows, 20 to encode columns) Individual bits assigned by detecting if bit-plane (or its inverse) is brighter Decoding algorithm: Gray code binary code integer row/column index

38 Calibration of a Projector-Camera System Camera calibration (matlab and OpenCV) Extention for projector calibration (matlab) Idea: Consider the projector as an inverse camera, which maps 2D image points into 3D rays, thus making the calibration of a projector the same as that of a camera

39 Calibration of a Projector-Camera System Calibrate the camera using [Zhang 2000] Recover calibration plane in camera coordinate system Project a checkerboard on the calibration board and detect corners Apply ray-plane intersection to recover 3D position for each projected corner Calibrate the projector using the correspondences between the 2D points of the projector image and the recovered 3D position of projected points (Zhang s method).

40 Dynamic Scanning Single-shot patterns (N-arrays, grids, random, etc.) Spatial encoding strategies [Chen et al. 2007] De Bruijn sequences [Zhang et al. 2002] Pseudorandom and M-arrays [Griffin 1992] J. Salvi, J. Pagès, and J. Batlle. Pattern Codification Strategies in Structured Light Systems. Pattern Recognition, 2004 Phase-shifting [Zhang Ivoet Ihrke al. 2004] / Winter 2013

41 Popular example: Kinect I [Matthew Fisher] [Andres Reza]

42 Continuum of Triangulation Methods Multi-stripe Multi-frame Single-stripe Single-frame Slow, robust Fast, fragile

43 Active Light Photometric Stereo The Basics and Some Hints Towards More Complex Issues

44 Photometric Stereo Idea: One camera, multiple (known) light sources s 1 s 2 n s 3 v in matrix form Lambertian reflection

45 The Reflectance Map Reflectance Map: image intensity of a pixel as a function of surface orientation of the corresponding scene point p, q,1 p s, qs,1 s p, q,1 p q,1 s, s Surface normal Light source direction cos s p, q,1 p, q,1 1 p 1 p 2 2 q q 2 1 p s 2 p q s s 1 p q 1 p 2 s 2 s s q q 2 s 2 s Angle between surface normal and direction to point light source

46 The Reflectance Map II Reflectance map of Lambertian surface for p s 0.2, q 0. 4 s R(p,q): Reflectance for all surface orientations (p,q) for a given light source distribution and a given surface material Lambertian surface R p, q cos s 1 p 1 p 2 q s 2 p q s q 1 p 2 s q 2 s

47 Photometric Stereo Point light source: lines of constant reflectance in reflectance map defined by second-order polynomial For each image point (x,y), the possible surface orientation (p,q) is restricted to lie on the curve defined by the polynomial 3 different point sources give 3 second-order curves

48 Photometric Stereo II Light source from left Light source from top Light source from right

49 Photometric Stereo III p,q: surface normal direction For each pixel, the 3 measured intensities corresponding to the 3 different illumination directions lie on 3 different curves in the reflectance map that intersect in one point

50 Photometric Stereo IV ), ( ), ( q p R y x E 2 2 1,1), ( ˆ q p q p n 2 2 1,1), ( ˆ i i i i i q p q p s i i i E n s ˆ ˆ z y x z y x z y x s s s s s s s s s 3, 3, 3, 2, 2, 2, 1, 1, 1, ˆ ˆ ˆ s s s S E E E E n S E ˆ E S n ˆ 1 Surface normal Point light source direction Lambertian reflection Image irradiance 3 illumination directions/ substitution : surface albedo Linear set of 3 equations per pixel

51 Photometric Stereo Advantages A normal value for each pixel with very few images Approach captures details, like pores, wrinkles, or spots Disadvantages Normals must be integrated to get a 3D model (errors will accumulate) Needs controlled lighting or robust estimation of light direction

52 Photometric Stereo V Real Geometry Normal Map

53 Integrating the normals Normals of a height field surface: Tangent space is spanned by Laplacian operator discretize

54 Integrating the normals Solve system Compute right-hand side numerically Include Boundary conditions: e.g. von Neumann = 0 Modify system above it corresponds to boundary on the

55 Photometric Stereo VI Reconstruct normal directions Integrate normal directions to obtain depth map Triangulate depth map

56 Photometric Stereo Assumptions Assumptions: Surface has Lambertian reflectance characteristics Illumination directions are known Illumination sources are distant from object Secondary illumination is insignificant There are no cast shadows Distant camera (orthographic projection) Extensions Non-local illumination or views Non-Lambertian BRDF

57 Active Light Time-of-Flight Imaging

58 Time-of-Flight Pulse-based time-of-flight scanners [Gvili03] NOT triangulation based short infrared laser pulse is sent from camera reflection is recorded in a very short time frame (ps) results in depth profile (intensity image)

59 Time-of-Flight Pulse-based time-of-flight scanner examples accuracy 1-2 cm in a range of 4 7 m applications: "depth keying" replaces chroma keying 3D interaction large scale 3D scanning (LIDAR light detection and ranging)

60 Canesta/3DV Kinect 2 (?)

61 PMDs Photonic Mixer Devices Also called multi-bucket sensors Chip layout Schematic view Electric symbol Fast on-chip modulation [Luan 01]

62 PMDs Photonic Mixer Devices Working principle: Measure phase difference of emitted, modulated signal and received one [Metrilus] Current hardware: ~20MHz modulation frequency in practice square wave

63 [slide by Felix Heide] Photonic Mixer Device (PMD) sensor: P PMD Sensor dt

64 [slide by Felix Heide] Single-path Time-of-Flight depth imaging: P PMD Sensor dt

65 [slide by Felix Heide] Single-path Time-of-Flight depth imaging: P PMD Sensor dt distance-dependent correlation

66 Time-of-Flight Cameras LIDAR University of Washington NASA PMD Panasonic Fotonic ARTTS MESA See [Kolb et al. 10], EG STAR for more details

67 Active Light Transient Imaging

68 [slide by Felix Heide] What is Transient Imaging? a.k.a. Light-in-Flight Imaging Ultrafast imaging of nonstationary light distribution in scenes Tracking of wavefronts of light as they propagate in the scene Equivalent to imaging ultrashort pulses of light t =

69 [slide by Felix Heide] Applications of Transient Imaging Many applications by analyzing a transient image! Understanding light transport: Surface reflectance capture: Decomposing light transport: Reconstructing hidden object geometry (and motion): Wu et al 2012 Velten et al 2013 Naik et al 2011 Velten et al 2012 Pandharkar et al 2011

70 Visualizing Light in Motion Repetitive Event, 1012 FPS, 672 x 1000 pixel resolution, 1 ns+ capture time Light moves 0.6 mm per frame Ability to see light transport [Velten et al. 13]

71 Streak Cameras Picosecond time resolution 1D: 1x672 pixels Result as 2D image ( Streak Photo ) Hamamatsu [Velten et al. 13]

72 Experimental Setup Ti:Sapphire Laser [Velten et al. 13] Synchronization Camera Object Scanning Unit

73 Actual Setup [Velten et al. 13]

74 Streak Camera Picture Time [Velten et al. 13] Space

75 [Velten et al. 13]

76 [Velten et al. 13]

77 Example Result [Velten et al. 13]

78 [Velten et al. 13]

79 Scattering in Real Scenes [Velten et al. 13]

80 [Velten et al. 13]

81 Transient Imaging With PMD Devices linked to the multipath or mixed pixel problem PMD Panasonic Fotonic ARTTS MESA

82 [slides by Felix Heide] Multi-path contributions: P2 PMD Sensor P1 dt

83 [slides by Felix Heide] Recovering multi-path contributions: P2 PMD Sensor P1 dt

84 [slides by Felix Heide] Recovering multi-path contributions: P2 PMD Sensor P1 dt

85 [slides by Felix Heide] Recovering multi-path contributions: P2 PMD Sensor P1 dt

86 [slides by Felix Heide] Image formation model: Relation of Transient Image to PMD measurement: Linear system is ill-posed: assume model for temporal response (signal sparsity) Mixture Model: Gaussian + Exponential and spatial smoothness solve non-linear system multiple measurements Spatial coherence Temporal model

87 [slides by Felix Heide] PMDTechnologies CamBoard nano 2 Too slow LED unit Fixed frequency PLL 25MHz Control FPGA PMD sensor ADC

88 [slides by Felix Heide] Modifications to the hardware Prototype: Max mod. frequency 180 MHz, stable up to 110 MHz Novatech DDS9m Laser unit ic-hg + LPC826 2-ch. function generator MHz Control FPGA Control MOD PMD sensor TRIG ADC

89 Prototype setup MOD in Used for results in paper Camera Power TRIG out Laser unit Function generator

90 [slides by Felix Heide] Result Discoball : Color coding of strongest component:

91 [slides by Felix Heide] Results Discoball : Different time-steps:

92 [slides by Felix Heide] Results Discoball : Different time-steps:

93 [slides by Felix Heide] Results Discoball : Different time-steps:

94 [slides by Felix Heide] Results Discoball : Different time-steps:

95 [slides by Felix Heide] Results Discoball : Different time-steps:

96 [slides by Felix Heide] Result Bottles : Color coding of strongest component:

97 [slides by Felix Heide] Emerging Technologies prototype Camera unit Light source AD9958 DDS (function generator) MHz Control FPGA Control MOD PMD sensor TRIG ADC

98 [slides by Felix Heide] Camera Light source

99 [slides by Felix Heide] Results: People

100 References Curless, Levoy, Better optical triangulation through spacetime analysis, CVPR 1995 Levoy et al., The digital Michelangelo project: 3D scanning of large statues, SIGGRAPH 2000 [Gvili03] R. Gvili, A. Kaplan, E. Ofek, G. Yahav, "Depth Keying", SPIE ISOE 5006, 2003, pp Luan, Experimental Investigation of Photonic Mixer Device and Development of TOF 3D Ranging Systems Based on PMD Technology, PhD Thesis, Siegen University, 2001 Velten et al. Femto-photography: capturing and visualizing the propagation of light, SIGGRAPH 2013 Heide et al. Low-budget Transient Imaging using Photonic Mixer Devices., SIGGRAPH 2013

Active Light Time-of-Flight Imaging

Active Light Time-of-Flight Imaging Active Light Time-of-Flight Imaging Time-of-Flight Pulse-based time-of-flight scanners [Gvili03] NOT triangulation based short infrared laser pulse is sent from camera reflection is recorded in a very

More information

Structured Light II. Guido Gerig CS 6320, Spring (thanks: slides Prof. S. Narasimhan, CMU, Marc Pollefeys, UNC)

Structured Light II. Guido Gerig CS 6320, Spring (thanks: slides Prof. S. Narasimhan, CMU, Marc Pollefeys, UNC) Structured Light II Guido Gerig CS 6320, Spring 2013 (thanks: slides Prof. S. Narasimhan, CMU, Marc Pollefeys, UNC) http://www.cs.cmu.edu/afs/cs/academic/class/15385- s06/lectures/ppts/lec-17.ppt Variant

More information

L2 Data Acquisition. Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods

L2 Data Acquisition. Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods L2 Data Acquisition Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods 1 Coordinate Measurement Machine Touch based Slow Sparse Data Complex planning Accurate 2

More information

Surround Structured Lighting for Full Object Scanning

Surround Structured Lighting for Full Object Scanning Surround Structured Lighting for Full Object Scanning Douglas Lanman, Daniel Crispell, and Gabriel Taubin Brown University, Dept. of Engineering August 21, 2007 1 Outline Introduction and Related Work

More information

ENGN 2911 I: 3D Photography and Geometry Processing Assignment 2: Structured Light for 3D Scanning

ENGN 2911 I: 3D Photography and Geometry Processing Assignment 2: Structured Light for 3D Scanning ENGN 2911 I: 3D Photography and Geometry Processing Assignment 2: Structured Light for 3D Scanning Instructor: Gabriel Taubin Assignment written by: Douglas Lanman 26 February 2009 Figure 1: Structured

More information

Multi-View Stereo for Community Photo Collections Michael Goesele, et al, ICCV Venus de Milo

Multi-View Stereo for Community Photo Collections Michael Goesele, et al, ICCV Venus de Milo Vision Sensing Multi-View Stereo for Community Photo Collections Michael Goesele, et al, ICCV 2007 Venus de Milo The Digital Michelangelo Project, Stanford How to sense 3D very accurately? How to sense

More information

3D Object Representations. COS 526, Fall 2016 Princeton University

3D Object Representations. COS 526, Fall 2016 Princeton University 3D Object Representations COS 526, Fall 2016 Princeton University 3D Object Representations How do we... Represent 3D objects in a computer? Acquire computer representations of 3D objects? Manipulate computer

More information

Computer Vision. 3D acquisition

Computer Vision. 3D acquisition è Computer 3D acquisition Acknowledgement Courtesy of Prof. Luc Van Gool 3D acquisition taxonomy s image cannot currently be displayed. 3D acquisition methods Thi passive active uni-directional multi-directional

More information

3D Photography: Stereo

3D Photography: Stereo 3D Photography: Stereo Marc Pollefeys, Torsten Sattler Spring 2016 http://www.cvg.ethz.ch/teaching/3dvision/ 3D Modeling with Depth Sensors Today s class Obtaining depth maps / range images unstructured

More information

3D Scanning. Qixing Huang Feb. 9 th Slide Credit: Yasutaka Furukawa

3D Scanning. Qixing Huang Feb. 9 th Slide Credit: Yasutaka Furukawa 3D Scanning Qixing Huang Feb. 9 th 2017 Slide Credit: Yasutaka Furukawa Geometry Reconstruction Pipeline This Lecture Depth Sensing ICP for Pair-wise Alignment Next Lecture Global Alignment Pairwise Multiple

More information

3D Photography: Active Ranging, Structured Light, ICP

3D Photography: Active Ranging, Structured Light, ICP 3D Photography: Active Ranging, Structured Light, ICP Kalin Kolev, Marc Pollefeys Spring 2013 http://cvg.ethz.ch/teaching/2013spring/3dphoto/ Schedule (tentative) Feb 18 Feb 25 Mar 4 Mar 11 Mar 18 Mar

More information

Structured light , , Computational Photography Fall 2017, Lecture 27

Structured light , , Computational Photography Fall 2017, Lecture 27 Structured light http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 27 Course announcements Homework 5 has been graded. - Mean: 129. - Median:

More information

Optical Active 3D Scanning. Gianpaolo Palma

Optical Active 3D Scanning. Gianpaolo Palma Optical Active 3D Scanning Gianpaolo Palma 3D Scanning Taxonomy SHAPE ACQUISTION CONTACT NO-CONTACT NO DESTRUCTIVE DESTRUCTIVE X-RAY MAGNETIC OPTICAL ACOUSTIC CMM ROBOTIC GANTRY SLICING ACTIVE PASSIVE

More information

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe Structured Light Tobias Nöll tobias.noell@dfki.de Thanks to Marc Pollefeys, David Nister and David Lowe Introduction Previous lecture: Dense reconstruction Dense matching of non-feature pixels Patch-based

More information

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman Stereo 11/02/2012 CS129, Brown James Hays Slides by Kristen Grauman Multiple views Multi-view geometry, matching, invariant features, stereo vision Lowe Hartley and Zisserman Why multiple views? Structure

More information

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013 Lecture 19: Depth Cameras Visual Computing Systems Continuing theme: computational photography Cameras capture light, then extensive processing produces the desired image Today: - Capturing scene depth

More information

Stereo and structured light

Stereo and structured light Stereo and structured light http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 20 Course announcements Homework 5 is still ongoing. - Make sure

More information

3D Computer Vision. Structured Light I. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light I. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light I Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from? Binocular Stereo Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image Where does the depth information come from? Binocular stereo Given a calibrated binocular stereo

More information

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller 3D Computer Vision Depth Cameras Prof. Didier Stricker Oliver Wasenmüller Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de

More information

Overview of Active Vision Techniques

Overview of Active Vision Techniques SIGGRAPH 99 Course on 3D Photography Overview of Active Vision Techniques Brian Curless University of Washington Overview Introduction Active vision techniques Imaging radar Triangulation Moire Active

More information

TRANSIENT IMAGING. Freshman Imaging Project 2014

TRANSIENT IMAGING. Freshman Imaging Project 2014 1 TRANSIENT IMAGING Freshman Imaging Project 2014 2 Outline What is transient imaging? Previous work: MIT, UBC Possible applications of transient imaging Theory behind transient imaging Our goals Technology

More information

Other Reconstruction Techniques

Other Reconstruction Techniques Other Reconstruction Techniques Ruigang Yang CS 684 CS 684 Spring 2004 1 Taxonomy of Range Sensing From Brain Curless, SIGGRAPH 00 Lecture notes CS 684 Spring 2004 2 Taxonomy of Range Scanning (cont.)

More information

Range Sensors (time of flight) (1)

Range Sensors (time of flight) (1) Range Sensors (time of flight) (1) Large range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic sensors, infra-red sensors

More information

Perspective projection

Perspective projection Sensors Ioannis Stamos Perspective projection Pinhole & the Perspective Projection (x,y) SCENE SCREEN Is there an image being formed on the screen?

More information

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze

More information

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical

More information

3D Scanning. Lecture courtesy of Szymon Rusinkiewicz Princeton University

3D Scanning. Lecture courtesy of Szymon Rusinkiewicz Princeton University 3D Scanning Lecture courtesy of Szymon Rusinkiewicz Princeton University Computer Graphics Pipeline 3D Scanning Shape Motion Rendering Lighting and Reflectance Human time = expensive Sensors = cheap Computer

More information

A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect

A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect Diana Beltran and Luis Basañez Technical University of Catalonia, Barcelona, Spain {diana.beltran,luis.basanez}@upc.edu

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

CSc Topics in Computer Graphics 3D Photography

CSc Topics in Computer Graphics 3D Photography CSc 83010 Topics in Computer Graphics 3D Photography Tuesdays 11:45-1:45 1:45 Room 3305 Ioannis Stamos istamos@hunter.cuny.edu Office: 1090F, Hunter North (Entrance at 69 th bw/ / Park and Lexington Avenues)

More information

Image-based modeling (IBM) and image-based rendering (IBR)

Image-based modeling (IBM) and image-based rendering (IBR) Image-based modeling (IBM) and image-based rendering (IBR) CS 248 - Introduction to Computer Graphics Autumn quarter, 2005 Slides for December 8 lecture The graphics pipeline modeling animation rendering

More information

Stereo vision. Many slides adapted from Steve Seitz

Stereo vision. Many slides adapted from Steve Seitz Stereo vision Many slides adapted from Steve Seitz What is stereo vision? Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape What is

More information

The main problem of photogrammetry

The main problem of photogrammetry Structured Light Structured Light The main problem of photogrammetry to recover shape from multiple views of a scene, we need to find correspondences between the images the matching/correspondence problem

More information

Active Stereo Vision. COMP 4900D Winter 2012 Gerhard Roth

Active Stereo Vision. COMP 4900D Winter 2012 Gerhard Roth Active Stereo Vision COMP 4900D Winter 2012 Gerhard Roth Why active sensors? Project our own texture using light (usually laser) This simplifies correspondence problem (much easier) Pluses Can handle different

More information

Lecture 10: Multi view geometry

Lecture 10: Multi view geometry Lecture 10: Multi view geometry Professor Fei Fei Li Stanford Vision Lab 1 What we will learn today? Stereo vision Correspondence problem (Problem Set 2 (Q3)) Active stereo vision systems Structure from

More information

10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1.

10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1. Range Sensors (time of flight) (1) Range Sensors (time of flight) (2) arge range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic

More information

Traditional pipeline for shape capture. Active range scanning. New pipeline for shape capture. Draw contours on model. Digitize lines with touch probe

Traditional pipeline for shape capture. Active range scanning. New pipeline for shape capture. Draw contours on model. Digitize lines with touch probe Traditional pipeline for shape capture Draw contours on model Active range scanning Digitize lines with touch probe Fit smooth surfaces CAD/CAM Texturing and animation 1 2 New pipeline for shape capture

More information

Outline. ETN-FPI Training School on Plenoptic Sensing

Outline. ETN-FPI Training School on Plenoptic Sensing Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration

More information

Stereo. Many slides adapted from Steve Seitz

Stereo. Many slides adapted from Steve Seitz Stereo Many slides adapted from Steve Seitz Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image image 1 image 2 Dense depth map Binocular stereo Given a calibrated

More information

ENGN D Photography / Spring 2018 / SYLLABUS

ENGN D Photography / Spring 2018 / SYLLABUS ENGN 2502 3D Photography / Spring 2018 / SYLLABUS Description of the proposed course Over the last decade digital photography has entered the mainstream with inexpensive, miniaturized cameras routinely

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

Other approaches to obtaining 3D structure

Other approaches to obtaining 3D structure Other approaches to obtaining 3D structure Active stereo with structured light Project structured light patterns onto the object simplifies the correspondence problem Allows us to use only one camera camera

More information

Structured light 3D reconstruction

Structured light 3D reconstruction Structured light 3D reconstruction Reconstruction pipeline and industrial applications rodola@dsi.unive.it 11/05/2010 3D Reconstruction 3D reconstruction is the process of capturing the shape and appearance

More information

Surround Structured Lighting for Full Object Scanning

Surround Structured Lighting for Full Object Scanning Surround Structured Lighting for Full Object Scanning Douglas Lanman, Daniel Crispell, and Gabriel Taubin Department of Engineering, Brown University {dlanman,daniel crispell,taubin}@brown.edu Abstract

More information

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems EECS 442 Computer vision Stereo systems Stereo vision Rectification Correspondence problem Active stereo vision systems Reading: [HZ] Chapter: 11 [FP] Chapter: 11 Stereo vision P p p O 1 O 2 Goal: estimate

More information

3D Time-of-Flight Image Sensor Solutions for Mobile Devices

3D Time-of-Flight Image Sensor Solutions for Mobile Devices 3D Time-of-Flight Image Sensor Solutions for Mobile Devices SEMICON Europa 2015 Imaging Conference Bernd Buxbaum 2015 pmdtechnologies gmbh c o n f i d e n t i a l Content Introduction Motivation for 3D

More information

2 Depth Camera Assessment

2 Depth Camera Assessment 2 Depth Camera Assessment The driving question of this chapter is how competitive cheap consumer depth cameras, namely the Microsoft Kinect and the SoftKinetic DepthSense, are compared to state-of-the-art

More information

3D Modeling of Objects Using Laser Scanning

3D Modeling of Objects Using Laser Scanning 1 3D Modeling of Objects Using Laser Scanning D. Jaya Deepu, LPU University, Punjab, India Email: Jaideepudadi@gmail.com Abstract: In the last few decades, constructing accurate three-dimensional models

More information

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography Computational Photography Matthias Zwicker University of Bern Fall 2009 Today From 2D to 3D using multiple views Introduction Geometry of two views Stereo matching Other applications Multiview geometry

More information

Spacetime Stereo: A Unifying Framework for Depth from Triangulation. Abstract

Spacetime Stereo: A Unifying Framework for Depth from Triangulation. Abstract Spacetime Stereo: A Unifying Framework for Depth from Triangulation James Davis Diego Nehab Ravi Ramamoorthi Szymon Rusinkiewicz Honda Research Institute Princeton University Columbia University Princeton

More information

Computational Imaging for Self-Driving Vehicles

Computational Imaging for Self-Driving Vehicles CVPR 2018 Computational Imaging for Self-Driving Vehicles Jan Kautz--------Ramesh Raskar--------Achuta Kadambi--------Guy Satat Computational Imaging for Self-Driving Vehicles Jan Kautz--------Ramesh Raskar--------Achuta

More information

Non-line-of-sight imaging

Non-line-of-sight imaging Non-line-of-sight imaging http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 25 Course announcements Homework 6 will be posted tonight. - Will

More information

Robust and Accurate One-shot 3D Reconstruction by 2C1P System with Wave Grid Pattern

Robust and Accurate One-shot 3D Reconstruction by 2C1P System with Wave Grid Pattern Robust and Accurate One-shot 3D Reconstruction by 2C1P System with Wave Grid Pattern Nozomu Kasuya Kagoshima University Kagoshima, Japan AIST Tsukuba, Japan nozomu.kasuya@aist.go.jp Ryusuke Sagawa AIST

More information

Depth Camera for Mobile Devices

Depth Camera for Mobile Devices Depth Camera for Mobile Devices Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Stereo Cameras Structured Light Cameras Time of Flight (ToF) Camera Inferring 3D Points Given we have

More information

Lecture 8 Active stereo & Volumetric stereo

Lecture 8 Active stereo & Volumetric stereo Lecture 8 Active stereo & Volumetric stereo In this lecture, we ll first discuss another framework for describing stereo systems called active stereo, and then introduce the problem of volumetric stereo,

More information

Depth Sensors Kinect V2 A. Fornaser

Depth Sensors Kinect V2 A. Fornaser Depth Sensors Kinect V2 A. Fornaser alberto.fornaser@unitn.it Vision Depth data It is not a 3D data, It is a map of distances Not a 3D, not a 2D it is a 2.5D or Perspective 3D Complete 3D - Tomography

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction

More information

3D shape from the structure of pencils of planes and geometric constraints

3D shape from the structure of pencils of planes and geometric constraints 3D shape from the structure of pencils of planes and geometric constraints Paper ID: 691 Abstract. Active stereo systems using structured light has been used as practical solutions for 3D measurements.

More information

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and

More information

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few... STEREO VISION The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Bill Freeman and Antonio Torralba (MIT), including their own

More information

3D Computer Vision 1

3D Computer Vision 1 3D Computer Vision 1 Multiview Stereo Multiview Stereo Multiview Stereo https://www.youtube.com/watch?v=ugkb7itpnae Shape from silhouette Shape from silhouette Shape from silhouette Shape from silhouette

More information

3D photography. Digital Visual Effects, Spring 2007 Yung-Yu Chuang 2007/5/15

3D photography. Digital Visual Effects, Spring 2007 Yung-Yu Chuang 2007/5/15 3D photography Digital Visual Effects, Spring 2007 Yung-Yu Chuang 2007/5/15 with slides by Szymon Rusinkiewicz, Richard Szeliski, Steve Seitz and Brian Curless Announcements Project #3 is due on 5/20.

More information

Handy Rangefinder for Active Robot Vision

Handy Rangefinder for Active Robot Vision Handy Rangefinder for Active Robot Vision Kazuyuki Hattori Yukio Sato Department of Electrical and Computer Engineering Nagoya Institute of Technology Showa, Nagoya 466, Japan Abstract A compact and high-speed

More information

CS5670: Computer Vision

CS5670: Computer Vision CS5670: Computer Vision Noah Snavely, Zhengqi Li Stereo Single image stereogram, by Niklas Een Mark Twain at Pool Table", no date, UCR Museum of Photography Stereo Given two images from different viewpoints

More information

Photometric Stereo. Lighting and Photometric Stereo. Computer Vision I. Last lecture in a nutshell BRDF. CSE252A Lecture 7

Photometric Stereo. Lighting and Photometric Stereo. Computer Vision I. Last lecture in a nutshell BRDF. CSE252A Lecture 7 Lighting and Photometric Stereo Photometric Stereo HW will be on web later today CSE5A Lecture 7 Radiometry of thin lenses δa Last lecture in a nutshell δa δa'cosα δacos β δω = = ( z' / cosα ) ( z / cosα

More information

Lecture 9 & 10: Stereo Vision

Lecture 9 & 10: Stereo Vision Lecture 9 & 10: Stereo Vision Professor Fei- Fei Li Stanford Vision Lab 1 What we will learn today? IntroducEon to stereo vision Epipolar geometry: a gentle intro Parallel images Image receficaeon Solving

More information

Multiple View Geometry

Multiple View Geometry Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V

More information

Non-line-of-sight imaging

Non-line-of-sight imaging Non-line-of-sight imaging http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 23 Course announcements Homework 6 posted, due November 30 th. -

More information

Sensing Deforming and Moving Objects with Commercial Off the Shelf Hardware

Sensing Deforming and Moving Objects with Commercial Off the Shelf Hardware Sensing Deforming and Moving Objects with Commercial Off the Shelf Hardware This work supported by: Philip Fong Florian Buron Stanford University Motivational Applications Human tissue modeling for surgical

More information

Metrology and Sensing

Metrology and Sensing Metrology and Sensing Lecture 4: Fringe projection 2016-11-08 Herbert Gross Winter term 2016 www.iap.uni-jena.de 2 Preliminary Schedule No Date Subject Detailed Content 1 18.10. Introduction Introduction,

More information

Optical Imaging Techniques and Applications

Optical Imaging Techniques and Applications Optical Imaging Techniques and Applications Jason Geng, Ph.D. Vice President IEEE Intelligent Transportation Systems Society jason.geng@ieee.org Outline Structured light 3D surface imaging concept Classification

More information

Midterm Examination CS 534: Computational Photography

Midterm Examination CS 534: Computational Photography Midterm Examination CS 534: Computational Photography November 3, 2016 NAME: Problem Score Max Score 1 6 2 8 3 9 4 12 5 4 6 13 7 7 8 6 9 9 10 6 11 14 12 6 Total 100 1 of 8 1. [6] (a) [3] What camera setting(s)

More information

Registration of Moving Surfaces by Means of One-Shot Laser Projection

Registration of Moving Surfaces by Means of One-Shot Laser Projection Registration of Moving Surfaces by Means of One-Shot Laser Projection Carles Matabosch 1,DavidFofi 2, Joaquim Salvi 1, and Josep Forest 1 1 University of Girona, Institut d Informatica i Aplicacions, Girona,

More information

Photometric Stereo.

Photometric Stereo. Photometric Stereo Photometric Stereo v.s.. Structure from Shading [1] Photometric stereo is a technique in computer vision for estimating the surface normals of objects by observing that object under

More information

Light Transport CS434. Daniel G. Aliaga Department of Computer Science Purdue University

Light Transport CS434. Daniel G. Aliaga Department of Computer Science Purdue University Light Transport CS434 Daniel G. Aliaga Department of Computer Science Purdue University Topics Local and Global Illumination Models Helmholtz Reciprocity Dual Photography/Light Transport (in Real-World)

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Lecture 10: Multi-view geometry

Lecture 10: Multi-view geometry Lecture 10: Multi-view geometry Professor Stanford Vision Lab 1 What we will learn today? Review for stereo vision Correspondence problem (Problem Set 2 (Q3)) Active stereo vision systems Structure from

More information

Compact and Low Cost System for the Measurement of Accurate 3D Shape and Normal

Compact and Low Cost System for the Measurement of Accurate 3D Shape and Normal Compact and Low Cost System for the Measurement of Accurate 3D Shape and Normal Ryusuke Homma, Takao Makino, Koichi Takase, Norimichi Tsumura, Toshiya Nakaguchi and Yoichi Miyake Chiba University, Japan

More information

Volumetric Scene Reconstruction from Multiple Views

Volumetric Scene Reconstruction from Multiple Views Volumetric Scene Reconstruction from Multiple Views Chuck Dyer University of Wisconsin dyer@cs cs.wisc.edu www.cs cs.wisc.edu/~dyer Image-Based Scene Reconstruction Goal Automatic construction of photo-realistic

More information

Resolution adjustable 3D scanner based on using stereo cameras

Resolution adjustable 3D scanner based on using stereo cameras Resolution adjustable D scanner based on using stereo cameras zung-han Lin * * Graduate Institute of Color and Illumination echnology, National aiwan University of Science and echnology, aiwan. E-mail:

More information

Recap from Previous Lecture

Recap from Previous Lecture Recap from Previous Lecture Tone Mapping Preserve local contrast or detail at the expense of large scale contrast. Changing the brightness within objects or surfaces unequally leads to halos. We are now

More information

Project 2 due today Project 3 out today. Readings Szeliski, Chapter 10 (through 10.5)

Project 2 due today Project 3 out today. Readings Szeliski, Chapter 10 (through 10.5) Announcements Stereo Project 2 due today Project 3 out today Single image stereogram, by Niklas Een Readings Szeliski, Chapter 10 (through 10.5) Public Library, Stereoscopic Looking Room, Chicago, by Phillips,

More information

Multiple View Geometry

Multiple View Geometry Multiple View Geometry CS 6320, Spring 2013 Guest Lecture Marcel Prastawa adapted from Pollefeys, Shah, and Zisserman Single view computer vision Projective actions of cameras Camera callibration Photometric

More information

Lecture 8 Active stereo & Volumetric stereo

Lecture 8 Active stereo & Volumetric stereo Lecture 8 Active stereo & Volumetric stereo Active stereo Structured lighting Depth sensing Volumetric stereo: Space carving Shadow carving Voxel coloring Reading: [Szelisky] Chapter 11 Multi-view stereo

More information

Introduction to Computer Vision

Introduction to Computer Vision Introduction to Computer Vision Michael J. Black Nov 2009 Perspective projection and affine motion Goals Today Perspective projection 3D motion Wed Projects Friday Regularization and robust statistics

More information

Basilio Bona DAUIN Politecnico di Torino

Basilio Bona DAUIN Politecnico di Torino ROBOTICA 03CFIOR DAUIN Politecnico di Torino Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The

More information

Project 3 code & artifact due Tuesday Final project proposals due noon Wed (by ) Readings Szeliski, Chapter 10 (through 10.5)

Project 3 code & artifact due Tuesday Final project proposals due noon Wed (by  ) Readings Szeliski, Chapter 10 (through 10.5) Announcements Project 3 code & artifact due Tuesday Final project proposals due noon Wed (by email) One-page writeup (from project web page), specifying:» Your team members» Project goals. Be specific.

More information

Epipolar geometry contd.

Epipolar geometry contd. Epipolar geometry contd. Estimating F 8-point algorithm The fundamental matrix F is defined by x' T Fx = 0 for any pair of matches x and x in two images. Let x=(u,v,1) T and x =(u,v,1) T, each match gives

More information

3D scanning. 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects.

3D scanning. 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects. Acquiring 3D shape 3D scanning 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects. The produced digital model is formed by geometric

More information

CS4670/5760: Computer Vision Kavita Bala Scott Wehrwein. Lecture 23: Photometric Stereo

CS4670/5760: Computer Vision Kavita Bala Scott Wehrwein. Lecture 23: Photometric Stereo CS4670/5760: Computer Vision Kavita Bala Scott Wehrwein Lecture 23: Photometric Stereo Announcements PA3 Artifact due tonight PA3 Demos Thursday Signups close at 4:30 today No lecture on Friday Last Time:

More information

Epipolar Geometry and Stereo Vision

Epipolar Geometry and Stereo Vision Epipolar Geometry and Stereo Vision Computer Vision Jia-Bin Huang, Virginia Tech Many slides from S. Seitz and D. Hoiem Last class: Image Stitching Two images with rotation/zoom but no translation. X x

More information

CSE 4392/5369. Dr. Gian Luca Mariottini, Ph.D.

CSE 4392/5369. Dr. Gian Luca Mariottini, Ph.D. University of Texas at Arlington CSE 4392/5369 Introduction to Vision Sensing Dr. Gian Luca Mariottini, Ph.D. Department of Computer Science and Engineering University of Texas at Arlington WEB : http://ranger.uta.edu/~gianluca

More information

DD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication

DD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication DD2423 Image Analysis and Computer Vision IMAGE FORMATION Mårten Björkman Computational Vision and Active Perception School of Computer Science and Communication November 8, 2013 1 Image formation Goal:

More information

What have we leaned so far?

What have we leaned so far? What have we leaned so far? Camera structure Eye structure Project 1: High Dynamic Range Imaging What have we learned so far? Image Filtering Image Warping Camera Projection Model Project 2: Panoramic

More information

Acquisition and Visualization of Colored 3D Objects

Acquisition and Visualization of Colored 3D Objects Acquisition and Visualization of Colored 3D Objects Kari Pulli Stanford University Stanford, CA, U.S.A kapu@cs.stanford.edu Habib Abi-Rached, Tom Duchamp, Linda G. Shapiro and Werner Stuetzle University

More information

Multi-View 3D-Reconstruction

Multi-View 3D-Reconstruction Multi-View 3D-Reconstruction Cedric Cagniart Computer Aided Medical Procedures (CAMP) Technische Universität München, Germany 1 Problem Statement Given several calibrated views of an object... can we automatically

More information

Ping Tan. Simon Fraser University

Ping Tan. Simon Fraser University Ping Tan Simon Fraser University Photos vs. Videos (live photos) A good photo tells a story Stories are better told in videos Videos in the Mobile Era (mobile & share) More videos are captured by mobile

More information

Understanding Variability

Understanding Variability Understanding Variability Why so different? Light and Optics Pinhole camera model Perspective projection Thin lens model Fundamental equation Distortion: spherical & chromatic aberration, radial distortion

More information

Metrology and Sensing

Metrology and Sensing Metrology and Sensing Lecture 4: Fringe projection 2018-11-09 Herbert Gross Winter term 2018 www.iap.uni-jena.de 2 Schedule Optical Metrology and Sensing 2018 No Date Subject Detailed Content 1 16.10.

More information