Advanced Digital Sensors/Photography Geometry Capture Lecture #10 September 21, 2017 Donald P. Greenberg

Similar documents
Advanced Digital Photography Geometry and Motion Capture Lecture #11 September 29, 2015 Donald P. Greenberg

Advanced Digital Photography and Geometry Capture. Visual Imaging in the Electronic Age Lecture #10 Donald P. Greenberg September 24, 2015

Advanced Digital Photography and Geometry Capture. Visual Imaging in the Electronic Age Lecture #10 Donald P. Greenberg September 24, 2015

Visual Imaging in the Electronic Age Assignment #3 Draft Geometry Capture

Amateur photography was once largely a personal RECONSTRUCTING ROME

Comparing Aerial Photogrammetry and 3D Laser Scanning Methods for Creating 3D Models of Complex Objects

Introduction to Computer Vision. Srikumar Ramalingam School of Computing University of Utah

Large Scale 3D Reconstruction by Structure from Motion

Comparing Aerial Photogrammetry and 3D Laser Scanning Methods for Creating 3D Models of Complex Objects

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Chaplin, Modern Times, 1936

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Solid-State Hybrid LiDAR for Autonomous Driving Product Description

Epipolar Geometry and Stereo Vision

Computer Vision. 3D acquisition

Image correspondences and structure from motion

Multiple View Geometry

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F

The raycloud A Vision Beyond the Point Cloud

AUTOMATED 3D MODELING OF URBAN ENVIRONMENTS

CSc Topics in Computer Graphics 3D Photography

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

Step-by-Step Model Buidling

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Optical Sensors: Key Technology for the Autonomous Car

High Definition Modeling of Calw, Badstrasse and its Google Earth Integration

Depth Sensors Kinect V2 A. Fornaser

Introduction to 3D Machine Vision

Image Based Reconstruction II

Homographies and RANSAC

CS5670: Computer Vision

EVOLUTION OF POINT CLOUD

LASER SCANNING AND PHOTOGRAMMETRY: A HYBRID APPROACH FOR HERITAGE DOCUMENTATION

DETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS

3D Modeling of Objects Using Laser Scanning

Final project bits and pieces

3D Fusion of Infrared Images with Dense RGB Reconstruction from Multiple Views - with Application to Fire-fighting Robots

3D Scanning. Qixing Huang Feb. 9 th Slide Credit: Yasutaka Furukawa

Photo Tourism: Exploring Photo Collections in 3D

Ninio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29,

Miniature faking. In close-up photo, the depth of field is limited.

Computer Vision Lecture 17

Computer Vision Lecture 17

Trimble Engineering & Construction Group, 5475 Kellenburger Road, Dayton, OH , USA

Mobile Mapping and Navigation. Brad Kohlmeyer NAVTEQ Research

The Light Field and Image-Based Rendering

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac

Noah Snavely Steven M. Seitz. Richard Szeliski. University of Washington. Microsoft Research. Modified from authors slides

Image-Based Modeling and Rendering

Sensing Deforming and Moving Objects with Commercial Off the Shelf Hardware

3D Photography: Active Ranging, Structured Light, ICP

Collaborative Mapping with Streetlevel Images in the Wild. Yubin Kuang Co-founder and Computer Vision Lead

COMPARISON OF LASER SCANNING, PHOTOGRAMMETRY AND SfM-MVS PIPELINE APPLIED IN STRUCTURES AND ARTIFICIAL SURFACES

A Whole New World of Mapping and Sensing: Uses from Asset to Management to In Vehicle Sensing for Collision Avoidance

Pattern Recognition for Autonomous. Pattern Recognition for Autonomous. Driving. Freie Universität t Berlin. Raul Rojas

SIMPLE ROOM SHAPE MODELING WITH SPARSE 3D POINT INFORMATION USING PHOTOGRAMMETRY AND APPLICATION SOFTWARE

Image-Based Rendering

AUTOMOTIVE ENVIRONMENT SENSORS

Active Stereo Vision. COMP 4900D Winter 2012 Gerhard Roth

Photo Tourism: Exploring Photo Collections in 3D

CS4670: Computer Vision

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Camera Calibration. COS 429 Princeton University

Epipolar Geometry and Stereo Vision

TLS Parameters, Workflows and Field Methods

Multi-view stereo. Many slides adapted from S. Seitz

EE795: Computer Vision and Intelligent Systems

Large Scale 3D Reconstruction (50 mins) Yasutaka Washington University in St. Louis

Image-Based Modeling and Rendering

Overview of Active Vision Techniques

TLS Parameters, Workflows and Field Methods

TLS Parameters, Workflows and Field Methods

Recap from Previous Lecture

Project 3 code & artifact due Tuesday Final project proposals due noon Wed (by ) Readings Szeliski, Chapter 10 (through 10.5)

What is Computer Vision? Introduction. We all make mistakes. Why is this hard? What was happening. What do you see? Intro Computer Vision

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

3D Object Representations. COS 526, Fall 2016 Princeton University

Dense 3D Reconstruction. Christiano Gava

Course Outline (1) #6 Data Acquisition for Built Environment. Fumio YAMAZAKI

CS6670: Computer Vision

ENY-C2005 Geoinformation in Environmental Modeling Lecture 4b: Laser scanning

Photogrammetry: DTM Extraction & Editing

CS5670: Computer Vision

12/3/2009. What is Computer Vision? Applications. Application: Assisted driving Pedestrian and car detection. Application: Improving online search

Acoustic/Lidar Sensor Fusion for Car Tracking in City Traffic Scenarios

CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

Advanced Driver Assistance: Modular Image Sensor Concept

Leica Systems Overview

3D Photography: Stereo

Stereo vision. Many slides adapted from Steve Seitz

3D BUILDINGS MODELLING BASED ON A COMBINATION OF TECHNIQUES AND METHODOLOGIES

Photo Tourism: Exploring Photo Collections in 3D

An Introduction to Lidar & Forestry May 2013

Applications of Mobile LiDAR and UAV Sourced Photogrammetry

Solid State LiDAR for Ubiquitous 3D Sensing

Camera Drones Lecture 3 3D data generation

Option Driver Assistance. Product Information

Lecture 10: Multi view geometry

CS4670: Computer Vision

Transcription:

Visual Imaging in the Electronic Age Advanced Digital Sensors/Photography Geometry Capture Lecture #10 September 21, 2017 Donald P. Greenberg

Digital Geometry Capture Photographic methods Laser scanning Time of Flight Sensors

Camera Definition Camera View direction Model The camera location, view direction, and frustum must be defined relative to the object.

Donald P. Greenberg - Cornell Program of Computer Graphics Mapping a Viewing Frustum to a Standard Viewbox Z e Y e Y s Z s X s X e z e = D Frustum of vision z e = F 1 1 1 Screen coordinate system

In general, what are the unknown variables? Unknown observer position- Xe, Ye, Ze Unknown viewer direction- Φ, θ, Ψ Unknown focal length- f

Early Work - 1975

Computer Vision The science and technology of machines that see Can the machine extract desired information from an image? What did I see?

Simple case Known camera positions (x e, y e, z e ), camera optics, Identify corresponding points in each image. Jeremiah Fairbank. View dependent perspective images. Master's thesis, Cornell University, August 2005.

Sagan House

Sagan House

ReMake Autodesk

ReMake Autodesk

Reconstructing Rome 1 The advent of digital photography and the recent growth of photo-sharing websites ( ) have brought about the seismic change in photography and the use of photo collections. 1 A search for the word Rome on returns two million photos. This collection, or others like it, capture every popular site, facade, statue, fountain, interior, café, etc.

Characteristics of Typical Photo Sets The photos are unstructured No particular order or distribution of camera viewpoints The photos are uncalibrated Nothing is known about the camera settings (exposure, focal length, etc.) The scale is enormous (millions, not thousands of photos) and We need to do this fast!

Correspondence and 3D Structure from Different Camera Positions Note: The pictures are in correspondence 2D dots with same color correspond to the same 3D points.

3D Structure from Different Camera Positions Camera 1 Camera 2 Camera 3

3D Structure from Different Camera Positions Camera 3 Camera 1 Camera 2 Assuming the position of the red dot is known, there is reprojection error in Camera 3.

Change the Problem to an optimization problem Minimize the sum of the squares of the reprojection errors. This non-linear least squares problem is difficult to solve due to local minima and maxima. Authors selectively started with a few choice cameras and points and grew scenes incrementally (a process known as bundle adjustment).

Trevi Fountain, Rome Italy http://en.wikipedia.org/wiki/trevi_fountain

Feature Detection and Matching A track corresponding to a point on the face of the central statue of Oceanus at the Trevi Fountain, the embodiment of a river encircling the world in Greek mythology. Sameer Agarwal, Yasutaka Furukawa, Naoh Snavely, Brian Curless, Steve M. Seitz, Richard Szeliski. Reconstructing Rome, IEEE Computer, June 2010.

Colosseum The Colosseum (Rome) Reconstructed dense 3D point models. For places with many available images, reconstruction quality is very high. Sameer Agarwal, Yasutaka Furukawa, Naoh Snavely, Brian Curless, Steve M. Seitz, Richard Szeliski. Reconstructing Rome, IEEE Computer, June 2010.

Cornell Campus, McGraw Hall - Noah Snavely

Digital Geometry Capture Photographic methods Laser scanning Time of Flight Sensors

Cyberware Scanner

Cyberware Scanner Diagram

Cyberware Scanner

Uncle Don

Cyberware vs. 1 2 3 Catch

Digital Geometry Capture Photographic methods Laser scanning Time of Flight Sensors

Pulsed Modulation

Kinect for xbox 360

Matterport 2016

Google Street View The world contains roughly 50 million miles of roads, paved and unpaved, across 219 countries (ref.) This is equivalent to circumnavigating the globe 1250 times. To date, hundreds of cities in many countries across four continents have been captured. Google has developed several vehicular platforms and texture information in the project s seven year history. Drafomir Anguelov, Carole Dulong, Daniel Filip, Christian Frueh, Stepheane Lafon, Richard Lyon, Abhijit Ogale, Luc Vincent, Josh Weaver. Google Street View: Capturing The World At Street Level, IEEE Computer, June 2010.

Google Street View and Google Maps 2007-2012 In 2007, Larry Page requests Thrun and Levandowski to create a virtual map of the U.S. Engineers jury-rigged some vans with GPS and rooftop cameras which shot 360 panoramas for any address. They equipped 100 cars which were sent around the U.S. Data was put together with a program written by Marc Levoy. In 2011, Google announced it would start charging (large) commercial sites In 2012, Google allows users to post photos and reviews of locations By October 2012, Google will have updated 250,000 miles of U.S. roads Note: They have also added Google Moon and Google Mars

R7 Street View Camera System - 2009 The system is a rosette of 15 small, outward-looking cameras using 5-megapixel CMOS image sensors and custom, low-flare, controlled-distortion lenses. Drafomir Anguelov, Carole Dulong, Daniel Filip, Christian Frueh, Stepheane Lafon, Richard Lyon, Abhijit Ogale, Luc Vincent, Josh Weaver. Google Street View: Capturing The World At Street Level, IEEE Computer, June 2010.

Google Street View Car Fleet October 15, 2012

Will We Have Autonomous Driving Vehicles? Every decade another bit of automation is introduced: 1950s Power steering 1970s Cruise control 1980s Anti-lock brakes 1990s Electronic stability control 2000s The first self-parking cars Now Detection of lane lines Distance from car ahead Night vision Blind spot detection Stereo cameras to identify pedestrians

Autonomous Driving Vehicles Pre-2000 There was no way, before 2000, to make something interesting The sensors weren t there The computers weren t there The mapping wasn t there Radar was a device on a hilltop that cost $200M Sebastian Thrun Founder of the Google Car Project

Digital Geometry Capture Photographic methods Laser scanning Sonar Time of Flight All of the Above

Google s Autonomous Driving Vehicle 2010 http://erictric.com/2010/10/09/google-secretly-develops-tests-automated-cars/

Google s Autonomous Driving Vehicle 2013 Uses multiple sensors, each with a different view of the world Laser 64 beams @ 10 revolutions/second scanning 1.3 million points in concentric waves starting 8 feet from the car It can spot a 14 object at a distance of 160 feet Radar Has twice the range of the Laser, but much less precision Photography Excellent at identifying road signs, turn signals, colors and lights

Google s Autonomous Driving Vehicle 2014-2015 New laser sensors 2 X range 30 X resolution @ 300 can spot a metal plate <2 thick Size of a coffee mug Cost $10,000 (less than current model @ $80,000)

Data from Sensors on Google s ADV

Google Street View The world contains roughly 50 million miles of roads, paved and unpaved, across 219 countries (ref.) This is equivalent to circumnavigating the globe 1250 times. To date, hundreds of cities in many countries across four continents have been captured. Google has developed several vehicular platforms and texture information in the project s seven year history. Drafomir Anguelov, Carole Dulong, Daniel Filip, Christian Frueh, Stepheane Lafon, Richard Lyon, Abhijit Ogale, Luc Vincent, Josh Weaver. Google Street View: Capturing The World At Street Level, IEEE Computer, June 2010.

Google Street View 2012 2011 Google announced it would start charging (large) commercial sites 2012 Google allows users to post photos and reviews of locations By October 2012, Google will have updated 250,000 miles of U.S. roads Note: They have also added Google Moon and Google Mars

Lombard Street, San Francisco

Google Camel View

Google Street View Acquisition Map 2012

Google Street View Acquisition Map 2016

Using Street View data to enhance user walk-through experiences in Google Earth. Original 3D models of a New York City scene from airborne data only. Fused 3D model with high-resolution facades. The visual quality is considerably higher, and many storefronts and signs can now be easily identified and recognized. Drafomir Anguelov, Carole Dulong, Daniel Filip, Christian Frueh, Stepheane Lafon, Richard Lyon, Abhijit Ogale, Luc Vincent, Josh Weaver. Google Street View: Capturing The World At Street Level, IEEE Computer, June 2010.

Imagery from new Street View Vehicle is accompanied by laser range data - which is aggregated and simplified by robustly fitting it in a coarse mesh that models the dominant scene surfaces. Drafomir Anguelov, Carole Dulong, Daniel Filip, Christian Frueh, Stepheane Lafon, Richard Lyon, Abhijit Ogale, Luc Vincent, Josh Weaver. Google Street View: Capturing The World At Street Level, IEEE Computer, June 2010.

Uber 2016

Velodyne s Lidar Sensors (VLP-16) 2016 $8,000 (single mode) Range: 100 meters Depth Accuracy ±3cm Field of View Horizontal 360 Vertical ±15 Data Rate Rotates 20x/sec. (64 laser/detector pairs) Angular Resolution Horizontal 0.1-0.4 Vertical 2 Data Rate.3M (single mode) 1.3M (quad mode) points/second

Time of Flight Point Cloud

Meshes from Point Clouds

Autodesk ReCap 360

The Game of Drones Science Magazine, 30.Jan.2015 The End of Privacy

End