Non-line-of-sight imaging

Similar documents
Non-line-of-sight imaging

Last Lecture. Bayer pattern. Focal Length F-stop Depth of Field Color Capture. Prism. Your eye. Mirror. (flipped for exposure) Film/sensor.

Computational Imaging for Self-Driving Vehicles

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Multiple View Geometry

Signature redacted. Signature redacted. Capturing the Human Figure Through a Wall LIBRARIES ARCHEIV. Chen-Yu Hsu MAR

Computational light transport

Image correspondences and structure from motion

Face Cyclographs for Recognition

3D Computer Vision 1

More computational light transport

Focal stacks and lightfields

3D Scanning. Qixing Huang Feb. 9 th Slide Credit: Yasutaka Furukawa

3D Modeling of Objects Using Laser Scanning

Range Sensors (time of flight) (1)

Overview of Active Vision Techniques

Multi-View Stereo for Community Photo Collections Michael Goesele, et al, ICCV Venus de Milo

3D Photography: Active Ranging, Structured Light, ICP

Computational Imaging for Self-Driving Vehicles

Computer Vision Lecture 17

HISTOGRAMS OF ORIENTATIO N GRADIENTS

CIS 580, Machine Perception, Spring 2015 Homework 1 Due: :59AM

Depth Camera for Mobile Devices

CS 4495/7495 Computer Vision Frank Dellaert, Fall 07. Dense Stereo Some Slides by Forsyth & Ponce, Jim Rehg, Sing Bing Kang

Project 2 due today Project 3 out today. Readings Szeliski, Chapter 10 (through 10.5)

Multi-view stereo. Many slides adapted from S. Seitz

Multi-view Stereo. Ivo Boyadzhiev CS7670: September 13, 2011

Computer Vision Lecture 17

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1.

Gaze interaction (2): models and technologies

CPSC 425: Computer Vision

3D Photography: Stereo

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

12/3/2009. What is Computer Vision? Applications. Application: Assisted driving Pedestrian and car detection. Application: Improving online search

3D object recognition used by team robotto

New Sony DepthSense TM ToF Technology

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Stereo and structured light

Local features: detection and description. Local invariant features

Light transport matrices

CS4670: Computer Vision

Active Light Time-of-Flight Imaging

Local features: detection and description May 12 th, 2015

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera

Other Reconstruction Techniques

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Light Transport CS434. Daniel G. Aliaga Department of Computer Science Purdue University

Capturing light. Source: A. Efros

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Recap from Previous Lecture

Image-based Lighting

IRIS SEGMENTATION OF NON-IDEAL IMAGES

Stereo vision. Many slides adapted from Steve Seitz

Epipolar geometry contd.

Perceptual Effects in Real-time Tone Mapping

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

CS635 Spring Department of Computer Science Purdue University

All human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them.

Capturing, Modeling, Rendering 3D Structures

ToF Camera for high resolution 3D images with affordable pricing

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision

What have we leaned so far?

/$ IEEE

A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India

CMPSCI 670: Computer Vision! Image formation. University of Massachusetts, Amherst September 8, 2014 Instructor: Subhransu Maji

Light source estimation using feature points from specular highlights and cast shadows

Ray Tracing. CSCI 420 Computer Graphics Lecture 15. Ray Casting Shadow Rays Reflection and Transmission [Ch ]

CS5670: Computer Vision

Dense 3D Reconstruction. Christiano Gava

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

CSE 167: Lecture #7: Color and Shading. Jürgen P. Schulze, Ph.D. University of California, San Diego Fall Quarter 2011

Structured light , , Computational Photography Fall 2017, Lecture 27

Measurements using three-dimensional product imaging

CS201 Computer Vision Lect 4 - Image Formation

Light Field Occlusion Removal

Dense 3D Reconstruction. Christiano Gava

Structured light 3D reconstruction

Introduction to 3D Machine Vision

BIL Computer Vision Apr 16, 2014

COMP371 COMPUTER GRAPHICS

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Phys 102 Lecture 17 Introduction to ray optics

Geometric camera models and calibration

3D scanning. 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects.

Laser Eye a new 3D sensor for active vision

Advanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation

Experimental Humanities II. Eye-Tracking Methodology

CS354 Computer Graphics Ray Tracing. Qixing Huang Januray 24th 2017

Basic distinctions. Definitions. Epstein (1965) familiar size experiment. Distance, depth, and 3D shape cues. Distance, depth, and 3D shape cues

Structure from Motion and Multi- view Geometry. Last lecture

Light, Photons, and MRI

CS 563 Advanced Topics in Computer Graphics Camera Models. by Kevin Kardian

Visual Perception Sensors

Active Wearable Vision Sensor: Recognition of Human Activities and Environments

CSE 4392/5369. Dr. Gian Luca Mariottini, Ph.D.

(A) Electromagnetic. B) Mechanical. (C) Longitudinal. (D) None of these.

New Sony DepthSense TM ToF Technology

Transcription:

Non-line-of-sight imaging http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 25

Course announcements Homework 6 will be posted tonight. - Will be due Sunday 10th. - Almost no coding, only data capture and using existing code. - You will need high-sensitivity cameras, so use the DSLRs you have picked up. Final project report deadline moved to December 15 th. - Originally was December 11 th.

Overview of today s lecture The non-line-of-sight (NLOS) imaging problem. Active NLOS imaging using time-of-flight imaging. Active NLOS imaging using WiFi. Passive NLOS imaging using accidental pinholes. Passive NLOS imaging using accidental reflectors. Passive NLOS imaging using corners.

Slide credits Many of these slides were directly adapted from: Shree Nayar (Columbia). Fadel Adib (MIT). Katie Bouman (MIT).

The non-line-of-sight (NLOS) imaging problem

Time-of-flight (ToF) imaging

Looking around the corner

Looking around the corner

Active NLOS imaging using time-of-flight imaging

Processing a single-pixel transient VD VS light source single-pixel transient camera intensity time

Processing a single-pixel transient VD VS light source single-pixel transient camera intensity t 1 = 10 ps time

Processing a single-pixel transient VD VS l in l out light source single-pixel transient camera intensity l nlos = t 1 c - l in - l out Where in the world can we find points creating this pathlength? t 1 = 10 ps time

Processing a single-pixel transient VD VS l in l out light source Ellipse with foci on the virtual source and detector and pathlength l nlos single-pixel transient camera intensity l nlos = t 1 c - l in - l out t 1 = 10 ps time

Processing a single-pixel transient VD VS l in l out light source single-pixel transient camera Ellipse with foci on the virtual source and detector and pathlength l nlos intensity l nlos = t 2 c - l in - l out t 2 = 20 ps time

Processing another single-pixel transient VD VS l in l out light source single-pixel transient camera intensity What assumption are we making here? Is there any other information we are not using? t 2 = 20 ps time

Processing another single-pixel transient VD VS l in Photons only interact with the NLOS scene once (3-bounce paths) l out light source single-pixel transient camera intensity What assumption are we making here? Is there any other information we are not using? t 2 = 20 ps time

Processing another single-pixel transient VD VS l in Photons only interact with the NLOS scene once (3-bounce paths) l out light source single-pixel transient camera What assumption are we making here? intensity Is there any other information we are not using? t 2 = 20 ps We need to also account for intensity time

Elliptic backprojection VD VS l in l out light source single-pixel transient camera

Elliptic backprojection

Elliptic backprojection

Reconstructing Hidden Rooms Visible wall Invisible walls

Reconstructing Hidden Rooms: One plane at a time Visible wall Invisible walls

Reconstructing Hidden Rooms: One plane at a time Visible wall Invisible walls

Reconstructing Hidden Rooms: One plane at a time Visible wall Invisible walls

Reconstructing Hidden Rooms: One plane at a time Visible wall Invisible walls

Reconstructing Hidden Rooms: One plane at a time Visible wall Invisible walls

Reconstructing Hidden Rooms: One plane at a time Visible wall Invisible walls

Reconstructing Rectangular Rooms Original room Reconstructed room Average AICP error for all the walls is TBD mm (TBD %). Normalized with average room length of 1.1m

Reconstructing Complex Shape and Reflectance wall hidden object camera what a regular camera sees what shape our camera sees wall hidden object camera what a regular camera sees what depth our camera sees

Active NLOS imaging using WiFi

Imaging through occlusions

Imaging through occlusions using radio frequencies

Key Idea

Challenges Wall refection is 10,000x stronger than reflections coming from behind the wall Tracking people from their reflections

Wi-Vi: Small, Low-Power, Wi-Fi Eliminate the wall s reflection Track people from reflections Gesture-based interface Implemented on software radios

How Can We Eliminate the Wall s Reflection?

Idea: transmit two waves that cancel each other when they reflect off static objects but not moving objects Wall is static disappears People tend to move detectable

Eliminating the Wall s Reflection Receive Antenna: x αx Transmit Antennas

Eliminating the Wall s Reflection Receive Antenna: y = h 1 x + h 2 αx 0 x h 1 α = -h 1 / h 2 αx h 2

Eliminating All Static Reflections

Eliminating All Static Reflections x αx h 1 h 2 y = h 1 x + h 2 αx Static objects (wall, furniture, etc.) have constant channels y = h 1 x + h 2 (- h 1 / h 2 )x 0 People move, therefore their channels change y = h 1 x + h 2 (- h 1 / h 2 )x Not Zero

How Can We Track Using Reflections?

RF source Tracking Motion θ Direction of reflection Antenna Array

Tracking Motion Direction of motion θ At any point in time, we have a single measurement Antenna Array

Tracking Motion Direction of motion θ θ Direction of motion Antenna Array

Tracking Motion Direction of motion θ Antenna Array Direction of motion Human motion emulates antenna array θ

A Through-Wall Gesture Interface Sending Commands with Gestures Two simple gestures to represent bit 0 and bit 1 Can combine sequence of gestures to convey longer message

Gesture Encoding Bit 0 : step forward followed by step backward Bit 1 : step backward followed by step forward Step Forward Step Backward

Gesture Decoding Matched Filter Peak Detector Bit 0 Bit 1 Gesture interface that works through walls and none-line-of-sight

Imaging through occlusions using radio frequencies Our output Camera image Head Left Arm Chest Right Arm Our RF sensor Lower Left Arm Lower Right Arm Feet

Traditional Imaging RF Imaging Cannot image through occlusions like walls Walls are transparent and can image through them Form 2D images using lenses No lenses at these frequencies

Imaging with RF No lens at these frequencies Antenna cannot distinguish bounces from different directions Antenna

Imaging with RF Beamforming: Use multiple antennas to scan reflections within a specific beam Antenna Array Extend to 3D with time-of-flight measurements by repeating this at every depth

Coarse-to-fine Scan Larger aperture (more antennas) means finer resolution Used antennas

Coarse-to-fine Scan Reflectors Used antennas

Coarse-to-fine Scan Scanned region Used antennas

Coarse-to-fine Scan Scanned regions Used antennas

Traditional Imaging RF Imaging Cannot image through occlusions like walls Walls are transparent and can image through them Form 2D images using lenses No lenses at these frequencies Get a reflection from all points: can image all the body No reflections from most points: all reflections are specular

Challenge: Don t get reflections from most points in RF Output of 3D RF Scan Blobs of reflection power

Challenge: Don t get reflections from most points in RF At frequencies that traverse walls, human body parts are specular (pure mirror)

Antennas Challenge: Don t get reflections from most points in RF At frequencies that traverse walls, human body parts are specular (pure mirror) Cannot Capture Reflection At every point in time, get reflections from only a subset of body parts

Solution Idea: Exploit Human Motion and Aggregate over Time RF Imaging Array

Solution Idea: Exploit Human Motion and Aggregate over Time RF Imaging Array

Human Walks toward Sensor 3m 2.5m 2m Chest (Largest Convex Reflector) Use it as a pivot: for motion compensation and segmentation

Human Walks toward Sensor 3m 2.5m 2m Right arm Head Left arm Feet Lower Torso Chest (Largest Convex Reflector) Use it as a pivot: for motion compensation and segmentation Combine the various snapshots

Human Walks toward Sensor

Implementation Hardware 2D Antenna Array Built RF circuit 1/1,000 power of WiFi USB connection to PC Rx Tx Software Coarse-to-fine algorithm implemented in GPU to generate reflection snapshots in real-time

Evaluation SENSOR SETUP Sensor is Placed Behind a Wall Snapshots are Sensor RF-Capture sensor placed behind the wall 15 participants Use Kinect as baseline when needed

Sample Captured Figures through Walls

y-axis (meters) y-axis (meters) y-axis (meters) y-axis (meters) Sample Captured Figures through Walls 2 2 2 2 1.5 1.5 1.5 1.5 1 1 1 1 0.5 0.5 0.5 0.5 0-0.2 0 0.2 0.4 0.6 0.8 x-axis (meters) 0-0.2 0 0.2 0.4 0.6 0.8 x-axis (meters) 0-0.2 0 0.2 0.4 0.6 0.8 x-axis (meters) 0-0.2 0 0.2 0.4 0.6 0.8 x-axis (meters)

Tracking result

Writing in the air Kinect (in red) Median Accuracy is 2cm

Passive NLOS imaging using accidental pinholes

What does this image say about the world outside?

Accidental pinhole camera

Accidental pinhole camera projected pattern on the wall window with smaller gap upside down view outside window window is an aperture

Accidental pinspeck camera

Passive NLOS imaging using accidental reflectors

Ophthalmology Anatomy, physiology [Helmholtz 09; ] Computer Vision Iris recognition [Daugman 93; ] HCI Gaze as an interface [Bolt 82; ] Computer Graphics Vision-realistic Rendering [Barsky et al. 02; ]

Corneal Imaging System

Geometric Model of the Cornea Iris Pupil Sclera Cornea R=7.6mm t b 2.18mm r L eccentricity = 0.5 5.5mm

Self-calibration: 3D Coordinates, 3D Orientation

Viewpoint Loci camera pupil X cornea Z

Viewpoint Loci camera pupil X cornea Z

Viewpoint Loci camera pupil cornea

Resolution and Field of View gaze direction X cornea Z

Resolution and Field of View resolution gaze direction X FOV cornea Z

Resolution and Field of View resolution gaze direction 27 person s FOV 45 X cornea Z FOV

Environment Map from an Eye

What Exactly You are Looking At Eye Image: Computed Retinal Image:

Corneal Stereo System right cornea left cornea epipolar plane camera pupil epipolar curve

From Two Eyes in an Image Y X Z X Epipolar Curves Reconstructed Structure (frontal and side view) Y

Eyes Reveal Where the person is What the person is looking at The structure of objects

Implications Human Affect Studies: Social Networks Security: Human Localization Advanced Interfaces: Robots, Computers Computer Graphics: Relighting [SIGGRAPH 04]

Dynamic Illumination in a Video

Point Source Direction from the Eye intensity

Point Source Trajectory from the Eye (deg.) (deg.) (deg.) (deg.)

Inserting a Virtual Object

Sampling Appearance using Eyes

Sampling Appearance using Eyes

Computed Point Source Trajectory intensity

Fitting a Reflectance Model

3D Model Reconstruction

Relighting under Novel Illumination

VisualEyes TM http://www.cs.columbia.edu/cave/ with Akira Yanagawa

Passive NLOS imaging using corners

References Basic reading: Kirmani et al., Looking around the corner using ultrafast transient imaging, ICCV 2009 and IJCV 2011. Velten et al., Recovering three-dimensional shape around a corner using ultrafast time-offlight imaging, Nature Communications 2012. the two papers showing how ToF imaging can be used for looking around the corner. Abib and Katabi, See Through Walls with Wi-Fi!, SIGCOMM 2013. Abib et al., Capturing the Human Figure Through a Wall, SIGGRAPH Asia 2015. the two papers showing that WiFi can be used to see through walls. Torralba and Freeman, Accidental Pinhole and Pinspeck Cameras, CVPR 2012. the paper discussing passive NLOS imaging using accidental pinholes. Nishimo and Nayar, Corneal Imaging System: Environment from Eyes, IJCV 2006. the paper discussing passive NLOS imaging using accidental reflectors. Bouman et al., Turning corners into cameras: Principles and Methods, ICCV 2017. the paper discussing passive NLOS imaging using corners. Additional reading: Pediredla et al., Reconstructing rooms using photon echoes: A plane based model and reconstruction algorithm for looking around the corner, ICCP 2017. the paper on NLOS room reconstruction using ToF imaging. Nishimo and Nayar, Eyes for relighting, SIGGRAPH 2004. a follow-up paper to the paper on corneal imaging, show how similar ideas can be used for relighting and other image-based rendering tasks.