VIDEO FOR VIRTUAL REALITY LIGHT FIELD BASICS JAMES TOMPKIN

Similar documents
Image-Based Rendering

But, vision technology falls short. and so does graphics. Image Based Rendering. Ray. Constant radiance. time is fixed. 3D position 2D direction

Computational Photography

VOLUMETRIC VIDEO // PLENOPTIC LIGHTFIELD // MULTI CAMERA METHODOLOGIES JORDAN HALSEY // VR PLAYHOUSE

More and More on Light Fields. Last Lecture

Modeling Light. Slides from Alexei A. Efros and others

Modeling Light. Michal Havlik : Computational Photography Alexei Efros, CMU, Fall 2007

Real Time Rendering. CS 563 Advanced Topics in Computer Graphics. Songxiang Gu Jan, 31, 2005

Algorithms for Image-Based Rendering with an Application to Driving Simulation

Capturing and View-Dependent Rendering of Billboard Models

Image Base Rendering: An Introduction

Hybrid Rendering for Collaborative, Immersive Virtual Environments

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

The Light Field and Image-Based Rendering

Morphable 3D-Mosaics: a Hybrid Framework for Photorealistic Walkthroughs of Large Natural Environments

Image-based modeling (IBM) and image-based rendering (IBR)

IMMERSIVE VIRTUAL REALITY FOR LIVE-ACTION VIDEO USING CAMERA ARRAYS

Modeling Light. Michal Havlik

Focal stacks and lightfields

Image-Based Modeling and Rendering

Modeling Light. Michal Havlik : Computational Photography Alexei Efros, CMU, Fall 2011

Computational Cameras: Exploiting Spatial- Angular Temporal Tradeoffs in Photography

Modeling Light. Michal Havlik

GPU assisted light field capture and processing

TREE-STRUCTURED ALGORITHM FOR EFFICIENT SHEARLET-DOMAIN LIGHT FIELD RECONSTRUCTION. Suren Vagharshakyan, Robert Bregovic, Atanas Gotchev

Image-Based Rendering and Light Fields

Image-Based Modeling and Rendering. Image-Based Modeling and Rendering. Final projects IBMR. What we have learnt so far. What IBMR is about

CS 684 Fall 2005 Image-based Modeling and Rendering. Ruigang Yang

CSCI 1290: Comp Photo

A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India

Computational Photography

Image-Based Modeling and Rendering

A Warping-based Refinement of Lumigraphs

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting

Light Field Spring

Rendering by Manifold Hopping

Modeling, Combining, and Rendering Dynamic Real-World Events From Image Sequences

The Plenoptic videos: Capturing, Rendering and Compression. Chan, SC; Ng, KT; Gan, ZF; Chan, KL; Shum, HY.

Image or Object? Is this real?

Lecture 15: Image-Based Rendering and the Light Field. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011)

Linearizing the Plenoptic Space

Plenoptic camera and its Applications

CSc Topics in Computer Graphics 3D Photography

Structure from Motion and Multi- view Geometry. Last lecture

Depth Images: Representations and Real-time Rendering

Sashi Kumar Penta COMP Final Project Report Department of Computer Science, UNC at Chapel Hill 13 Dec, 2006

Traditional Image Generation. Reflectance Fields. The Light Field. The Light Field. The Light Field. The Light Field

A million pixels, a million polygons. Which is heavier? François X. Sillion. imagis* Grenoble, France

Image-Based Rendering. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen

View Synthesis for Multiview Video Compression

Active Scene Capturing for Image-Based Rendering with a Light Field Setup

3D Editing System for Captured Real Scenes

Visual Imaging in the Electronic Age Assignment #3 Draft Geometry Capture

Jingyi Yu CISC 849. Department of Computer and Information Science

Light Fields. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen

IMAGE-BASED RENDERING TECHNIQUES FOR APPLICATION IN VIRTUAL ENVIRONMENTS

Image-based rendering using plane-sweeping modelisation

IMAGE-BASED RENDERING

Light Field Techniques for Reflections and Refractions

Free-viewpoint video renderer

Mosaics, Plenoptic Function, and Light Field Rendering. Last Lecture

Re-live the Movie Matrix : From Harry Nyquist to Image-Based Rendering. Tsuhan Chen Carnegie Mellon University Pittsburgh, USA

Plenoptic Image Editing

Modeling Light. On Simulating the Visual Experience

A FREE-VIEWPOINT VIDEO SYSTEM FOR VISUALISATION OF SPORT SCENES

3-D Shape Reconstruction from Light Fields Using Voxel Back-Projection

Multi-view stereo. Many slides adapted from S. Seitz

High-Quality Interactive Lumigraph Rendering Through Warping

VIDEO-TO-3D. Marc Pollefeys, Luc Van Gool, Maarten Vergauwen, Kurt Cornelis, Frank Verbiest, Jan Tops

Image-Based Rendering using Image-Warping Motivation and Background

PAPER Three-Dimensional Scene Walkthrough System Using Multiple Acentric Panorama View (APV) Technique

Evolution of Imaging Technology in Computer Graphics. Related Areas

A million pixels, a million polygons: which is heavier?

CONVERSION OF FREE-VIEWPOINT 3D MULTI-VIEW VIDEO FOR STEREOSCOPIC DISPLAYS

PART IV: RS & the Kinect

CS 283: Assignment 3 Real-Time / Image-Based Rendering

3D Modeling of Objects Using Laser Scanning

Other approaches to obtaining 3D structure

DEPTH, STEREO AND FOCUS WITH LIGHT-FIELD CAMERAS

Ping Tan. Simon Fraser University

Tecnologie per la ricostruzione di modelli 3D da immagini. Marco Callieri ISTI-CNR, Pisa, Italy


Optimizing an Inverse Warper

Tecnologie per la ricostruzione di modelli 3D da immagini. Marco Callieri ISTI-CNR, Pisa, Italy

Volumetric Scene Reconstruction from Multiple Views

Point based Rendering

Plenoptic Image Editing

Computational Imaging for Self-Driving Vehicles

FAST ALGORITHM FOR CREATING IMAGE-BASED STEREO IMAGES

Geometric Modeling. Bing-Yu Chen National Taiwan University The University of Tokyo

Semantic 3D Reconstruction of Heads Supplementary Material

A Generalized Light-Field API and Management System

View Synthesis for Multiview Video Compression

Acquiring 4D Light Fields of Self-Luminous Light Sources Using Programmable Filter

LASER ECHO INTENSITY BASED TEXTURE MAPPING OF 3D SCAN MESH

A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect

AS the most important medium for people to perceive

3D Reconstruction with Tango. Ivan Dryanovski, Google Inc.

New Frontier in Visual Communication and Networking: Multi-View Imaging. Professor Tsuhan Chen 陳祖翰

Colored Point Cloud Registration Revisited Supplementary Material

Transcription:

VIDEO FOR VIRTUAL REALITY LIGHT FIELD BASICS JAMES TOMPKIN

WHAT IS A LIGHT FIELD? Light field seems to have turned into a catch-all term for many advanced camera/display technologies.

WHAT IS A LIGHT FIELD? Has become a trade mark: - Avegant Light Field Technology - Light Field Lab - LightField Studios [Avegant / TechCrunch.com]

WHAT IS A LIGHT FIELD? Volumetric video But usually not a regular sampling like volumetric MRI data 4D video But definitely not 4D cinema 6DoF video Multi-camera / camera array Free viewpoint video Video-based rendering Lumigraph (+unstructured) I m in the movie!

WHAT IS A LIGHT FIELD? Light field seems to have turned into a catch-all term for many advanced camera/display technologies. How should we think about this?

PLENOPTIC FUNCTION 5+ dimensional function representing all light everywhere x,y,z -> position in space θ,ɸ -> angle on sphere L -> amount of light (radiance) [Wikipedia Light Field Qef] [Adelson and Bergen 1991, McMillan & Bishop 1995]

PLENOPTIC FUNCTION 5+ dimensional function representing all light everywhere x,y,z -> position in space θ,ɸ -> angle on sphere L -> amount of light (radiance) Video needs t -> time [Wikipedia Light Field Qef] [Adelson and Bergen 1991, McMillan & Bishop 1995]

WHAT IS A LIGHT FIELD? Light field seems to have turned into a catch-all term for many advanced camera/display technologies. How should we think about this? Different camera configurations provide different samplings of the plenoptic function. [Chai et al., 2000, Lin et al. 2004]

ALL TECHNIQUES ARE SAMPLINGS Idealized 360 video Single x,y,z sample Complete sampling of θ,ɸ θ,ɸ x,y,z

ALL TECHNIQUES ARE SAMPLINGS Idealized stereo 3D 360 video Two x,y,z samples Complete sampling of θ,ɸ What happens when you tilt your head? θ,ɸ θ,ɸ x 1,y 1,z 1 x 2,y 2,z 2

ALL TECHNIQUES ARE SAMPLINGS A light field generally implies Many x,y,z samples Some sampling of θ,ɸ Many options for how to sample! θ,ɸ y z x

THE RIGHT WAY TO THINK ABOUT LIGHT FIELDS A space of possible samplings! Leads to a space of possible camera setups. Each is some trade-off in the sampling space. Different configurations for different applications. Each fails to sample some rays but we have a suite of algorithms to (try to) accommodate.

[Image Lytro] EXAMPLE: LYTRO CINEMA y Lenslet array, e.g., 60 fov Lenslet-based light field imager 755 MPixel sensor x Sensor [Levoy and Hanrahan 1996, Gortler et al. 1996, Ng et al. 2005]

[Image Lytro] EXAMPLE: LYTRO CINEMA y Lenslet array, e.g., 60 fov Lenslet-based light field imager 755 MPixel sensor [Levoy and Hanrahan 1996, Gortler et al. 1996, Ng et al. 2005] x Sensor Small baseline dense sampling Little parallax/head motion Small θ,ɸ extent Only sees part of scene

[Image Lytro] EXAMPLE: LYTRO CINEMA y Lenslet array, e.g., 60 fov Lenslet-based light field imager 755 MPixel sensor [Levoy and Hanrahan 1996, Gortler et al. 1996, Ng et al. 2005] x Sensor - Still a light field camera.but not intended for VR. - Helps post-production for synthetic shutter and aperture, scene segmentation.

[Image Lytro] EXAMPLE: LYTRO IMMERGE y x Multi-cam array, e.g., 120 fov Multi-camera light field imager 95 cameras

[Image Lytro] EXAMPLE: LYTRO IMMERGE y x Multi-cam array, e.g., 120 fov Multi-camera light field imager 95 cameras Larger baseline Sufficient parallax/head motion for seated(+) VR Still not 360 θ,ɸ extent Wider FOV lenses

[Image Lytro] EXAMPLE: LYTRO IMMERGE y x Multi-cam array, e.g., 120 fov Multi-camera light field imager 95 cameras Better for VR Larger headbox But need to interpolate farther between views.

VIEW INTERPOLATION What happens when my VR head goes in between the cameras? Must resample captured rays to render a novel view. Dense camera sampling, e.g., Lytro Cinema World Interpolated frame Capture camera Render camera Capture camera Looks ok if sampling is dense / object is far away. [Chai et al., 2000, Lin et al. 2004]

VIEW INTERPOLATION What happens when my VR head goes in between the cameras? Must resample captured rays to render a novel view. World Interpolated frame Capture camera Render camera Capture camera More and more ghosting. [Chai et al., 2000, Lin et al. 2004]

VIEW INTERPOLATION What happens when my VR head goes in between the cameras? Must resample captured rays to render a novel view. World Interpolated frame Capture camera Render camera Capture camera More and more ghosting. [Chai et al., 2000, Lin et al. 2004]

VIEW INTERPOLATION What happens when my VR head goes in between the cameras? Must resample captured rays to render a novel view. World Interpolated frame Capture camera Render camera Capture camera More and more ghosting. [Chai et al., 2000, Lin et al. 2004]

VIEW INTERPOLATION What happens when my VR head goes in between the cameras? Must resample captured rays to render a novel view. Sparser camera sampling, e.g., Lytro Immerge World Interpolation is untenable. Capture camera Render camera Capture camera Must do something else. [Chai et al., 2000, Lin et al. 2004]

VIEW INTERPOLATION What happens when my VR head goes beyond my baseline? Reveal space not captured by any cameras. Can fake it or hallucinate inpainting only goes so far. Basically impossible with sampling-based approach. VR head position

PYRAMID OF ANALYSIS / SYNTHESIS Sparser samplings More world reconstruction [Magnor et al., Video-based Rendering, SIGGRAPH 2005 Course]

COMMON REPRESENTATIONS Point Clouds Dynamic Mesh + View-dependent Texture [Point-based Graphics, Gross and Pfister 2007] e.g., Nurulize Atom View [Collet et al., High-quality Streamable Free-viewpoint Video, ACM ToG 2015]

DEPTH MAP EXTRACTION Point clouds come from computing world depth. Multi-view stereo. Many techniques: Yucer et al., SIGGRAPH 2016 Wang et al., TPAMI 2016 Jeon et al., CVPR 2015 Kim et al., SIGGRAPH 2013, 3DV 2016 LF input Wang et al.

PHOTOGRAMMETRY / GEOMETRY RECONSTRUCTION Merge depth maps / fit surface to point cloud. One reconstruction per frame. Errors: Geometry edge flickering, phantom volumes Tools: RealityCapture, ImageModeler, PhotoScan. [ Mathworks] [Kazhdan et al., Poisson Surface Reconstruction, SGP 2006]

TIME-EVOLVING MESHES Prada et al., Spatiotemporal atlas parameterization for evolving meshes. SIGGRAPH 2017

OTHER TECHNOLOGIES: LIDAR, DEPTH CAMERAS, OTHER SCANNERS Even with dense sampling, sometimes we can t reconstruct depth or geometry: Complex materials Dynamic events Use other modes to fill in: Geometry from LIDAR to generate clean backplate for dynamic depth. Fit known geometric models to scene. [FARO Focus laser scanner]

[Point-based Graphics, Gross and Pfister 2007] DO WE EVEN NEED SURFACES? [ Nurulise, 2017]

RECORDING LIGHT FIELDS A lot of data! 100x the cameras = 100x the data VR needs high frame rate 90+ fps 100 camera 1080p 8bit RGB 90fps = 52 Gbytes per second RAW Terabytes of data for even a movie short.

PLAYING LIGHT FIELDS Real-time rendering in game engine 90 Hz High data bandwidth, complex geometry. Big machine. Alternative: holographic video [OTOY holographic videos - https://home.otoy.com/render/light-fields/]

LIGHT FIELD VIDEO GOALS High immersion 6 degrees of movement freedom. Correct stereo vision everywhere, not just horizontal. High realism Looks closer to real life than a real-time rendering But artifacts often not solid, e.g., geometric edge flickering. Starts to lose the look of video.

LYTRO MOON [Images Lytro - http://blog.lytro.com/vr-announcing-moon/]

SONY JOSHUA BELL [Images Sony - https://youtu.be/poi6e8o4jmq]

SONY JOSHUA BELL [Images Sony - https://youtu.be/poi6e8o4jmq]

SONY JOSHUA BELL [Images Sony - https://youtu.be/poi6e8o4jmq]

VR VILLAGE @ SIGGRAPH 2017 Hallelujah: Creating a Breakthrough VR Experience with Lytro Immerge [Lytro/Within, Richter, Halvorson, Kolar, 2017]

[Lytro/Within, Richter, Halvorson, Kolar, 2017]

[Lytro/Within, Richter, Halvorson, Kolar, 2017]

[Lytro/Within, Richter, Halvorson, Kolar, 2017]

EDITING LIGHT FIELDS Many operations we take for granted on 1-camera video are not well defined on light fields! Any editing operation must be efficient. Example: filtering

EXAMPLE UNPROCESSED LIGHT FIELD VIDEO...... 90 1 2 3 4 5 6 42 43 44 45 46 47 86 87 88 89 91...... [Big Buck Bunny / Kovács et al. / Holografika]

LIGHT FIELD VIDEO CYCLE THROUGH CAMERA VIEWS [Bonneel, Tompkin, Wang et al., Eurographics 2017, Consistent Video Filtering for Camera Arrays] [Big Buck Bunny / Kovács et al. / Holografika]

LIGHT FIELD VIDEO FILTERING INTRINSIC DECOMPOSITION (SHADING) [Bonneel, Tompkin, Wang et al., Eurographics 2017, Consistent Video Filtering for Camera Arrays] Inconsistency across space and across time! [Big Buck Bunny / Kovács et al. / Holografika]

LIGHT FIELD VIDEO FILTERING INTRINSIC DECOMPOSITION (SHADING) Input Processed [Bonneel, Tompkin, Wang et al., Eurographics 2017, Consistent Video Filtering for Camera Arrays] [Big Buck Bunny / Kovács et al. / Holografika]

LIGHT FIELD VIDEO FILTERING SPATIO-TEMPORAL CONSISTENCY Our result [Bonneel, Tompkin, Wang et al., Eurographics 2017, Consistent Video Filtering for Camera Arrays] [Big Buck Bunny / Kovács et al. / Holografika]

6DoF Production Jordan Halsey