Re-live the Movie Matrix : From Harry Nyquist to Image-Based Rendering. Tsuhan Chen Carnegie Mellon University Pittsburgh, USA

Similar documents
New Frontier in Visual Communication and Networking: Multi-View Imaging. Professor Tsuhan Chen 陳祖翰

A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India

But, vision technology falls short. and so does graphics. Image Based Rendering. Ray. Constant radiance. time is fixed. 3D position 2D direction

Image-Based Rendering

Modeling Light. Michal Havlik : Computational Photography Alexei Efros, CMU, Fall 2007

Image or Object? Is this real?

A Survey on Image-Based Rendering

Modeling Light. Michal Havlik : Computational Photography Alexei Efros, CMU, Fall 2011

Modeling Light. Michal Havlik

Multiple View Geometry

Light Field Spring

Modeling Light. On Simulating the Visual Experience

Image-Based Modeling and Rendering

Hybrid Rendering for Collaborative, Immersive Virtual Environments

Modeling Light. Michal Havlik

Image-Based Rendering and Light Fields

Image-Based Modeling and Rendering. Image-Based Modeling and Rendering. Final projects IBMR. What we have learnt so far. What IBMR is about

Optimization of the number of rays in interpolation for light field based free viewpoint systems

Multi-View Imaging: Capturing and Rendering Interactive Environments

Active Scene Capturing for Image-Based Rendering with a Light Field Setup

Modeling Light. Slides from Alexei A. Efros and others

EFFICIENT REPRESENTATION OF LIGHTING PATTERNS FOR IMAGE-BASED RELIGHTING

Image-based modeling (IBM) and image-based rendering (IBR)

Active Rearranged Capturing of Image-Based Rendering Scenes Theory and Practice

Capturing and View-Dependent Rendering of Billboard Models

Mosaics, Plenoptic Function, and Light Field Rendering. Last Lecture

TREE-STRUCTURED ALGORITHM FOR EFFICIENT SHEARLET-DOMAIN LIGHT FIELD RECONSTRUCTION. Suren Vagharshakyan, Robert Bregovic, Atanas Gotchev

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

Image-Based Modeling and Rendering

IMAGE-BASED RENDERING TECHNIQUES FOR APPLICATION IN VIRTUAL ENVIRONMENTS

Image-Based Rendering. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen

Rendering by Manifold Hopping

More and More on Light Fields. Last Lecture

Image Base Rendering: An Introduction

Volumetric Scene Reconstruction from Multiple Views

CSc Topics in Computer Graphics 3D Photography

Image-Based Rendering. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen

An object-based approach to plenoptic videos. Proceedings - Ieee International Symposium On Circuits And Systems, 2005, p.

DECODING COMPLEXITY CONSTRAINED RATE-DISTORTION OPTIMIZATION FOR THE COMPRESSION OF CONCENTRIC MOSAICS

Representing Moving Images with Layers. J. Y. Wang and E. H. Adelson MIT Media Lab

Image-Based Rendering. Image-Based Rendering

Lecture 15: Image-Based Rendering and the Light Field. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011)

1-2 Feature-Based Image Mosaicing

Georgios Tziritas Computer Science Department

A Characterization of Inherent Stereo Ambiguities

A unified approach for motion analysis and view synthesis Λ

Multi-view stereo. Many slides adapted from S. Seitz

Computational Photography

Image-based Modeling and Rendering: 8. Image Transformation and Panorama

BIL Computer Vision Apr 16, 2014

VIDEO FOR VIRTUAL REALITY LIGHT FIELD BASICS JAMES TOMPKIN

Real Time Rendering. CS 563 Advanced Topics in Computer Graphics. Songxiang Gu Jan, 31, 2005

3-D Shape Reconstruction from Light Fields Using Voxel Back-Projection

Textureless Layers CMU-RI-TR Qifa Ke, Simon Baker, and Takeo Kanade

Light Field Appearance Manifolds

Determining Reflectance Parameters and Illumination Distribution from a Sparse Set of Images for View-dependent Image Synthesis

A million pixels, a million polygons. Which is heavier? François X. Sillion. imagis* Grenoble, France

Structure from Motion and Multi- view Geometry. Last lecture

Capturing the plenoptic function in a swipe

Spatial Data Structures

The Plenoptic videos: Capturing, Rendering and Compression. Chan, SC; Ng, KT; Gan, ZF; Chan, KL; Shum, HY.

Rate-distortion Optimized Streaming of Compressed Light Fields with Multiple Representations

Feature Tracking and Optical Flow

Supplementary materials of Multispectral imaging using a single bucket detector

RDTC Optimized Streaming for Remote Browsing in Image-Based Scene Representations

On the Use of Ray-tracing for Viewpoint Interpolation in Panoramic Imagery

Spatial Data Structures

Announcements. Mosaics. How to do it? Image Mosaics

View Generation for Free Viewpoint Video System

Light Fields. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen

Stereo Vision. MAN-522 Computer Vision

PAPER Three-Dimensional Scene Walkthrough System Using Multiple Acentric Panorama View (APV) Technique

Real-Time Video- Based Modeling and Rendering of 3D Scenes

Processing 3D Surface Data

Focal stacks and lightfields

Announcements. Light. Properties of light. Light. Project status reports on Wednesday. Readings. Today. Readings Szeliski, 2.2, 2.3.

Project Updates Short lecture Volumetric Modeling +2 papers

A Panoramic Walkthrough System with Occlusion Culling

WATERMARKING FOR LIGHT FIELD RENDERING 1

Face Re-Lighting from a Single Image under Harsh Lighting Conditions

Image-Based Deformation of Objects in Real Scenes

Mosaics. Today s Readings

Jingyi Yu CISC 849. Department of Computer and Information Science

Midterm Examination CS 534: Computational Photography

The Video Z-buffer: A Concept for Facilitating Monoscopic Image Compression by exploiting the 3-D Stereoscopic Depth map

The Light Field and Image-Based Rendering

EE795: Computer Vision and Intelligent Systems

Processing 3D Surface Data

Compression of Lumigraph with Multiple Reference Frame (MRF) Prediction and Just-in-time Rendering

Re-rendering from a Dense/Sparse Set of Images

Projective Texture Mapping with Full Panorama

Image Mosaicing with Motion Segmentation from Video

CS5670: Computer Vision

Visual motion. Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys

Morphable 3D-Mosaics: a Hybrid Framework for Photorealistic Walkthroughs of Large Natural Environments

Multiview Reconstruction

Dense Lightfield Disparity Estimation using Total Variation Regularization

Efficient Depth-Compensated Interpolation for Full Parallax Displays

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Optimizing an Inverse Warper

Transcription:

Re-live the Movie Matrix : From Harry Nyquist to Image-Based Rendering Tsuhan Chen tsuhan@cmu.edu Carnegie Mellon University Pittsburgh, USA Some History IEEE Multimedia Signal Processing (MMSP) Technical Committee, 1996~ IEEE MMSP Workshops Princeton 1997, Los Angeles 1998, Copenhagen 1999, Cannes 2001, St. Thomas 2002, Siena 2004, IEEE International Conf. on Multimedia and Expo. (ICME) New York 2000, Tokyo 2001, Lausanne 2002, Baltimore 2003, Taipei 2004, Proceedings of IEEE, Special Issue on MMSP, 1998 IEEE Transactions on Multimedia, March 1999~ Special issues: networked multimedia 2001, multimedia database 2002, multimodal interface 2003, streaming media 2004, 1

Convergence Image Processing Multimedia Computer Vision Computer Graphics Convergence of Image, Vision, and Graphics Yah, right... The truth is Between vision and graphics communities Vision is more solid research than graphics But graphics people make more money! Graphics and vision communities say: Image processing is low-level processing Image processing community says: Graphics is only some fancy toys Vision is things that don t work! 2

Some Attempts MPEG-4 promised the convergence Started out as model-based coding Encoding: analysis using vision Decoding: synthesis using graphics Well, almost Settled with 2D shape-based coding Model-based for limited content, e.g., faces AFX (Animated Framework Extension) Next wave is image-based rendering Fundamentals of Image-Based Rendering 3

Model-Based Rendering vs. Image-Based Rendering [Digital Michelangelo Project, Stanford] Light field of Michelangelo's statue of Night Example Demo 4

7D Plenoptic Function f ( V, V, V, θ, ψ, λ, t) x y z [Adelson 91] (θ,ψ) (V x,v y y,v z z ) Representations 7D: f ( Vx, Vy, Vz, θ, ψ, λ, t) [Adelson 91] 5D: Stationary and monochrome [McMillan 95] 4D: Scene inside a bounded region Lumigraph [Gortler 96] Lightfield [Levoy 96] EyeVision [Kanade 01] 3D: Viewpoint on a plane Concentric Mosaics [Shum 99] BulletTime [ Matrix ] 2D: Viewpoint at a single point Panorama [Chen 95], QuickTime VR 5

Lumigraph/Lightfield [Gortler 96] [Levoy 96] p p p light ray z(v,t) s u Outer plane (u 0,v 0 ) (s 0,t 0 ) Inner plane t v z 4D f ft/z(v,t) 0 v v' 0 v t 0 2D Simplification v t Captured Images t v u s 4D Image Array 6

Example: EyeVision 4D (incl. time) [Kanade 01] Before Correction After Correction SuperBowl Concentric Mosaics [Shum 99] 3D 7

3D Examples Lobby (1350x320x240) Kids (1462x352x288) Demo IBR: A Sampling Problem Given a set of discrete samples (complete or incomplete) from the plenoptic function, the goal of image based rendering is to generate a continuous representation of that function [McMillan 95] Q: This is a sampling problems How many samples and where? A: Need Nyquist Sampling Theorem for IBR 8

Sampling Theorem ωs > 2ω Nyquist Rate M Shannon, 1949 (in communication theory) Whittaker, 1964 (in math) 2WT numbers (in Fourier series) to represent a function of duration T and highest frequency W Nyquist, 1928 Gabor, 1946 Recall u,v-s,t Parameterization p p p light ray z(v,t) s Outer plane (s 0,t 0 ) t u (u 0,v 0 ) Inner plane v z f ft/z(v,t) 0 v v' 0 v t 0 v t Need multidimensional spectral analysis 9

Spectral Analysis on t-v Plane Intensity on t-v plane Spectrum Intensity on t-v plane Spectrum Sampling for IBR Ω v d max Ωt fωv = 0 π Optimal rendering depth determined by: d opt Ω fω = 0 t v Ω v π Ω t Ω t B π (a) dmin Ωt fωv = 0 Lambertian surface No occlusion Truncating window analysis [Chai et. al, 2000] π (b) Lowpass Filter 10

Optimal Sampling [Zhang and Chen, 2003] Ω v Ω v B π π π (a) 2x more compact 50% fewer samples Ω t π Ω t Fan Filter (b) Same rate as rectangular sampling Easier to design the filter Rectangular Sampling v V 3 = 0 0 1 t 11

Optimal Sampling v V 3 = 1 0 2 t Hexagonal/Quincunx Sampling!!! Beating Nyquist 12

Beating Nyquist Plenoptic functions are non-stationary Non-Lambertian surfaces Occlusion Non-uniform sampling is preferred Active IBR Determine where to capture the images Resulting a non-uniform sampling scheme A Naive Approach Initial views Voxel space to represent the scene Camera plane For each voxel, check the consistency to the initial views 13

Initial views Voxel space to represent the scene Accumulate all the consistency of visible voxels, compare it with a Camera certain plane threshold:. Consistent: Stop. Inconsistent: Split Initial views New views Voxel space to represent the scene Camera plane Add five new views and split! 14

Initial views New views Voxel space to represent the scene Camera For each plane four neighboring images, check consistency! Initial views New views Voxel space to represent the scene Camera Find the plane quad that is the most inconsisent, split! 15

Initial views Old views New views Voxel space to represent the scene Camera plane The pink views are newly added Initial views Old views New views Voxel space to represent the scene Camera plane Iterate 16

Initial views Old views New views Voxel space to represent the scene Camera plane Iterate Color Inconsistency P C1 Captured Views C2 α 3 α 4 α 2 Virtual View C C3 C4 17

Why Inconsistency? P Occlusion Non Lambertian P C1 Captured Views C2 Virtual View C3 C C4 Poor Geometry C1 Virtual View Captured Views C2 C3 C C4 P C1 Captured Views C2 Virtual View C C4 C3 Progressive Capturing (PCAP) Camera positions Image pair A split of the image pair Q: Where to split, i.e., add one more image? A: The image pair with highest inconsistency 18

Rearranged Capturing (RCAP) Force from the left Force from the right Force proportional to inconsistency Test Scene I: Capturing Non-Uniform Uniform 19

Test Scene I: Rendering Results (I) Non-Uniform Uniform Test Scene II: Capturing Non-Uniform Uniform 20

Test Scene II: Rendering Results (I) Non-Uniform Uniform Self-Reconfigurable Camera Array [Stanford] [Zhang and Chen, CMU] [MIT] 21

Details 48 webcams (embedded processors) 2 step-motors each for translation and pan Real-time capturing/calibration/rendering New architecture may be needed Future: Mirror/Lens Array Many applications 22

Future: Transparent Material Camera Array 3D Display Many applications Future Research Directions IBR composition Merging objects/ibr, alignment, relighting IBR compression Sampling/compressing IBR IBR communication Streaming IBR, error resilience, etc. 23

We can beat Nyquist if we can Reconstruct One Single Image 16 12 8 Number of all possible 16 12 images = 2 >> number of all possible face images [Baker and Kanade, Hallucinating Faces ] >> 30 60 60 24 365 human history world population We can beat Nyquist with prior What does this say? [http://www.palmyra.demon.co.uk] Human is the best sampler!!! 24

Advanced Multimedia Processing Lab Please visit us at: http://amp.ece.cmu.edu 25