recording plane (U V) image plane (S T)

Similar documents
Reconstruction PSNR [db]

Dimensional Imaging IWSNHC3DI'99, Santorini, Greece, September SYNTHETIC HYBRID OR NATURAL FIT?

Figure 1. Block-based depth map estimation. Figure 2. Disparity-compensated image prediction. two images being recorded from arbitrary positions P 1,

voxel volume camera views

3-D Shape Reconstruction from Light Fields Using Voxel Back-Projection

Rate-distortion Optimized Streaming of Compressed Light Fields with Multiple Representations

Rate-distortion Optimized Streaming of Compressed Light Fields with Multiple Representations

Multi-View Image Coding in 3-D Space Based on 3-D Reconstruction

A Warping-based Refinement of Lumigraphs

Compression of Light Field Images using Projective 2-D Warping method and Block matching

Facial Expression Analysis for Model-Based Coding of Video Sequences

Mesh Based Interpolative Coding (MBIC)

Capturing and View-Dependent Rendering of Billboard Models

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 15, NO. 4, APRIL

Rate Distortion Optimization in Video Compression

Light Field Techniques for Reflections and Refractions

Model-Aided Coding: A New Approach to Incorporate Facial Animation into Motion-Compensated Video Coding

DECODING COMPLEXITY CONSTRAINED RATE-DISTORTION OPTIMIZATION FOR THE COMPRESSION OF CONCENTRIC MOSAICS

Compression of Lumigraph with Multiple Reference Frame (MRF) Prediction and Just-in-time Rendering

Review and Implementation of DWT based Scalable Video Coding with Scalable Motion Coding.

Jingyi Yu CISC 849. Department of Computer and Information Science

A High Quality/Low Computational Cost Technique for Block Matching Motion Estimation

Lazy decompression of surface light fields for precomputed global illumination

Light Fields. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen

Optimized Progressive Coding of Stereo Images Using Discrete Wavelet Transform

Compression of Stereo Images using a Huffman-Zip Scheme

EFFICIENT INTRA PREDICTION SCHEME FOR LIGHT FIELD IMAGE COMPRESSION

Stereo Image Compression

I. DARIBO, R. FURUKAWA, R. SAGAWA, H. KAWASAKI, S. HIURA, and N. ASADA. 1. Introduction IS4-15 : 1385

View Generation for Free Viewpoint Video System

IN the early 1980 s, video compression made the leap from

View Synthesis for Multiview Video Compression

High-Quality Interactive Lumigraph Rendering Through Warping

Image-Based Rendering

Model-based Enhancement of Lighting Conditions in Image Sequences

View Synthesis for Multiview Video Compression

Comparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV

Overview: motion-compensated coding

Differential Compression and Optimal Caching Methods for Content-Based Image Search Systems

Depth Estimation for View Synthesis in Multiview Video Coding

Morphable 3D-Mosaics: a Hybrid Framework for Photorealistic Walkthroughs of Large Natural Environments

RDTC Optimized Streaming for Remote Browsing in Image-Based Scene Representations

LONG-TERM MEMORY PREDICTION USING AFFINE MOTION COMPENSATION

Modified SPIHT Image Coder For Wireless Communication

Rate-Distortion Analysis for Light Field. Coding and Streaming

signal-to-noise ratio (PSNR), 2

Model-Aided Coding: A New Approach to Incorporate Facial Animation into Motion-Compensated Video Coding

The Video Z-buffer: A Concept for Facilitating Monoscopic Image Compression by exploiting the 3-D Stereoscopic Depth map

QUALITY-BASED IMPROVEMENT OF QUANTIZATION FOR LIGHT FIELD COMPRESSION

2014 Summer School on MPEG/VCEG Video. Video Coding Concept

( ) ; For N=1: g 1. g n

A Low Bit-Rate Video Codec Based on Two-Dimensional Mesh Motion Compensation with Adaptive Interpolation

The Plenoptic videos: Capturing, Rendering and Compression. Chan, SC; Ng, KT; Gan, ZF; Chan, KL; Shum, HY.

An Optimized Template Matching Approach to Intra Coding in Video/Image Compression

Extensions of H.264/AVC for Multiview Video Compression

View Synthesis Prediction for Rate-Overhead Reduction in FTV

Video Compression An Introduction

Wavelet-Based Video Compression Using Long-Term Memory Motion-Compensated Prediction and Context-Based Adaptive Arithmetic Coding

Decoded. Frame. Decoded. Frame. Warped. Frame. Warped. Frame. current frame

Network Image Coding for Multicast

Image Base Rendering: An Introduction

MRT based Fixed Block size Transform Coding

International Journal of Emerging Technology and Advanced Engineering Website: (ISSN , Volume 2, Issue 4, April 2012)

Image or Object? Is this real?

DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS

Real Time Rendering. CS 563 Advanced Topics in Computer Graphics. Songxiang Gu Jan, 31, 2005

DEPTH MAPS CODING WITH ELASTIC CONTOURS AND 3D SURFACE PREDICTION

University of Bristol - Explore Bristol Research. Link to published version (if available): /s

Video Transcoding Architectures and Techniques: An Overview. IEEE Signal Processing Magazine March 2003 Present by Chen-hsiu Huang

Real-time Shading: Sampling Procedural Shaders

INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO

Multiview Depth-Image Compression Using an Extended H.264 Encoder Morvan, Y.; Farin, D.S.; de With, P.H.N.

ECE 417 Guest Lecture Video Compression in MPEG-1/2/4. Min-Hsuan Tsai Apr 02, 2013

An Embedded Wavelet Video. Set Partitioning in Hierarchical. Beong-Jo Kim and William A. Pearlman

Image Segmentation Techniques for Object-Based Coding

A combined fractal and wavelet image compression approach

Acquisition and Visualization of Colored 3D Objects

Light-Dependent Texture Mapping

Efficient Rendering of Glossy Reflection Using Graphics Hardware

Express Letters. A Simple and Efficient Search Algorithm for Block-Matching Motion Estimation. Jianhua Lu and Ming L. Liou

Efficient Free Form Light Field Rendering

International Journal of Advanced Research in Computer Science and Software Engineering

Advanced Video Coding: The new H.264 video compression standard

Outline Introduction MPEG-2 MPEG-4. Video Compression. Introduction to MPEG. Prof. Pratikgiri Goswami

A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India

Ingo Scholz, Joachim Denzler, Heinrich Niemann Calibration of Real Scenes for the Reconstruction of Dynamic Light Fields

Speech Driven Synthesis of Talking Head Sequences

WATERMARKING FOR LIGHT FIELD RENDERING 1

New Techniques for Improved Video Coding

Image and Video Watermarking

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

Compression of RADARSAT Data with Block Adaptive Wavelets Abstract: 1. Introduction

Efficient Image Compression of Medical Images Using the Wavelet Transform and Fuzzy c-means Clustering on Regions of Interest.

1740 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. 7, JULY /$ IEEE

A Hybrid Temporal-SNR Fine-Granular Scalability for Internet Video

An Embedded Wavelet Video Coder. Using Three-Dimensional Set. Partitioning in Hierarchical Trees. Beong-Jo Kim and William A.

BLOCK MATCHING-BASED MOTION COMPENSATION WITH ARBITRARY ACCURACY USING ADAPTIVE INTERPOLATION FILTERS

Very Low Bit Rate Color Video

CSEP 521 Applied Algorithms Spring Lossy Image Compression

More and More on Light Fields. Last Lecture

Transcription:

IEEE ONFERENE PROEEDINGS: INTERNATIONAL ONFERENE ON IMAGE PROESSING (IIP-99) KOBE, JAPAN, VOL 3, PP 334-338, OTOBER 1999 HIERARHIAL ODING OF LIGHT FIELDS WITH DISPARITY MAPS Marcus Magnor and Bernd Girod Telecommunications Laboratory, University of Erlangen-Nuremberg auerstraße 7, 91058 Erlangen, Germany email: fmagnor,girodg@nte-technikuni-erlangende ABSTRAT A coder for light fields is presented Due to the large amount of data needed to represent a complete light field, a hierarchical decomposition is employed The full light field is built up by recursively predicting intermediate images from a small subset of light-field images Intermediate images are predicted by disparitycompensating multiple surrounding images The predicted images are refined using DT coding Ratedistortion measurements for two standard light fields verify the efficiency of the proposed scheme ompression ratios of 1000:1 are achieved at acceptable quality of the rendered views 1 INTRODUTION ommon computer graphics rendering schemes are based on descriptions of object geometry, texture properties and illumination conditions Photorealistic rendering results require very detailed geometry models and the computationally demanding simulation of global illumination effects Acquisition of geometry information from natural objects is difficult and can even prove impossible if objects exhibit ill-defined surfaces (eg, fur, hair, smoke, etc) In 1996, a new technique called Light Field Rendering (LFR) [1, 2] was proposed that promises to overcome the limitations of traditional geometry-based rendering Instead of 3-D geometry descriptions, LFR relies on a set of ordinary 2-D images of a static scene, enabling easy data acquisition from virtual as well as real-world scenes If the light field is sufficiently densely sampled, LFR can provide photorealistic rendered images at interactive frame rates, regardless of scene complexity Unfortunately, huge amounts of data are required to attain photorealistic rendering results: tens to hundreds of thousands of full-sized color images need to be compressed efficiently while maintaining fast access to arbitrary data segments In [1], vector quantization of the light field yields compression ratios of 27:1 A block-based DT coding scheme is used in [3] to compress light-field data by a factor of 24:1 For fast data transmission and rendering on standard hardware, much higher compression ratios are desirable We propose a light-field coder that exploits interimage similarities by compensating disparity between light-field images All images are coded in hierarchical order for fast access to multiple resolution levels Images that are not contained in the recorded light-field image set can be estimated by disparity compensation to facilitate rendering of photorealistic views from subsampled light fields This paper is organized as follows In the next section, we describe the light-field data structure and its sub-sampled representation We go on to explain how disparity maps are derived for all images and how lightfield images are predicted The light-field coding algorithm is described and experimental results are presented to validate the proposed coding scheme 2 THE LIGHT FIELD An illuminated, static object fills the surrounding space with light reflected off its surfaces, establishing the object's light field [4] The light field can be captured by recording images from multiple view points around the object (Fig 1) The light-field images' recording positions are arranged on planes facing the object [1, 2] On each plane, the recording positions are arranged in a regular grid, so the light-field structure resembles a two-dimensional array of images Optimal rendering quality is achieved if disparity between adjacent images does not exceed one pixel Respecting this constraint, the number of images needed to capture a complete light field is proportional to image resolution For example, to record a light field at 256 2 pixels resolution requires ß 262000 images, which

recording plane (U V) image plane (S T) Recording a light field Light Field Rendering Figure 1: An object's light field is recorded by taking a number of conventional images from regularly spaced positions in a plane During Light Field Rendering, the viewing rays' intersection points with the recording plane and image plane are calculated, and the image array is resampled corresponds to 48 GBytes, given that each pixel is represented by 24 bits Acquisition of such a large number of images is practically impossible Therefore, light fields are almost always a sub-sampled representation of the complete light-field information During rendering, the image array is resampled by tracing rays from the view point along the viewing direction The rays' intersection points with recording and image plane determine which light-field segments to render Because intersection coordinates do not, in general, coincide with image recording positions, inbetween image segments must be estimated based on the available sub-sampled light-field data To avoid aliasing, quadralinear interpolation is applied in [1, 2], resulting in low-pass filtered (blurred) rendered images To avoid unsharp rendering results, the available array of light-field images has to be refined by disparity-compensated intermediate images The presented coder allows to estimate disparity-compensated intermediate images, enhancing rendering quality from sub-sampled light fields 3 DISPARITY MAP ESTIMATION When comparing two adjacent light-field images, a point on the surface of the depicted object appears at different positions in the images This displacement is the point's parallax or disparity From known image recording positions, the disparity direction between two images can be inferred The disparity magnitude is a scalar value associated with each pixel that describes the amount of shift along the disparity direction An image's disparity map is an array of scalar values that lists the amount of shift for each pixel when compared to neighboring images As object geometry is not recorded during light field acquisition, disparity information must be retrieved from the light-field images Disparity maps are estimated on image blocks (Fig 2) Because the light-field images are arranged in a regular grid of equal spacing, an image block's true disparity magnitude is the same when comparing the block to all four neighboring images along the respective disparity direction To find the magnitude of disparity for each block, a number of disparity values within a predefined search range are considered For each disparity value, the corresponding blocks from all four neighboring images are extracted The extracted blocks are averaged and compared to the original block The disparity value resulting in the smallest prediction error is chosen as the block's disparity magnitude Full-pel disparity accuracy was found to be sufficient, in accordance with considerations in [5] for multi-hypothesis compensation A fixed Huffmancode tabe is used to code the disparity values In contrast to geometry reconstruction algorithms, the described method converges regardless of light-field scene characteristics The estimated low-resolution disparity maps allow optimal disparity compensation of light-field images from surrounding images, and disparity-compensated prediction of intermediate (missing) images becomes possible oding efficiency can be optimized by varying the disparity maps' block size If the light-field object's full 3-D geometry can be estimated, high resolution disparity maps are available Instead of Huffman-coding each disparity map entry, arbitrarily many maps can be generated from the object model at constant geometry coding bit-rate Results for geometry-aided light-field coding are presented in [6] in these proceedings

Disparity map estimation Disparity compensation Figure 2: Disparity maps are estimated for all light-field images by comparing image blocks to adjacent images Image prediction is performed by first estimating the target image's disparity map which is used for disparitycompensating from all reference images 4 DISPARITY OMPENSATION The disparity maps are estimated with one-pixel accuracy in respect to neighboring images If reference and target images are farther apart, full-pixel accurate compensation is possible only at lower image resolution Therefore, prediction over greater distances is performed after the images have been downsampled by an appropriate factor The disparity-compensated image is upsampled again to yield the target image's prediction Given multiple reference images and their corresponding disparity maps, it was experimentally found that a target image and its disparity map can best be predicted in the following manner (Fig 2): 1 All reference images' disparity maps are low-pass filtered and forward-compensated to the target image's position, yielding an estimate for the target image's disparity map 2 The estimated disparity map is used to backwardcompensate the target image from all reference images The disparity maps are low-pass filtered to yield smooth disparity fields If reference and target positions are farther apart, the smoothed disparity maps are downsampled appropriately The disparity maps are then forward-compensated to the target image's position The resulting disparity maps are averaged and undefined patches, if any, are filled by interpolation By upsampling to original image size, an estimate for the target image's disparity map is derived The target image itself is predicted by using the estimated disparity map and backward-compensating from all reference images If reference and target images are not adjacent, the reference images are downsampled All reference images are disparitycompensated, and the prediction results are averaged Finally, the estimated image is upsampled to yield the target image 5 THE ODER Prior to coding a light field containing U V images, a minimum reconstruction quality parameter q min is set All reconstructed light-field images must meet this quality criterion Image quality is measured as the peak-signal-to-noise ratio (PSNR) of the image's luminance signal First, the image array's four corner images (1; 1), (U; 1), (U; V ), (1;V) are intra-coded using a standard block-dt scheme adopted from still-image compression (positions A in Fig 3) For each image, the DT quantization parameter is individually chosen to ensure that the reconstructed image meets the reconstruction value q min The corner images' disparity maps are Huffman-coded applying a fixed table From the four reconstructed corner images, the center image (U=2;V=2) and its disparity map (position B in Fig 3) is predicted as described If the image estimate meets the reconstruction criterion q min, no further information regarding the center image is coded Otherwise, the center image is again predicted using its original disparity map which is then Huffman-coded If image quality still doesn't suffice, the prediction error is additionally DT-coded The DT quantizer level is thereby adjusted to yield minimum bit-rate for the required image quality Then, the four middle images on the array sides (U=2; 1), (U; V=2), (U=2;V), (1;V=2) (positions in Fig 3) are predicted from the reconstructed center image and the two closest corner images If necessary, the residual error and disparity

1 1 U A A B Figure 4: Dragon Images from the light fields Buddha and V A A Figure 3: oding order of the light field image array: from the corner images (A), the center image (B) is predicted The images at the middle of the sides () are predicted from the center image and the two closest corner images The array is subdivided into quadrants and each quadrant is coded likewise The algorithm keeps recursing until all images are coded maps are coded At this point, 9 light field images spanning the full recording plane are available The image array is now divided into four quadrants The four corner images of each quadrant are already coded and, as before, the center and side images in each quadrant are predicted The algorithm keeps recursing through the quadtree structure until all images are coded The proposed decoder allows access to different levels of increasing sampling density of the U V image array Further, the decoder can locally refine the light field by estimating disparity-compensated intermediate light-field images that were not originally recorded These interpolated images can greatly enhance rendering performance [7] 6 RESULTS The proposed coding scheme is validated using two publically available light fields (Buddha, Dragon, Fig 4) 1 Both light fields contain U V =32 32 images of S T = 256 256 24-bit RGB pixels, totalling 192 MBytes In both data sets, disparity between adjacent images ranges from -3 to 3 pixels Disparity maps with different block sizes are estimated for all light-field images Both light fields 1 www-graphicsstanfordedu/software/lightpack/lifshtml are coded for varying reconstruction quality parameter q min The reconstructed images are compared to the original light-field images, and the distortion is measured (PSNR) and averaged over all images The rate-distortion curves for various disparity map resolutions are shown in Fig 5 For both light fields, maps of block size 8 8 and 16 16 yield the best allocation of bit-rate between disparity information and residual error for all reconstruction distortion values considered oding efficiency depends on reconstruction quality and light-field scene characteristics The Dragon object exhibits detailed texture and complex geometry 569 kbytes are needed to code its light field at 36 db PSNR The Buddha statue appears smaller in the images and has a more compact geometry, enabling the light field to be compressed to 222 kbytes at 40 db PSNR Fig 6 depicts the number of light-field images that need encoding of the residual error At lower image reconstruction quality, many light-field images are pre- Number of coded images 1000 800 600 400 200 buddha light field dragon light field 0 30 35 40 45 Figure 6: Number of residual-error coded light-field images (disparity map block size 16 16 pixels)

Buddha Light Field Dragon Light Field 44 42 4x4 8x8 16x16 32x32 40 38 4x4 8x8 16x16 32x32 40 38 36 34 36 32 34 0 002 004 006 008 01 Bit Rate [bpp] 30 0 005 01 015 02 025 Bit Rate [bpp] Figure 5: Rate-distortion measurements for 1024 images in each light field at different disparity map block sizes dicted well enough by disparity compensation to neglect the residual error This way, the coder decimates the number of light-field images and only codes relevant images The discarded images are reconstructed by merely disparity-compensating surrounding images 7 ONLUSIONS A coding scheme has been presented that facilitates hierarchical decoding of the light-field data structure The presented algorithm allows disparity-compensated estimation of missing light-field information, yielding enhanced rendering results At reasonable reconstruction distortions, light fields can be compressed by factors of 1000:1 and higher if compared to the amount of recorded light-field data available as input With regard to the complete light-field information necessary for optimal rendering performance, compression ratios exceeding 10,000:1 are achieved These bit-rates allow convenient storage of all light-field data in local memory for rendering on standard hardware Also, for possible future 3D-television or Internet distribution applications, capacity requirements for transmitting light-field data is greatly reduced [3] Gavin Miller, Steven Rubin, and Dulce Ponceleon, Lazy decompression of surface light fields for precomputed global illumination", Rendering Techniques '98 (Eurographics Proceedings), pp 281 292, June 1998 [4] A Gershun, The light field", J Math Phys, vol 18, pp 51 151, 1939 [5] B Girod, Efficiency analysis of multi-hypothesis motion-compensated prediction for video coding", IEEE Trans Image Processing, in print [6] B Girod, P Eisert, M Magnor, E Steinbach, and T Wiegand, 3-d image models and compression - synthetic hybrid or natural fit?", Proc International onference on Image Processing IIP-99, Kobe, Japan, Oct 1999 [7] W Heidrich, H Schirmacher, H Kück, and H- P Seidel, A warping-based refinement of lumigraphs", Proc 6th International onference in entral Europe on omputer Graphics and Visualization, 1998 8 REFERENES [1] Marc Levoy and Pat Hanrahan, Light field rendering", omputer Graphics (SIGGRAPH '96 Proceedings), pp 31 42, Aug 1996 [2] Steven J Gortler, Radek Grzeszczuk, Richard Szelinski, and Michael F ohen, The Lumigraph", omputer Graphics (SIGGRAPH '96 Proceedings), pp 43 54, Aug 1996