View Generation for Free Viewpoint Video System

Similar documents
Key-Words: - Free viewpoint video, view generation, block based disparity map, disparity refinement, rayspace.

View Synthesis for Multiview Video Compression

A New Data Format for Multiview Video

DEPTH-LEVEL-ADAPTIVE VIEW SYNTHESIS FOR 3D VIDEO

View Synthesis for Multiview Video Compression

Joint Tracking and Multiview Video Compression

Client Driven System of Depth Image Based Rendering

FAST MOTION ESTIMATION WITH DUAL SEARCH WINDOW FOR STEREO 3D VIDEO ENCODING

View Synthesis Prediction for Rate-Overhead Reduction in FTV

Rate-distortion Optimized Streaming of Compressed Light Fields with Multiple Representations

Depth Estimation for View Synthesis in Multiview Video Coding

Rate-distortion Optimized Streaming of Compressed Light Fields with Multiple Representations

CONVERSION OF FREE-VIEWPOINT 3D MULTI-VIEW VIDEO FOR STEREOSCOPIC DISPLAYS

WATERMARKING FOR LIGHT FIELD RENDERING 1

3-D Shape Reconstruction from Light Fields Using Voxel Back-Projection

LBP-GUIDED DEPTH IMAGE FILTER. Rui Zhong, Ruimin Hu

Capturing and View-Dependent Rendering of Billboard Models

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

Extensions of H.264/AVC for Multiview Video Compression

A Novel Filling Disocclusion Method Based on Background Extraction. in Depth-Image-Based-Rendering

The Plenoptic videos: Capturing, Rendering and Compression. Chan, SC; Ng, KT; Gan, ZF; Chan, KL; Shum, HY.

Conversion of free-viewpoint 3D multi-view video for stereoscopic displays Do, Q.L.; Zinger, S.; de With, P.H.N.

5LSH0 Advanced Topics Video & Analysis

On the Adoption of Multiview Video Coding in Wireless Multimedia Sensor Networks

An Efficient Mode Selection Algorithm for H.264

Focus on visual rendering quality through content-based depth map coding

DECODING COMPLEXITY CONSTRAINED RATE-DISTORTION OPTIMIZATION FOR THE COMPRESSION OF CONCENTRIC MOSAICS

DEPTH PIXEL CLUSTERING FOR CONSISTENCY TESTING OF MULTIVIEW DEPTH. Pravin Kumar Rana and Markus Flierl

ARTICLE IN PRESS. Signal Processing: Image Communication

A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India

WITH the improvements in high-speed networking, highcapacity

Next-Generation 3D Formats with Depth Map Support

Multi-View Image Coding in 3-D Space Based on 3-D Reconstruction

Graph-based representation for multiview images with complex camera configurations

TREE-STRUCTURED ALGORITHM FOR EFFICIENT SHEARLET-DOMAIN LIGHT FIELD RECONSTRUCTION. Suren Vagharshakyan, Robert Bregovic, Atanas Gotchev

INTERNATIONAL ORGANISATION FOR STANDARISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO

Overview of Multiview Video Coding and Anti-Aliasing for 3D Displays

Reconstruction PSNR [db]

2D/3D Freeview Video Generation for 3DTV System

Video Communication Ecosystems. Research Challenges for Immersive. over Future Internet. Converged Networks & Services (CONES) Research Group

Scene Segmentation by Color and Depth Information and its Applications

Image-Based Rendering. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen

DEPTH IMAGE BASED RENDERING WITH ADVANCED TEXTURE SYNTHESIS. P. Ndjiki-Nya, M. Köppel, D. Doshkov, H. Lakshman, P. Merkle, K. Müller, and T.

But, vision technology falls short. and so does graphics. Image Based Rendering. Ray. Constant radiance. time is fixed. 3D position 2D direction

FRAME-RATE UP-CONVERSION USING TRANSMITTED TRUE MOTION VECTORS

Interactive Free Viewpoint Video Streaming Using Prioritized Network Coding

Image-Based Rendering. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen

DATA FORMAT AND CODING FOR FREE VIEWPOINT VIDEO

Research on Distributed Video Compression Coding Algorithm for Wireless Sensor Networks

Depth Map Boundary Filter for Enhanced View Synthesis in 3D Video

Coding of 3D Videos based on Visual Discomfort

MULTIVIEW video is capable of providing viewers

A Fast Region-Level 3D-Warping Method for Depth-Image-Based Rendering

Bilateral Depth-Discontinuity Filter for Novel View Synthesis

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 24, NO. 5, MAY

Analysis of Depth Map Resampling Filters for Depth-based 3D Video Coding

An overview of free viewpoint Depth-Image-Based Rendering (DIBR)

THE MULTIVIEW video is a collection of multiple videos

Efficient Stereo Image Rectification Method Using Horizontal Baseline

Upcoming Video Standards. Madhukar Budagavi, Ph.D. DSPS R&D Center, Dallas Texas Instruments Inc.

Depth Image-based Techniques for Compression, Transmission and Display of Auto-stereo Video

Template based illumination compensation algorithm for multiview video coding

FILTER BASED ALPHA MATTING FOR DEPTH IMAGE BASED RENDERING. Naoki Kodera, Norishige Fukushima and Yutaka Ishibashi

Morphable 3D-Mosaics: a Hybrid Framework for Photorealistic Walkthroughs of Large Natural Environments

OVERVIEW OF IEEE 1857 VIDEO CODING STANDARD

Enhanced Still 3D Integral Images Rendering Based on Multiprocessor Ray Tracing System

Real-Time Video- Based Modeling and Rendering of 3D Scenes

A Novel Statistical Distortion Model Based on Mixed Laplacian and Uniform Distribution of Mpeg-4 FGS

Multiview Image Compression using Algebraic Constraints

Module 7 VIDEO CODING AND MOTION ESTIMATION

Reducing/eliminating visual artifacts in HEVC by the deblocking filter.

Compression-Induced Rendering Distortion Analysis for Texture/Depth Rate Allocation in 3D Video Compression

Pseudo sequence based 2-D hierarchical coding structure for light-field image compression

Optimizing an Inverse Warper

A Semi-Automatic 2D-to-3D Video Conversion with Adaptive Key-Frame Selection

NEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING

Multiview Depth-Image Compression Using an Extended H.264 Encoder Morvan, Y.; Farin, D.S.; de With, P.H.N.

Multi-View Stereo for Static and Dynamic Scenes

Intermediate view synthesis considering occluded and ambiguously referenced image regions 1. Carnegie Mellon University, Pittsburgh, PA 15213

Fast frame memory access method for H.264/AVC

Enhanced View Synthesis Prediction for Coding of Non-Coplanar 3D Video Sequences

Image-Based Rendering. Image-Based Rendering

Optimizing the Deblocking Algorithm for. H.264 Decoder Implementation

RDTC Optimized Streaming for Remote Browsing in Image-Based Scene Representations

An object-based approach to plenoptic videos. Proceedings - Ieee International Symposium On Circuits And Systems, 2005, p.

BANDWIDTH REDUCTION SCHEMES FOR MPEG-2 TO H.264 TRANSCODER DESIGN

Usings CNNs to Estimate Depth from Stereo Imagery

Optimization of the number of rays in interpolation for light field based free viewpoint systems

Bit Allocation and Encoded View Selection for Optimal Multiview Image Representation

CONTENT ADAPTIVE SCREEN IMAGE SCALING

3D Autostereoscopic Display Image Generation Framework using Direct Light Field Rendering

Prediction Mode Based Reference Line Synthesis for Intra Prediction of Video Coding

Efficient Rendering of Glossy Reflection Using Graphics Hardware

The Light Field and Image-Based Rendering

INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO

What have we leaned so far?

Factorization Method Using Interpolated Feature Tracking via Projective Geometry

Stereo Image Rectification for Simple Panoramic Image Generation

A Low Bit-Rate Video Codec Based on Two-Dimensional Mesh Motion Compensation with Adaptive Interpolation

One-pass bitrate control for MPEG-4 Scalable Video Coding using ρ-domain

Transcription:

View Generation for Free Viewpoint Video System Gangyi JIANG 1, Liangzhong FAN 2, Mei YU 1, Feng Shao 1 1 Faculty of Information Science and Engineering, Ningbo University, Ningbo, 315211, China 2 Ningbo Institute of Technology, Zhejiang University, Ningbo, 315100, China Abstract: View generation is one of the most time consuming parts at user side in free viewpoint video (FVV) system. A user side oriented framework of FVV system is presented, focus on the requirement and capability of user side. The calculation of disparity map, which is required at the user side, is moved to the server side, based on the fact that the user side is a terminal with constrained resource environment and block based disparity map can be easily obtained at the server side due to the usage of disparity compensation prediction at the MVC encoder. Based on the block based disparity map transmitted from server side, a fast pixel-wise disparity refinement scheme is designed to accelerate arbitrary view rendering by using the disparity consistency test. Experimental results show that compared with the conventional block-based matching method the new method speeds up view generation significantly and the quality of virtual view is improved greatly. Keywords:Free viewpoint video; view generation; block based disparity map; disparity refinement; ray-space 1 Introduction It has been recognized that multi-view video signal processing is a key technology for a wide variety of future applications, including free viewpoint video (FVV) system, three-dimensional television, immersive teleconference and surveillance etc [1,2]. FVV can offer a 3D depth impression of the observed scenery and allow interactive selection of viewpoint and direction within a certain operating range, which is regarded as the most challenging audio-video application scenario[3,4]. A classical FVV system, as shown in Fig.1, normally includes the components of multi-view capture, pre-processing, multi-view video coding (MVC) encoder, transmission, MVC decoder, view generation and display[5]. At server side, the most difficult problem is how to improve the coding efficiency of multi-view video. While at user side, the most time-consuming parts are view generation and MVC decoding. Therefore, how to reduce the complexity of these two parts will alleviate the burden of a receiver. Fig.1. The framework of classical FVV system Image-based rendering refers to a collection of techniques and representations that allow 3D scenes and objects to be visualized in a realistic way without full 3D model reconstruction[6.7]. For interactive photo-realistic applications, accurate pixel-wise disparity information is generally used to synthesize high quality novel views[8]. Light field, ray-space and lumigraph are similar ideas which can model radiances of a scene by a 4D plenoptic function[9-13]. A novel view from an arbitrary position and direction can be generated by appropriately combing image pixels from the existing views[4]. However, these benefits have to be paid by dense sampling of the real world with plenty of original view images. In real situation, only sparse multi-view images are compressed by MVC scheme and transmitted to the user side, and dense intermediate view images are generated by some interpolating means[5]. To interpolate dense intermediate view images, accurate disparity map, which determines the quality of synthesized image, will be calculated beforehand. This paper is focused on arbitrary view rendering using block-based disparity map generated at server side, and it is organized as follows: Section 2 illustrates the main components of our framework. Section 3 describes a refinement scheme for block based disparity map at user side. Section 4 gives arbitrary viewpoint rendering used in our system. Section 5 reports some experimental results on different test images and Section 6 is the conclusion. 2 User Side Oriented Free Viewpoint Video System In many applications, user side is a terminal with constrained resource environment, such as TV set, ISBN: 978-960-6766-61-9 200 ISSN: 1790-5117

mobile phone or PDA etc. Its capability of memory and processing is not powerful to process complicated operations. To realize free viewpoint navigation in these devices will confront with many technical problems. This paper aims to view generation at user side with low complexity computation. Compared with the resource limitation at the user side, the complexity of server side can be left out of account. Thus, moving part of time-consuming task to the server side is an alternative scheme. Fig.2 gives the framework of our new system, in which block based disparity map is generated and encoded at the server side. At the user side, the received block based disparity map is refined for view generation, and spatial correlation in block based disparity map is utilized to accelerate the generation of pixel-wise disparity map. By using the refined disparity map, dense ray-space volume is obtained so that novel views can be rendered easily. Block based disparity map Pre-processing Encoder MVC encoder Network Decoder MVC decoder Pixel-wise disparity map View generation Fig.2. The framework of user side oriented FVV system The new framework is based on the following considerations: 1) To improve coding efficiency, correlations among multi-view videos are usually exploited. Disparity compensation prediction algorithm is used to calculate the correspondences between adjacent views. Thus the block based disparity map can be easily obtained at MVC encoding step without any extra computational efforts. 2) For view generation at the user side, pixel-wise disparity map is needed. So if pixel-wise disparity map is generated and transmitted to the user side, disparity estimation can be avoided for the receiver. However, in this case, the total bit rate will increase obviously because of the extra pixel-wise disparity map. Taking into account the tradeoff between the total bit rate and the resolution of disparity map, 8 8 block based disparity map is adopted in our scheme. Since the size of the disparity map is only 1/64 of the original image size, and the values of disparity map is restricted in the range of maximum disparity, the increased bit rate is small compared with original multi-view coding. While in the receiver, it is necessary to refine the block based disparity map to obtain pixel-wise disparity map. 3 Disparity Refinement Scheme At the server side, a disparity estimation method considering neighborhood constraints is used. Let I L and I R represent left image and right image respectively, (x,y) denote the current block to be estimated. An energy function including data item E data and smoothness item E smooth is defined as E E(x, y, dv)=e data (x, y, dv)+λe smooth (x, y, dv) (1) ( x, y, dv) = data w w L i= w j= w 3 smooth = dv dv i i= 0 I ( x + j, y + i) I ( x + dv + j, y + i) R (2) E (3) The smoothness item utilizes neighboring disparity vectors to enhance smoothness of disparity map. The relationship between dv and dv 0 ~dv 3 is illustrated in Fig.3. Factor λ is a constant which controls the influence of the smoothness item on the energy function, the higher value of λ leads to the smoother disparity map. Fig.4(a) and (b) give the original two views of Akko&Kayo, while Fig.4(c) and (d) show two corresponding disparity maps with block size of 8 8. Fig.4(c) is the result without smooth constraint, while Fig.4(d) is obtained by the above disparity estimation method considering neighborhood constraints. From the figures it is seen that the new dv 0 dv 1 dv 2 dv 3 dv Fig.3. Smoothness constraints between disparity vectors (a) original left view (b) oeiginal right view (c) without smooth (d) with smooth constraint constraint Fig.4. Results of block based disparity maps ISBN: 978-960-6766-61-9 201 ISSN: 1790-5117

method can achieve much smoother disparity in the regions belonging to the same object compared with the conventional method. At the user side, pixel-wise disparity map is necessary for view generation. Based on the block based disparity map received from the server side, fast disparity refinement scheme can be designed. Similar to normal 2D image, disparity map has strong spatial correlation. If disparity estimation is accurate, objects in a scene with same depth will have same value of disparity. To reduce complexity of disparity estimation, some blocks with accurate disparities are extracted directly from disparity map by using this smoothness assumption so that the re-estimation of these blocks can be avoided. The disparity refinement algorithm is carried out as the following steps: Step1. In the disparity map with the resolution of 8 8, if disparity dv of current block satisfies Eq.(4), where dv i (i [0, 7]) is disparity of eight neighborhood of current block, the current disparity dv is regarded as an accurate one, so dv will be copied to each pixel within the corresponding block of pixel-wise disparity map. Otherwise, the current 8 8 block will be split into four smaller blocks with the size of 4 4, and they are marked as inaccurate blocks. 1, if ( dv dvi ) < 2, i [0,7] f8 8 ( x, y) = (4) 0, otherwise Step2. In the disparity map with resolution of 4 4, for those inaccurate blocks marked in step1, the disparities are recalculated by the above disparity estimation method considering neighborhood constraints. If dv of current block satisfies Eq.(5), where dv i (i [0, 3]) is disparity of four neighborhood of 4 4 block with respect to dv, the current disparity dv is regarded to be accurate, and it will be copied to the corresponding position of pixel-wise disparity map. Otherwise, they are still marked as inaccurate blocks. 1, if ( dv dvi ) < 2, i [0,3] f 4 4 ( x, y) = (5) 0, otherwise Step3. In the pixel-wise disparity map, 4 4 disparity vector is applied to accelerate the refinement of these inaccurate regions. Let dv denote current disparity of an inaccurate block and it is used as initial position of searching process. A small range for searching the optimal disparity is defined as [dv-4, dv+4]. Therefore, for position that has initial block disparity, the computation complexity of disparity estimation is decreased greatly. Fig.5(a) gives disparity map after 8 8 block based consistency test. The white areas indicate the inaccurate blocks, while the gray regions are copied directly from the 8 8 disparity map. As we can see, the accurate blocks are either background of a scene or at middle part of foreground objects. Fig.5(b) shows the result of 4 4 block based disparity consistency test, in which the white areas indicate the accurate blocks while the grey ones need to be verified by the disparity estimation method considering neighborhood constraints. Fig.5(c) gives the verified 4 4 block based disparity map, the white pixels are those inaccurate blocks even after the verification. Fig.6 gives two pixel-wise disparity maps. Fig.6(a) is generated by the conventional block matching method, while Fig.6(b) is acquired by the proposed refinement method. Compared with the block matching method, the proposed method can save time more than 26 times. Moreover, disparities at smooth regions are more accurate than that of conventional method. (a)8 8 block based result (b)4 4 block based result (c) Verified 4 4 block based disparity map Fig.5. Results of disparity consistency test (a) block matching method (b) proposed method (19.63 seconds) (0.73 seconds) Fig.6. Dense pixel-wise disparity map ISBN: 978-960-6766-61-9 202 ISSN: 1790-5117

4 View Generation Scheme Intermediate view interpolation is vital for IBR(Image Based Rendering) based FVV system, since it determines the quality of synthesized virtual view. When the correspondences between view n and view n+1 are obtained, the dense intermediate disparity map at α position can be generated through the following steps. (1) For the pixel at position x in the view n with disparity d, the position of the corresponding pixel in the intermediate view can be obtained by simply scaling the disparity according to α which indicates the location of the intermediate view between view n and view n+1. Disparity d is assigned to intermediate view pixel at position x+αd. In this case, the values of the pixels between view n and view n+1 are interpolated in term of α. This is illustrated in Fig.7 by the pair of arrows which are opposite with each other. (2) The intermediate disparity map obtained by the above step will have some holes, as shown in Fig.8, in which the holes are indicated by white pixels. For the holes in corresponding intermediate view, the pixels within the holes are regarded as occluded pixels. These pixels can only be seen in one view, and a simple strategy is introduced to estimate their disparities. As we known, the disparities of background objects are normally smaller than that of the foreground objects, and in general, background objects may be occluded by foreground objects. Therefore, for the occluded pixels, the nearest matched pixels at the left and right direction are compared. Let d L and d R denote the corresponding disparities with respect to the left and right directions, then the disparity d of current occluded pixel is determined by between real space and ray-space. Intermediate views with respect to virtual cameras, denoted as the gray ones in the figure, are required to be rendered. The relationship between the virtual view and the ray-space can be explained as X = x + ztanθ (7) Therefore, when dense sampling of a scene is acquired, it is comparatively easy to synthesize a novel view. 1.0 α 0.0 right occluded view n+1 virtual view view n matched left occluded Fig.7. View interpolation Fig.8 Intermediate disparity map with holes d R, and left occluded, if d L d R d = (6) d L, and right occluded, otherwise In this case, the pixels are interpolated according to occlusion type. If current pixel is labeled as left occluded pixel, that is, d L d R, the pixel at the view n indicated by the disparity d R is directly copied to the interpolating pixel. The similar operation can be done to the right occluded pixel. Ray-space representation describes the rays in a scene as a 4D function f(x,y,θ,ϕ), where (θ,ϕ) denotes the direction of the ray, (x,y) denotes the intersection of the ray and the reference plane, and f(x,y,θ,ϕ) represents the intensity of the specific ray. An important feature in ray-space is that an image with respect to a certain viewpoint is given as a sub-space of ray-space. Fig.9 shows the relationship (a) real space (b) ray-space Fig.9. Relationship between real space and ray-space ISBN: 978-960-6766-61-9 203 ISSN: 1790-5117

5 Experimental Results Experiments are performed to evaluate the efficiency of the proposed algorithm. The test image Xmas with 101 parallel views is used to evaluate intermediate view interpolation without camera parameter. The 3DTV test data Akko&Kayo is used to render arbitrary view image. Akko&Kayo is captured by 3 5 camera array, here, 5 images in the middle row are selected, the maximum disparity between two anchor images is 20 pixels. Fig.10 shows PSNR and running time of interpolating intermediate view with different camera intervals. In the figures, the proposed method outperforms BMI(Block-based Matching Interpolation) method[14], because it can obtain more accurate disparities. Moreover, the proposed method runs much faster than BMI because it has initial disparity for searching. For rendering virtual viewpoint image, as illustrated in Fig.11, dense intermediate viewpoint images are firstly interpolated by the proposed method. Here, 19 intermediate viewpoint images are interpolated between two anchor view images according to the sampling theory. Then, these dense multi-view images are projected into ray-space. When given a virtual viewpoint position, the corresponding data can be extracted from the interpolated dense ray-space. Fig.11(a) shows a virtual view at the baseline, while Fig.11(b) and Fig.11(c) show the virtual views at forward and backward positions from the baseline respectively. As we move forward, details of a scene can be seen, meanwhile, the field of view becomes smaller. On the contrary, when we move backward, the field of view becomes larger. The emergence of black regions at top and bottom in Fig.11(c) is resulted from the fact that dense intermediate views are interpolated in horizontal direction but not vertical direction. From the rendered virtual view images, it is seen that the synthesized images have photo-reality without ghosting artifacts. (a) PSNR of interpolated intermediate view (b) run time of interpolating intermediate view (c)original intermediate image (d)bmi (e) 8 8 Proposed method (PSNR=28.36dB,TIME=32.8second) (PSNR=36.17dB,TIME=1.1second) Fig.10. Result of intermediate view generation of Xmas (maximum disparity=50) ISBN: 978-960-6766-61-9 204 ISSN: 1790-5117

(a) virtual view at baseline (b) virtual view forward 0.5m (c) virtual view backward 0.5m Fig.11. Results of arbitrary view rendering. 6 Conclusions In this paper, a user side oriented framework of free viewpoint video system is presented, which focus on the requirement and capability of user side. The calculation of disparity map, which is required at the user side, is moved to the server side, based on the fact that the user side is a terminal with constrained resource environment and block based disparity map can be easily obtained at the server side due to the usage of disparity compensation prediction at the MVC encoder. This change will alleviate the burden of view synthesis and it is more suitable for real time applications. Acknowledgement This work was supported by the Natural Science Foundation of China (grant 60472100, 60672073), the Program for New Century Excellent Talents in University (NCET-06-0537), Natural Science Foundation of Ningbo (grant 2007A610037). References: [1] Ishfaq Ahmad, Multiview Video: Get Ready for Next-Generation Television, IEEE Distributed Systems Online, vol. 8, no.3, 2007, art. no. 0703-o3006. [2] A. Smolic, D. McCutchen, 3DAV Exploration of Video-Based Rendering Technology in MPEG, IEEE Transactions on Circuits and Systems for Video Technology, vol. 14, no. 3, 2004, pp. 348-356. [3] H. Kimata, M. Kitahara, K. Kamikura, et al., Low-Delay Multiview Video Coding for Free-Viewpoint Video Communication, Systems and Computers in Japan, vol. 38, no. 5, 2007, pp.15-29. [4] A. Smolic, K. Mueller, P. Merkle, et al., 3D Video and Free Viewpoint Video - Technologies, Applications and MPEG Standards, IEEE Int. Conf on Multimedia and Expo, 2006, pp.2161-2164. [5] M. Tanimoto, T. Fujii, H. Kimata, S. Sakazawa, Proposal on Requirements for FTV, Joint Video Team (JVT) of ISO/IEC MPEG & ITU-T VCEG, JVT-W127, 23rd Meeting: San Jose, California, Apr. 2007. [6] S.C. Chan, H.-Y. Shum, K.-T. Ng, Image-Based Rendering and Synthesis, IEEE Signal Processing Magazine, vol.24, no.6, 2007, pp.22-33. [7] C. Zhang T. Chen, A Survey on Image-Based Rendering-Representation, Sampling and Compression, EURASIP Signal Processing: Image Commun., vol. 19, no. 1, 2004, pp. 1-28. [8] E. Kurutepe, M. R. Civanlar, A. M. Tekalp, Interactive Transport of Multi-view Videos for 3DTV Applications, Journal of Zhejiang University Science, vol. 7, no. 5, 2006, pp. 830-836. [9] M. Levoy, P. Hanrahan, Light Field Rendering, Proc. of Siggraph, 1996, pp. 31-42. [10] X. Tong, R. M. Gray, Interactive Rendering from Compressed Light Fields, IEEE Trans. Circuits Syst. Video Techn. vol.13, no.11, 2003, pp.1080-1091. [11] N. Fukushima, T. Yendo and T. Fujii, Real-Time Arbitrary View Interpolation and Rendering System Using Ray-Space, Proceedings of SPIE Three-Dimensional TV, Video and Display, 2005, pp. 250-261. [12] L. Zhong, M. Yu, Y. Zhou, G. Jiang, Quality and Performance Evaluation of Ray-Space Interpolation for Free Viewpoint Video Systems, Lecture Notes in Computer Science, LNCS3992, 2006, pp.367-370. [13] S. J. Gortler, R. Grzeszczuk, R. Szeliski, M. F. Cohen.: The Lumigraph, Proc. of SIGGRAPH-96, 1996, pp.31-42. [14] T. Yendo, T. Fujii, M. Tanimoto, Ray-Space Acquisition and Reconstruction within Cylindrical Objective Space, Proceedings of SPIE Stereoscopic Displays and Virtual Reality Ssytem, vol.6055, 2006, pp. 299-306. ISBN: 978-960-6766-61-9 205 ISSN: 1790-5117