Image-based rendering using plane-sweeping modelisation

Similar documents
Real-time Plane-Sweep with local strategy

Real-Time Video-Based Rendering from Multiple Cameras

Multi-view stereo. Many slides adapted from S. Seitz

Real-Time Free Viewpoint from Multiple Moving Cameras

But, vision technology falls short. and so does graphics. Image Based Rendering. Ray. Constant radiance. time is fixed. 3D position 2D direction

Online Multiple View Computation for Autostereoscopic Display

Some books on linear algebra

3-D Shape Reconstruction from Light Fields Using Voxel Back-Projection

Some books on linear algebra

On-line Free-viewpoint Video: From Single to Multiple View Rendering

Photo Tourism: Exploring Photo Collections in 3D

Static Scene Reconstruction

Volumetric Scene Reconstruction from Multiple Views

Capturing and View-Dependent Rendering of Billboard Models

An Efficient Visual Hull Computation Algorithm

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

Polyhedral Visual Hulls for Real-Time Rendering

Image Base Rendering: An Introduction

VIDEO-TO-3D. Marc Pollefeys, Luc Van Gool, Maarten Vergauwen, Kurt Cornelis, Frank Verbiest, Jan Tops

Evaluation of Large Scale Scene Reconstruction

Occlusion Detection of Real Objects using Contour Based Stereo Matching

Hardware-Accelerated Rendering of Photo Hulls

Appearance-Based Virtual View Generation From Multicamera Videos Captured in the 3-D Room

Real Time Rendering. CS 563 Advanced Topics in Computer Graphics. Songxiang Gu Jan, 31, 2005

Modeling, Combining, and Rendering Dynamic Real-World Events From Image Sequences

Foundations of Image Understanding, L. S. Davis, ed., c Kluwer, Boston, 2001, pp

CS 684 Fall 2005 Image-based Modeling and Rendering. Ruigang Yang

Morphable 3D-Mosaics: a Hybrid Framework for Photorealistic Walkthroughs of Large Natural Environments

Using Shape Priors to Regularize Intermediate Views in Wide-Baseline Image-Based Rendering

More and More on Light Fields. Last Lecture

Active Scene Capturing for Image-Based Rendering with a Light Field Setup

VOLUMETRIC MODEL REFINEMENT BY SHELL CARVING

Multiview Reconstruction

3D Photography: Stereo Matching

A Warping-based Refinement of Lumigraphs

Structure from Motion and Multi- view Geometry. Last lecture

On-line and Off-line 3D Reconstruction for Crisis Management Applications

Acquisition and Visualization of Colored 3D Objects

Hybrid Rendering for Collaborative, Immersive Virtual Environments

Interactive 3D Scene Reconstruction from Images

Efficient View-Dependent Image-Based Rendering with Projective Texture-Mapping

Photo Tourism: Exploring Photo Collections in 3D

Hardware-Accelerated Visual Hull Reconstruction and Rendering

Image-Based Modeling and Rendering

Stereo Vision. MAN-522 Computer Vision

Prof. Trevor Darrell Lecture 18: Multiview and Photometric Stereo

Layered Depth Panoramas

Noah Snavely Steven M. Seitz. Richard Szeliski. University of Washington. Microsoft Research. Modified from authors slides

Mobile Point Fusion. Real-time 3d surface reconstruction out of depth images on a mobile platform

Multi-View Imaging: Capturing and Rendering Interactive Environments

View Synthesis for Multiview Video Compression

View Synthesis using Convex and Visual Hulls. Abstract

A Statistical Consistency Check for the Space Carving Algorithm.

Efficient View-Dependent Sampling of Visual Hulls

Image-Based Modeling and Rendering. Image-Based Modeling and Rendering. Final projects IBMR. What we have learnt so far. What IBMR is about

Real-Time Consensus-Based Scene Reconstruction using Commodity Graphics Hardware

Geometric Reconstruction Dense reconstruction of scene geometry

What have we leaned so far?

Mikio Terasawa* Yasushi Yamaguchiy. Kinji Odakaz. *College of Economics, Nihon University Misaki-cho, Chiyoda-ku, Tokyo Japan

Miniature faking. In close-up photo, the depth of field is limited.

Applications of Explicit Early-Z Culling

camera path backtracking i-1 look-ahead search range viewpoint surface

Joint 3D-Reconstruction and Background Separation in Multiple Views using Graph Cuts

Real-Time Consensus-Based Scene Reconstruction using Commodity Graphics Hardware

CS451Real-time Rendering Pipeline

Image-Based Rendering

Visual-Geometric Scene Reconstruction from Image Streams

Image-Based Rendering. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen

First Steps in Hardware Two-Level Volume Rendering

3D Surface Reconstruction from 2D Multiview Images using Voxel Mapping

Modeling Light. Michal Havlik : Computational Photography Alexei Efros, CMU, Fall 2007

Free-viewpoint video renderer

Dense Matching of Multiple Wide-baseline Views

Optimizing an Inverse Warper

Image-Based Rendering and Light Fields

On the Use of Ray-tracing for Viewpoint Interpolation in Panoramic Imagery

A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India

MTMesh: Image Based Mesh Reconstruction and Rendering

A million pixels, a million polygons. Which is heavier? François X. Sillion. imagis* Grenoble, France

A virtual tour of free viewpoint rendering

Plenoptic Image Editing

Keywords:Synthetic Data, IBR, Data Generation Tool. Abstract

Real-Time Reflection on Moving Vehicles in Urban Environments

Multiple View Geometry

CSc Topics in Computer Graphics 3D Photography

Visual Shapes of Silhouette Sets

Department of Computer Engineering, Middle East Technical University, Ankara, Turkey, TR-06531

Image or Object? Is this real?

Efficient Rendering of Glossy Reflection Using Graphics Hardware

Image-Based Modeling and Rendering

Multiple View Geometry

Sashi Kumar Penta COMP Final Project Report Department of Computer Science, UNC at Chapel Hill 13 Dec, 2006

Multi-View 3D-Reconstruction

Ingo Scholz, Joachim Denzler, Heinrich Niemann Calibration of Real Scenes for the Reconstruction of Dynamic Light Fields

A Real-Time Distributed Light Field Camera

Lecture 8 Active stereo & Volumetric stereo

A unified approach for motion analysis and view synthesis Λ

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

Traditional Image Generation. Reflectance Fields. The Light Field. The Light Field. The Light Field. The Light Field

Interactive Free-Viewpoint Video

Transcription:

Author manuscript, published in "IAPR Machine Vision and Applications MVA2005, Japan (2005)" Image-based rendering using plane-sweeping modelisation Vincent Nozick, Sylvain Michelin and Didier Arquès Marne la Vallée University, Gaspard Monge Institut, SISAR team 6 cours du Danube, 77700 Serris, France {vnozick, michelin, arques}@univ-mlv.fr Abstract : This paper presents a method to create new images of a scene from a small set of calibrated images using color matching. While most of the existing techniques attempt to build a 3d reconstruction of the scene, the only geometric constraint our color matching method requires is camera calibration. This paper describes a new adaptation of the plane-sweeping algorithm using a local dynamic scoring method that handle occlusions. We also propose a real-time implementation of our method using fragment programs. These modifications are well suited for on-line real-time rendering of dynamic scenes. Keywords : real-time vision, image-based rendering, fragment programs 1 Introduction Realistic rendering can be achieved using two approaches : computer graphics where an image is generated from a scene description and computer vision where the image is generated from pictures of real objects. The second approach often uses geometric reconstruction like texture map rendering [SS02, PVG00] or volumetric techniques [SD97], which implies geometric interpretation of the scene. These techniques become more and more accurate but they require a large set of input images and a lengthy computation time before visualisation. They also need complex points and surfaces reconstruction algorithms, accurate feature extraction and a robust calibration method. However these techniques allow a fluid navigation in the reconstructed scene. The main drawback of these methods is the cost of the computation which rules out their use for animation in the case of a dynamic scene. Implicit geometry methods like [MB95, GSC96, LH96] generate new views without any geometric interpretation of the scene and achieve real-time rendering using interpolation between views but they need a large set of input images. [MBR + 00] propose a real-time reconstruction that shades visual hulls from silouhette image data but therefore can not handle concave objects. Finally, view-dependent geometry methods provide realistic rendering without geometric reconstruction nor correspondence estimation between different images. Some methods such as [IHA02, FWZ03, YWB02] generate new views without any other geometric information than the calibration matrices of the input images. These methods are based on color matching betwen differents views. They do not explicitly compute 3d geometry but this implicitly appears in the resulting images. Our algorithm belongs to this latter familly and follows a plane-sweeping algorithm. We will first present the basic plane-sweeping algorithm and some related work. Then, we will introduce our new scoring method and [explain/detail] our implementation. 2 Basic plane-sweeping algorithm and related work Given a small set of calibrated images from video cameras, we wish to generate a new view of the scene from a new viewpoint. Considering a scene where objects are exclusively diffuse, we first place the virtual camera and divide space in parallel planes D i in front of the camera as shown in Fig. 1. We project the input images onto each plane D i in a back to front order. Let s consider a visible object of the scene lying on one of these planes at a point M. The input cameras will project on M the same color (ie the object color). Therefore, points on the planes D i where projected colors match together potentially correspond to an object of the scene.

far I1 near I2 cam1 camx (a) Plane-sweeping algorithm : geometric configuration. cam2 Figure 1: Plane-sweeping algorithm (b) Four pictures associated to four steps of the algorithm using two input cameras. Let I 1... I n denote a set of n images and P 1...P n their respective 3 4 projection matrices. I x is the new image to be computed, cam x is its virtual pinhole projective camera and P x its associated projection matrix. We define a near plane and a far plane parallel to cam x image plane such that all the objects of the scenes lie between near and far. For each pixel of each plane D i, a score and a color are computed according to the matching of the colors. The plane-sweeping algorithm can be explained as follows : initialize I x s score for each plane D i from far to near project all the input image I 1...I n on D i as textures project D i multi-textured on I x for each pixel p of I x draw I x - compute a score and a color according to the coherence of the colors from each camera s contribution - if the score is better than the previous ones then update the score and the color of p Yang et al. [YWB02] propose an implementation of the plane-sweeping algorithm using pixel shaders and reach real-time rendering [for dynamic scenes] using five input cameras. For the scoring stage, they choose a reference camera that is closest to cam x and pairwise compare the contribution of each input image with the reference image. According to the small number of instructions, this method provides good speed results, however the occlusions are not handled. This method also implies geometric constraints on the position of the cameras : the input cameras have to be close to each other and the navigation of the virtual camera should lie between the viewpoints of the input cameras, otherwise the reference camera may not be representative of cam x. Lastly, there may appear discontinuities in the computed images when the virtual camera moves and changes its reference camera.

3 New color scores The score computation is a crucial step in the plane-sweeping algorithm. Both visual results and speedy computation depend on it. We propose a new method to compute a score according to all the input image colors instead of computing by pairs. For this purpose, we use multi-texturing fonctions to access to each input camera color contribution. For each pixel of I x, we propose a finite number of positions X in the scene (one per plane D). A score is computed for each position and this score depends on the color C i of the projections of X in each input image. We propose an iterative algorithm to reject outliers colors and compute an optimal score. This method first computes the variance v of the color set S = {C i } i=1...n and finds the color C f S the furthest from the average. If this color is further than a defined distance d, then it is removed from S. This step is repeated until stability or until S contains only 2 elements. The returned score is the variance found in the last step. This algorithm can be summarized as follows : bool stable = false S = {C i } i=1...n a = average(s) v = variance(s,a) do find the farest color C f S from a if dist(c f,a) > d then - S = S-{C f } - a = average(s) - v = variance(s,a) else stable = true while Card(S) 2 and stable = false score = v color = a If an object is viewed by most of the input cameras then it will be correctly displayed since this method rejects outliers colors, hence this method handles occlusions. Again, this method does not need a reference camera i.e. a camera near to the virtual camera and with similar properties. Thus we can place the virtual camera anywhere in front of the scene and it does not need to lie between the input cameras. Finally, our method avoids discontinuities which can happen when the virtual camera is moving and changes its reference camera. 4 Implementation and results Our plane-sweeping modelisation method has been implemented using OpenGL 1.5 and OpenGL Shading Language for the scoring stage. The texture coordinates are computed using projected textures from the camera matrices P i. We use multitexturing in order to get access to each texture during the scoring stage. The scores are stored in the []gl_fragdepth and the colors in the gl_fragcolor. Hence we let OpenGL select best scores with the z-test and update the color in the frame buffer. Most of the work is done by the graphic card and the CPU is free for others tasks. Figure 2 shows three calibrated images taken from three video cameras. Two pictures generated in real-time from this set of input images are shown in Figure 3. Despite the apparition of noise, we can see that the resulting image is close to the real model. The computation of a 325 260 final image using 30 depth samples along cam x s principal ray provide a 15 fps framerate on an Pentium4 2.8 GHz and an ATI Radeon 9800 Pro. This application is dedicated to virtual reality using stereo display. In that case, the humain brain removes noise errors during the fusion of the two images.

5 Conclusion We have presented in this article an adaptation of an image-based method using a small set of input images whereas other methods require a large set. The final image is quite realistic compare to the real model. Moreover, since the computation of an image does not depend on the previous generated images, our method allows the rendering of dynamic scenes for a low resolution final image. This method can be implemented for every consumer graphic harware that supports fragment shaders and therefore frees the CPU for other tasks. Since our scoring method takes into account all input images together and reject outliers, our method handle occlusions. Finally, our scoring method enhances robustness and implies fewer constraints on the position of the virtual camera, i.e. it does not need to lie between the input camera s area. We propose to extend our research on a stereoscopic adaptation of our method that compute the second view for low cost. Figure 2: Three input calibrated images. Figure 3: Two real-time images from three input cameras (20 fps with 30 planes). References [FWZ03] Andrew Fitzigibbon, Yonatan Wexler, and Andrew Zisserman. Image-based rendering using imagebased priors. 9th IEEE International Conference on Computer Vision, pages 1176 1183, 2003. [GSC96] J. Gortlerand R. Grzeszczuk, R. Szeliski, and M. F. Cohen. The lumigraph. SIGGRAPH, pages 43 54, 1996. [IHA02] M. Irani, T. Hassner, and P. Anandan. What does the scene look like from a scene point? Proc. ECCV, pages 883 897, 2002. [LH96] Marc Levoy and Pat Hanrahan. Light field rendering. SIGGRAPH, pages 31 42, 1996. [MB95] Leonard McMillan and Gary Bishop. Plenoptic modeling: An image-based rendering system. SIG- GRAPH, pages 39 46, 1995. [MBR + 00] Wojciech Matusik, Chris Buehler, Ramesh Raskar, Steven J. Gortler, and Leonard McMillan. Imagebased visual hulls. Proc ACM SIGGRAPH, pages 369 374, 2000. [PVG00] Marc Pollefeys, Maarten Vergauwen, and Luc Van Gool. Automatic 3d modeling from image sequences. proc. ISPRS congress, Amsterdam, pages 7 42, 2000.

[SD97] [SS02] [YWB02] Steven M. Seitz and Charles R. Dyer. Photorealistic scene reconstruction by voxel coloring. Proc. CVPR, pages 1067 1073, 1997. Daniel Scharstein and Richard Szeliski. A taxonomy and evaluation of dense two-frame stereo correspondence algorithms. IJCV, 47, pages 7 42, 2002. Ruigang Yang, Greg Welch, and Gary Bishop. Real-time consensus-based scene reconstruction using commodity graphics hardware. Proc. of Pacific Graphics, pages 225 234, 2002.