Real-Time Video- Based Modeling and Rendering of 3D Scenes
|
|
- Ralf James
- 6 years ago
- Views:
Transcription
1 Image-Based Modeling, Rendering, and Lighting Real-Time Video- Based Modeling and Rendering of 3D Scenes Takeshi Naemura Stanford University Junji Tago and Hiroshi Harashima University of Tokyo In research on 3D image communications and virtual reality, developing techniques for synthesizing arbitrary views has become an important technical issue. Given an object s structural model (such as a polygon or volume model), it s relatively easy to synthesize arbitrary views. Generating a structural model of an object, however, isn t necessarily easy. For this reason, research has been progressing on a technique called image-based modeling and rendering (IBMR) that avoids this problem. To date, researchers have performed studies on various IBMR techniques. (See the Related Work sidebar for more specific information.) Using densely arranged Our work targets 3D scenes in motion. In this article, we propose a cameras, we generate a method for view-dependent layered representation of 3D dynamic dynamic layered model of a scenes. Using densely arranged cameras, we ve developed a system 3D scene in real time, that can perform processing in real time from image pickup to interactive display, using video sequences suppressing visible errors in instead of static images, at 10 frames the synthetic virtual views. per second (frames/sec). In our system, images on layers are view dependent, and we update both the shape and image of each layer in real time. This lets us use the dynamic layers as the coarse structure of the dynamic 3D scenes, which improves the quality of the synthesized images. In this sense, our prototype system may be one of the first full real-time IBMR systems. Our experimental results show that this method is useful for interactive 3D rendering of real scenes. System configuration Our camera array system consists of 16 cameras and is versatile in arranging cameras in several configurations. We designed our input system to capture 16 video sequences simultaneously and to suppress individual differences between cameras. We use a real-time depth estimation system that can synthesize 30 depth maps per second from the input images. Versatile camera array system First, we constructed a versatile camera array system named Camera Array Pursuits for Plenoptic Acquisition (CAPPA). Figure 1a shows the modular camera head units we developed. By connecting several units, we can freely construct an array of NTSC color camera heads (Sony XC-333). For example, cameras can be densely arranged in a lattice (see Figure 1b) or connected in a row (see Figure 1c). Moreover, we can arrange cameras sparsely by inserting empty units between camera head units. We can also change lenses to capture light rays from several viewing angles. The distance between neighboring cameras is about 31 mm. With an empty unit between two camera head units, the distance is 62 mm approximately the 1 Versatile camera array system CAPPA, which consists of Sony XC-333 camera heads and (a) our original modular units. Cameras can be densely arranged in (b) a lattice or (c) a row. (a) (b) (c) 66 March/April /02/$ IEEE
2 Related Work A key advantage of IBMR is that using actual images makes it possible to synthesize real images even when estimating object structure is difficult. A system that makes absolutely no use of an object s structural elements can therefore be thought of as the most illustrative example of IBMR. 1 It has a close relationship with 3D display technology as methods of reproducing light rays. 2 Nevertheless, researchers have proposed systems that use an object s structural elements either in a partial or supplementary manner. These include techniques that interpolate between light-ray data using a rough structural model obtained from the object s silhouette 3 and that synthesize images for virtual views by surrounding the object with cameras and generating structural models of even higher accuracy. 4 Techniques exist at various levels, from ones using no structural characteristics of the object at all to conventional techniques that depend on such structural characteristics. An IBMR system can be roughly divided into two modules: One that performs data input and data generation from image capture to data interpolation and structure estimation. One that synthesizes desired images from the data obtained by the first module. The first module requires more processing time as complexity increases. It also requires some technique to make transmitting and storing data more efficient as the number of input images increases. The second module, on the other hand, can synthesize images in real-time interactively. Most research has targeted static space. In other words, an image is displayed only after capturing an object from various viewpoints and processing the data obtained, which takes time. Even in research targeting dynamic space in which scenes change with time, processing is applied only after multiview video input has been recorded. 4 However, some research on data input and generation modules that take real-time characteristics into account have appeared recently. 5-8 Naemura and Harashima 5 implemented a system that can simultaneously capture 4 4 multiview video sequences using five quad processors and a single video capturing device (Onyx2 DIVO) at the cost of resolution. They realized a real-time method of interactively synthesizing arbitrary views of dynamic 3D scenes from 4 4 camera array input. The method is a kind of ray-based representation of visual cues 1 that we ll extend in this article. Kunita et. al 6 developed a hardware system that adaptively extracts scan lines from videos captured by 12 cameras arranged in a row and composes them into an appropriate video sequence. However, while this approach can avoid the resolution problem in Naemura and Harashima, 5 it s limited to cameras arranged in a row. Ooi et. al 7 designed a smart image sensor specialized for image-based rendering. Setting an array of lenses in front of the sensor, a multiview image is projected onto the sensor. Then, appropriate pixels are selected with the method Naemura and Harashima 5 describe and a composed image is output directly from the sensor as a video sequence. Although this approach minimizes the processing delay, it still requires further improvements to enhance its resolution. All these approaches utilize densely arranged cameras to capture numerous light rays. In contrast, Matusik et. al 8 proposed a method that uses sparsely arranged cameras and generates visual hulls a better approximation of 3D dynamic scenes in real time. Alternatively, we use densely arranged cameras to generate a dynamic layered model of 3D scenes in real time, suppressing visible errors in the synthesized images. References 1. M. Levoy and P. Hanrahan, Light Field Rendering, Computer Graphics (Proc. Siggraph 96), ACM Press, New York, 1996, pp T. Naemura and H. Harashima, Ray-based Approach to Integrated 3D Visual Communication, Three- Dimensional Video and Display: Devices and Systems, SPIE Press, Bellingham, Wash., vol. CR76, 2000, pp S. Gortler et al., The Lumigraph, Computer Graphics (Proc. Siggraph 96), ACM Press, New York, 1996, pp T. Kanade, P. Rander, and P.J. Narayanan, Virtualized Reality: Constructing Virtual Worlds from Real Scenes, IEEE MultiMedia, vol. 4, no. 1, Jan. Mar. 1997, pp T. Naemura and H. Harashima, Real-Time Video-Based Rendering for Augmented Spatial Communication, Proc. Visual Comm. and Image Process (VCIP 99), vol. 3653, SPIE Press, Bellingham, Wash., 1999, pp Y. Kunita et al., Real-Time Rendering System of Moving Objects, Proc. IEEE Workshop Multiview Modeling and Analysis of Visual Scenes (MVIEW 99), IEEE CS Press, Los Alamitos, Calif., 1999, pp R. Ooi et al., Pixel Independent Random Access Image Sensor for Real Time Image-Based Rendering System, Proc. IEEE Int l Conf. Image Processing (ICIP 2001), IEEE Signal Process. Soc., Piscataway, N.J., 2001, Vol. II, pp W. Matusik et al., Image-Based Visual Hulls, Computer Graphics (Proc. Siggraph 2000), ACM Press, New York, 2000, pp IEEE Computer Graphics and Applications 67
3 Image-Based Modeling, Rendering, and Lighting 2 System configuration. Blue lines indicate a simultaneous capturing path, and red lines indicate a realtime depth estimation path. Quad processors Camera array DIVO Combined 16 views Onyx2 PC FZ930 Estimated depth map TCP/IP Viewpoint Interactive display puter as the number of cameras increases, which isn t conducive to real-time processing. To solve this problem, we adopted the system configuration in Figure 2. In Figure 2, the quad processor (Sony YS-Q430) is a piece of equipment that combines the video sequences from four cameras and outputs a video sequence divided into four screens (the blue lines in the figure). Video from the 16 cameras therefore consists of four-screen sequences obtained through the use of four quad-processor units. A fifth quad processor combines these four sequences so that the video from 16 cameras becomes a single 16-screen sequence. An SGI Onyx2 workstation (four 400-MHz R12000) with a DIVO (digital video option) captures and processes this single 16-screen sequence. In this regard, when connecting quad-processor units in an L-level cascade in this way, one video board can accommodate 4 L cameras worth of video in the computer. Here, however, the resolution of each video sequence drops, and system design must consequently consider an appropriate balance between resolution and the number of cameras. In this study, L = 2, which means that the horizontal resolution of each contracted image becomes about 720/4 = 180 pixels. 3 Effect of realtime adjustment. (a) Input from a fifth quad processor. (b) Adjusted 16- screen image. (a) Real-time adjustment Unless we can prepare extremely sophisticated hardware, however, there is always the possibility of a slight offset in each image s position. Typical reasons for this are differences between extrinsic camera parameters, such as camera position and direction; differences between intrinsic camera parameters, such as lens distortion and the relationship between the lens and charge-coupled device (CCD) surface; and individual differences in the way that images are divided in the quad processors. (b) distance between the left and right eyes which is generally regarded as a good condition for stereoscopic imaging. In this article, we focus on the camera configuration in Figure 1b, by which we can capture light rays at 4 4 sampling points on a 93 mm 93 mm square. Simultaneous capture Our system aims to process multiview video input obtained from multiple cameras simultaneously. One method of achieving this is to install in a computer as many video-capturing boards as the number of cameras. However, this approach increases the load on the com- Taking all these effects into account would require nonlinear processing. However, considering our emphasis on real-time characteristics, we perform the correction through translation operations. This isn t enough in some cases but quite effective within the limitation of real-time processing. We determine the amount of translation here for each image. In particular, the amounts of translation in the horizontal and vertical directions for the image captured by the camera C k (k = 1,, 16) are denoted by o x(k) and o y(k), respectively. o x(k) and o y(k) are characteristic parameters of the capturing system. We determine them only once after constructing the system. Figure 3a shows an example of a 16-screen combined video sequence. Here, we ve selected corresponding points manually and connected them with white lines. We can make the points line up correctly by determin- 68 March/April 2002
4 ing the appropriate amount of translation from the least squares method. Figure 3b shows the results of performing a translation on each image. The manual operations consist of setting the camera system s characteristic parameters. However, we only need to do these once after constructing the camera system. A A % 0% 50% 100% Object plane A 3 A 2.5 A 2 50% 0% 0% 100% 50% 0% 100% Real-time depth estimation We input nine of the 16 video sequences into the real-time depth estimation system Komatsu FZ930 ( Stereo/stereoj/index.htm [in Japanese]; see the red lines in Figure 2). FZ930 is a commercially available peripheral component interconnect (PCI) board that can estimate a depth map by employing the stereo matching method between the central image and the eight neighboring images. We use the set of eight neighboring images to overcome difficult occlusion problems by using a single neighboring image. FZ930 can provide 30 depth maps ( pixels, 8 bits) per second. We transfer this depth map to the Onyx2 workstation through a 100- Mbps Ethernet. The Onyx2 workstation interactively synthesizes and displays views that correspond to the user s viewpoint from the 16-color video sequences and an estimated depth-map sequence. Image synthesis basics Now let s move on to how we synthesize arbitrary views from the light field captured by the camera array (see Figure 4). We want to synthesize a view I v corresponding to the viewpoint P v from the images I i (i = 1, 2, 3) captured by cameras C i (i = 1, 2, 3), respectively. We can assume that the light ray P 1P v that passes through the point P 1 will arrive at P v as camera C 1 captures it. The problem is that we must interpolate the light rays that pass between cameras such as P 1.5P v and P 2.5P v in Figure 4. Blending images For the purpose of interpolation, we approximate the shape of the object by a plane. For example, a point A 1.5 on the object plane is the cross point of the object plane and the light ray P 1.5P v. This point A 1.5 is captured by both cameras C 1 and C 2 as pixel data denoted by I 1(A 1.5) and I 2(A 1.5). Here, we consider the case where we can use the intrinsic parameters of the cameras (C 1 and C 2) and the positions of P 1, P 2, and A 1.5 to read out the pixel value corresponding to A 1.5 from images I 1 and I 2. In this article, we use the average value of those pixels as the color data of the light ray P 1.5P v that is, I v(a 1.5). More generally, we calculate the color data of the light ray P 1 + ap v that passes a point P 1 + a (0 a 1) with this equation: I v(a 1 + a) = (1 a)i 1(A 1 + a) + ai 2(A 1 + a) P 1 P 1.5 C 1 C 2 P 2 P 2.5 Viewpoint P v C 3 Camera plane In other words, to synthesize a view I v at the viewpoint P v, we blend the two images I 1 and I 2. This blending effect suppresses the discontinuities between stitched views. Figure 4 shows how to blend more images. We use image I 2 captured by the camera C 2 between I v(a 1) and I v(a 3). This image is more effective than the other images (I 1 and I 3) between I v(a 1.5) and I v(a 2.5). More generally, we use the image I i between I v(a i 1) and I v(a i+1) but most effectively between I v(a i 0.5) and I v(a i+0.5). We can adopt the alpha-blending architecture to realize this bilinear interpolation in real time. Results Although our discussion so far has concentrated on the case where cameras are arranged in a row, we can easily extend this concept to cameras arranged in a lattice. Figure 5 (next page) shows the experimental results of this method. By moving the object plane to the position of the object closer to the camera (a toy pig), we can synthesize a view I v, as Figure 5b shows. We synthesized this image from portions of 4 4 multiview images by using the alpha-blending technique. In Figure 5a, the red squares denote the areas that correspond to those between I v(a i 0.5) and I v(a i+0.5) in our earlier discussion of the white squares between I v(a i 1)and I v(a i+1). We can see visible errors in the background object (a clock). Figures 5c and 5d show the result when we adjust the object plane to the clock, which blurs the toy pig. These results demonstrate a limitation of the plane approximation method. Our system takes into account that users can interactively move their viewpoints and the location and direction of the object plane. In other words, the focal plane is variable. 1 Processing for this interactive view synthesis of dynamic scenes was performed at 47 frames/sec. P 3 4 Blending several images to generate synthetic views. IEEE Computer Graphics and Applications 69
5 Image-Based Modeling, Rendering, and Lighting 5 Examples of the plane approximation method. (a) Input views used for synthesizing (b). (b) The toy pig in focus. (c) Input views used for synthesizing (d). (d) The clock in focus. (a) (b) (c) (d) This rate is greater than the 30 frames/sec of video input, which suggests the possibility of applying even more complex processing. Obviously, the single plane approximation of scene objects isn t good enough to render higher quality synthetic images. Dynamic layered representation To improve the quality of view synthesis, we adopt the depth map provided by the FZ930 system to better approximate the shape of the scene objects. Figure 6 illustrates the flowchart of this method. FZ930 estimates a depth map corresponding to the view indicated by the white square, which is the center of the nine views indicated by the white dashed square in the figure s input view. In the depth map, the brighter points indicate bigger depth values but mean that the object is closer to the camera array system. Because 8-bit depth values represent the depth map, it might be possible to approximate the shape by 256 planes (layers). From the viewpoint of real-time processing, however, we limit the number of layers to three in this article. Each layer is illustrated by red, blue, or green in Figure 6. View-dependent video layers Let s represent the red, blue, and green layers as L 1, L 2, and L 3, respectively. Each layer has its own depth value. D i(i = 1, 2, 3) denotes the depth value of layer L i, and D 1 < D 2 < D 3. (L 2 is farther from the camera array than L 3 but closer than L 1.) For each layer L i, we can synthesize an appropriate layer image T i by using the method in Figure 5. It s important that the layer image T i is view dependent in the sense that it varies according to the user s viewpoint. 2 Moreover, T i is a kind of video sequence that is updated periodically by the video-based rendering method. Figures 7a, 7b, and 7c show the results of synthesizing T i. In the layer image T 1 synthesized for the background layer L 1, we can see that background objects are in focus, but others (a toy cat and a toy pig) are blurry. In the middle layer L 2, the toy cat is in focus, but the toy pig is still blurry and the background has visible errors in T 2. In T 3 for L 3, all the objects except the toy pig have visible errors. Every focused part of T 1, T 2, and T 3 should be stitched together to synthesize a better quality view of the scene. For this purpose, we need a kind of mask for each layer to superimpose or blend the layer images T i. 70 March/April 2002
6 FZ930 Depth map Graph 1 Input 16 view Output (layer) Input (depth) User's viewpoint Layered alpha map 6 Flowchart of layered videobased rendering. Graph 2 L 1 L 2 T 1 α 1 (1 α 3 )(1 α 2 )α 1 T 1 Blending ratio D 1 D 2 D 3 Depth T 2 α 2 (1 α 3 )α 2 T 2 L 3 T 3 a 3 α 3 T 3 View-dependent layer images Alpha values for layer images Synthesized view α 3 T 3 + (1 α 3 ){α 2 T 2 + (1 α 2 )α 1 T 1 } Dynamic layer integration To synthesize arbitrary views of 3D scenes, a threelayer representation is obviously better than a single plane approximation. However, this might not be enough to describe complex scenes. To suppress the effect of this simplification, we adopt the alphablending technique. This concept is illustrated in Figure 6 s Graph 2. It is straightforward to render the object whose depth is close to D i simply by using the layer image T i. The problem is how to render such an object located between D 1 and D 2. For example, let s consider the case where the object is located at the depth value of D = a(d 1 + b) + (1 a)(d 2 b), where 0 a 1 and 0 b < (D 2 D 1)/2. We can render the view of this object I v by blending T 1 and T 2 according to this equation: (a) (c) (b) (d) 7 Result of dynamic layered representation. (a) Background layer image T 1, (b) a toy cat image layer T 2, (c) a toy pig image layer T 3, and (d) the result of blending the layers. I v = at 1 + (1 a) T 2 This approach isn t accurate but is a useful way of suppressing visible errors. Graph 2 illustrates the relationship between the depth of object and blending ratio of each layer. We can set the parameter b interactively while viewing the synthesized images. In this article, b = (D 2 D 1)/3. To implement this concept, we use the depth map gen- IEEE Computer Graphics and Applications 71
7 Image-Based Modeling, Rendering, and Lighting 8 The real scene captured for use in Figure 7. By comparing the images, we can see the relationship between scene objects and a camera array. 9 A plot of the frame rate as a function of the number of layers. Frames per second Number of layers erated by FZ930. We first synthesize filtered depth maps using the transfer function Graph 1 in Figure 6. We regard these maps as alpha maps for blending layers. Then, we translate and scale them so that they fit the layer images T i according to the relationship between the user s viewpoint and each layer s position. The translated alpha maps are denoted as α i in Figure 6. Note that α i isn t a static mask for each layer but dynamically updated. Finally, we render the synthetic view I v according to this blending equation: I v = α 3T 3 + (1 α 3){α 2T 2 + (1 α 2) α 1T 1} Results Figure 7d shows the final result. Most of the scene objects have better quality in Figure 7d than in Figures 7a, 7b, or 7c. Figure 8 shows the real scene captured and rendered in the experiment. We can see the relationship between scene objects and find that their size in Figure 7d is appropriate. The reason why the toy pig seems larger than the toy cat in the input 16 view in Figure 6 is that the camera array is close to the toy pig. The toy pig in Figure 7 isn t as large as in the input 16 view because we set the viewpoint behind the camera array as shown in Figure 4 during the rendering process. This indicates that the perspective produced by our system is consistent with the virtual viewpoint s position. Processing for this interactive view synthesis of dynamic scenes was performed at 10 frames/sec. This means that we achieved all these processes in real time: capturing 16 video sequences; estimating the depth map from nine of the 16 video sequences; synthesizing three view-dependent layer images T 1, T 2, and T 3 from the 16 video sequences; synthesizing layered alpha maps α 1, α 2, and α 3 from the estimated depth map; and blending the three layers. Users can move their viewpoint interactively to observe a 3D dynamic scene. Figure 9 illustrates the relationship between the number of layers and the processing speed (frames/sec). Even the 10-layer representation lets us observe the 3D dynamic scene interactively. However, we selected the three-layer representation for the experiment because we believe the real-time application should perform at more than 10 frames/sec. Future work One of the most important tasks of future work is to enhance the resolution of the synthetic images. For this purpose, it s possible to control the quad processors according to the user s viewpoint. This means that we should be able to adaptively input portions of the 16 video sequences into the computer. On the other hand, PC clusters might also be a practical solution for processing several views simultaneously. Moreover, we can utilize an optical system of integral photography and an HDTV camera to capture light fields. 3 We plan to keep investigating these approaches to find the most efficient way to capture and process multiple video sequences. Another important area for future work is to change the camera array configurations by using our versatile camera head units. The 4 4 camera array in this article is just one possible configuration. We plan to utilize the units to realize a panoramic stereo video system March/April 2002
8 Acknowledgments We thank Takuya Nitta for his support in system development. image-based rendering and augmented spatial communications. He has a BE, ME, and PhD in electronic engineering from the University of Tokyo. He is an IEEE member. References 1. A. Isaksen, L. McMillan, and S.J. Gortler, Dynamically Reparameterized Light Fields, Computer Graphics (Proc. Siggraph 2000), ACM Press, New York, 2000, pp P.E. Debevec, C.J. Taylor, and J. Malik, Modeling and Rendering Architecture from Photographs: A Hybrid Geometry- and Image-Based Approach, Computer Graphics (Proc. Siggraph 96), ACM Press, New York, 1996, pp T. Naemura, T. Yoshida, and H. Harashima, 3D Computer Graphics Based on Integral Photography, Optics Express, vol. 8, Feb. 2001, pp , opticsexpress.org/oearchive/source/30085.htm. 4. T. Naemura, M. Kaneko, and H. Harashima, Multi-User Immersive Stereo, Proc. IEEE Int l Conf. Image Process. (ICIP 98), IEEE CS Press, Los Alamitos, Calif., 1998, pp Takeshi Naemura is a visiting assistant professor in computer graphics at Stanford University and supported by the Japan Society for Promotion of Science (JSPS) Postdoctoral Fellowships for Research Abroad. His research interests include Junji Tago is a graduate student in the Department of Electronic Engineering at the University of Tokyo. His research interests include imagebased rendering and multimodal dialog systems. He has a BE in information and communication engineering from the University of Tokyo. Hiroshi Harashima is a professor in the Interfaculty Initiative in Information Studies at the University of Tokyo. His research interests include communication theory, image coding and processing, and human communication engineering. He has a BE, ME, and PhD in electrical engineering from the University of Tokyo. He is the president of the Virtual Reality Society of Japan. He is an IEEE member. Readers may contact Takeshi Naemura at Stanford Univ., Gates Computer Science Building 3B-390, Stanford, CA , naemura@graphics.stanford.edu Editorial Calendar January March: Advances in Multimedia Multimedia means different things to different communities. For researchers, it might be databases, search engines, or indexing tools, whereas content providers might be more concerned with streaming audio and video, compression techniques, and content distribution methods. This issue will offer various viewpoints, practices, evolving standards, and innovative projects in multimedia, covering current practice and research work. April June: Content-Based Multimedia Indexing and Retrieval Important research areas in multimedia indexing include audio, video, image, textual, and information retrieval. This special issue will cover the state of the art in multimedia indexing, especially image indexing, video indexing, user access and annotation, description of semantic content, and applications. July September: Multimedia R&D Multimedia systems and applications involve a broad range of topics, including hardware and software for media compression, media storage/transport, workstation support, data modeling, and abstractions to embedded multimedia in application programs. Even with this wide coverage, multimedia is still spreading its influence to nontraditional professional sections such as medicine, education, and the military. October December: Multimedia Trends Learn about the latest trends in multimedia and explore what researchers are developing for the next generation of multimedia applications. Find out what practitioners have learned in the field and what they plan to do next to improve the form and function of tomorrow s multimedia. IEEE Computer Graphics and Applications 73
Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images
Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images Abstract This paper presents a new method to generate and present arbitrarily
More informationRay-based approach to Integrated 3D Visual Communication
Ray-based approach to Integrated 3D Visual Communication Takeshi Naemura, Hiroshi Harashima 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656, JAPAN Dept. of Inform. & Commun. Eng., The Univ. of Tokyo ABSTRACT For
More informationBut, vision technology falls short. and so does graphics. Image Based Rendering. Ray. Constant radiance. time is fixed. 3D position 2D direction
Computer Graphics -based rendering Output Michael F. Cohen Microsoft Research Synthetic Camera Model Computer Vision Combined Output Output Model Real Scene Synthetic Camera Model Real Cameras Real Scene
More informationAcquisition and Visualization of Colored 3D Objects
Acquisition and Visualization of Colored 3D Objects Kari Pulli Stanford University Stanford, CA, U.S.A kapu@cs.stanford.edu Habib Abi-Rached, Tom Duchamp, Linda G. Shapiro and Werner Stuetzle University
More informationStereo pairs from linear morphing
Proc. of SPIE Vol. 3295, Stereoscopic Displays and Virtual Reality Systems V, ed. M T Bolas, S S Fisher, J O Merritt (Apr 1998) Copyright SPIE Stereo pairs from linear morphing David F. McAllister Multimedia
More informationConstruction of an Immersive Mixed Environment Using an Omnidirectional Stereo Image Sensor
Construction of an Immersive Mixed Environment Using an Omnidirectional Stereo Image Sensor Jun Shimamura, Naokazu Yokoya, Haruo Takemura and Kazumasa Yamazawa Graduate School of Information Science Nara
More informationLight Field Occlusion Removal
Light Field Occlusion Removal Shannon Kao Stanford University kaos@stanford.edu Figure 1: Occlusion removal pipeline. The input image (left) is part of a focal stack representing a light field. Each image
More informationReal-Time Video-Based Rendering from Multiple Cameras
Real-Time Video-Based Rendering from Multiple Cameras Vincent Nozick Hideo Saito Graduate School of Science and Technology, Keio University, Japan E-mail: {nozick,saito}@ozawa.ics.keio.ac.jp Abstract In
More informationLive Video Integration for High Presence Virtual World
Live Video Integration for High Presence Virtual World Tetsuro OGI, Toshio YAMADA Gifu MVL Research Center, TAO IML, The University of Tokyo 2-11-16, Yayoi, Bunkyo-ku, Tokyo 113-8656, Japan Michitaka HIROSE
More informationCapturing and View-Dependent Rendering of Billboard Models
Capturing and View-Dependent Rendering of Billboard Models Oliver Le, Anusheel Bhushan, Pablo Diaz-Gutierrez and M. Gopi Computer Graphics Lab University of California, Irvine Abstract. In this paper,
More informationImage-Based Rendering
Image-Based Rendering COS 526, Fall 2016 Thomas Funkhouser Acknowledgments: Dan Aliaga, Marc Levoy, Szymon Rusinkiewicz What is Image-Based Rendering? Definition 1: the use of photographic imagery to overcome
More informationStructure from Motion and Multi- view Geometry. Last lecture
Structure from Motion and Multi- view Geometry Topics in Image-Based Modeling and Rendering CSE291 J00 Lecture 5 Last lecture S. J. Gortler, R. Grzeszczuk, R. Szeliski,M. F. Cohen The Lumigraph, SIGGRAPH,
More informationHybrid Rendering for Collaborative, Immersive Virtual Environments
Hybrid Rendering for Collaborative, Immersive Virtual Environments Stephan Würmlin wuermlin@inf.ethz.ch Outline! Rendering techniques GBR, IBR and HR! From images to models! Novel view generation! Putting
More informationImage Base Rendering: An Introduction
Image Base Rendering: An Introduction Cliff Lindsay CS563 Spring 03, WPI 1. Introduction Up to this point, we have focused on showing 3D objects in the form of polygons. This is not the only approach to
More information3D Object Model Acquisition from Silhouettes
4th International Symposium on Computing and Multimedia Studies 1 3D Object Model Acquisition from Silhouettes Masaaki Iiyama Koh Kakusho Michihiko Minoh Academic Center for Computing and Media Studies
More informationDepth Estimation for View Synthesis in Multiview Video Coding
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Depth Estimation for View Synthesis in Multiview Video Coding Serdar Ince, Emin Martinian, Sehoon Yea, Anthony Vetro TR2007-025 June 2007 Abstract
More informationNatural Viewing 3D Display
We will introduce a new category of Collaboration Projects, which will highlight DoCoMo s joint research activities with universities and other companies. DoCoMo carries out R&D to build up mobile communication,
More informationModeling Light. Michal Havlik : Computational Photography Alexei Efros, CMU, Fall 2007
Modeling Light Michal Havlik 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 The Plenoptic Function Figure by Leonard McMillan Q: What is the set of all things that we can ever see? A: The
More information3-D Shape Reconstruction from Light Fields Using Voxel Back-Projection
3-D Shape Reconstruction from Light Fields Using Voxel Back-Projection Peter Eisert, Eckehard Steinbach, and Bernd Girod Telecommunications Laboratory, University of Erlangen-Nuremberg Cauerstrasse 7,
More informationOutdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera
Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera Tomokazu Sato, Masayuki Kanbara and Naokazu Yokoya Graduate School of Information Science, Nara Institute
More informationModeling, Combining, and Rendering Dynamic Real-World Events From Image Sequences
Modeling, Combining, and Rendering Dynamic Real-World Events From Image s Sundar Vedula, Peter Rander, Hideo Saito, and Takeo Kanade The Robotics Institute Carnegie Mellon University Abstract Virtualized
More information1-2 Feature-Based Image Mosaicing
MVA'98 IAPR Workshop on Machine Vision Applications, Nov. 17-19, 1998, Makuhari, Chibq Japan 1-2 Feature-Based Image Mosaicing Naoki Chiba, Hiroshi Kano, Minoru Higashihara, Masashi Yasuda, and Masato
More informationMore and More on Light Fields. Last Lecture
More and More on Light Fields Topics in Image-Based Modeling and Rendering CSE291 J00 Lecture 4 Last Lecture Re-review with emphasis on radiometry Mosaics & Quicktime VR The Plenoptic function The main
More informationA Stereo Vision-based Mixed Reality System with Natural Feature Point Tracking
A Stereo Vision-based Mixed Reality System with Natural Feature Point Tracking Masayuki Kanbara y, Hirofumi Fujii z, Haruo Takemura y and Naokazu Yokoya y ygraduate School of Information Science, Nara
More informationThe Plenoptic videos: Capturing, Rendering and Compression. Chan, SC; Ng, KT; Gan, ZF; Chan, KL; Shum, HY.
Title The Plenoptic videos: Capturing, Rendering and Compression Author(s) Chan, SC; Ng, KT; Gan, ZF; Chan, KL; Shum, HY Citation IEEE International Symposium on Circuits and Systems Proceedings, Vancouver,
More informationDD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication
DD2423 Image Analysis and Computer Vision IMAGE FORMATION Mårten Björkman Computational Vision and Active Perception School of Computer Science and Communication November 8, 2013 1 Image formation Goal:
More informationAccurate and Dense Wide-Baseline Stereo Matching Using SW-POC
Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Shuji Sakai, Koichi Ito, Takafumi Aoki Graduate School of Information Sciences, Tohoku University, Sendai, 980 8579, Japan Email: sakai@aoki.ecei.tohoku.ac.jp
More informationImage-Based Modeling and Rendering. Image-Based Modeling and Rendering. Final projects IBMR. What we have learnt so far. What IBMR is about
Image-Based Modeling and Rendering Image-Based Modeling and Rendering MIT EECS 6.837 Frédo Durand and Seth Teller 1 Some slides courtesy of Leonard McMillan, Wojciech Matusik, Byong Mok Oh, Max Chen 2
More informationCompression of 3D Objects with Multistage Color-Depth Panoramic Maps
Compression of 3D Objects with Multistage Color-Depth Panoramic Maps Chang-Ming Tsai 1,2, Wen-Yan Chang 1, Chu-Song Chen 1, Gregory Y. Tang 2 1 Institute of Information Science, Academia Sinica, Nankang
More informationStereo vision. Many slides adapted from Steve Seitz
Stereo vision Many slides adapted from Steve Seitz What is stereo vision? Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape What is
More informationAvatar Communication: Virtual Instructor in the Demonstration Exhibit
Avatar Communication: Virtual Instructor in the Demonstration Exhibit Tetsuro Ogi 1, 2, 3, Toshio Yamada 1, Takuro Kayahara 1, 2, Yuji Kurita 1 1 Telecommunications Advancement Organization of Japan 2
More informationCONVERSION OF FREE-VIEWPOINT 3D MULTI-VIEW VIDEO FOR STEREOSCOPIC DISPLAYS
CONVERSION OF FREE-VIEWPOINT 3D MULTI-VIEW VIDEO FOR STEREOSCOPIC DISPLAYS Luat Do 1, Svitlana Zinger 1, and Peter H. N. de With 1,2 1 Eindhoven University of Technology, P.O. Box 513, 5600 MB Eindhoven,
More informationAppearance-Based Virtual View Generation From Multicamera Videos Captured in the 3-D Room
IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 5, NO. 3, SEPTEMBER 2003 303 Appearance-Based Virtual View Generation From Multicamera Videos Captured in the 3-D Room Hideo Saito, Member, IEEE, Shigeyuki Baba, and
More informationDEVELOPMENT OF REAL TIME 3-D MEASUREMENT SYSTEM USING INTENSITY RATIO METHOD
DEVELOPMENT OF REAL TIME 3-D MEASUREMENT SYSTEM USING INTENSITY RATIO METHOD Takeo MIYASAKA and Kazuo ARAKI Graduate School of Computer and Cognitive Sciences, Chukyo University, Japan miyasaka@grad.sccs.chukto-u.ac.jp,
More informationImage Based Lighting with Near Light Sources
Image Based Lighting with Near Light Sources Shiho Furuya, Takayuki Itoh Graduate School of Humanitics and Sciences, Ochanomizu University E-mail: {shiho, itot}@itolab.is.ocha.ac.jp Abstract Recent some
More informationImage Based Lighting with Near Light Sources
Image Based Lighting with Near Light Sources Shiho Furuya, Takayuki Itoh Graduate School of Humanitics and Sciences, Ochanomizu University E-mail: {shiho, itot}@itolab.is.ocha.ac.jp Abstract Recent some
More informationView Generation for Free Viewpoint Video System
View Generation for Free Viewpoint Video System Gangyi JIANG 1, Liangzhong FAN 2, Mei YU 1, Feng Shao 1 1 Faculty of Information Science and Engineering, Ningbo University, Ningbo, 315211, China 2 Ningbo
More informationlecture 10 - depth from blur, binocular stereo
This lecture carries forward some of the topics from early in the course, namely defocus blur and binocular disparity. The main emphasis here will be on the information these cues carry about depth, rather
More informationEfficient View-Dependent Sampling of Visual Hulls
Efficient View-Dependent Sampling of Visual Hulls Wojciech Matusik Chris Buehler Leonard McMillan Computer Graphics Group MIT Laboratory for Computer Science Cambridge, MA 02141 Abstract In this paper
More informationComputational Photography
Computational Photography Photography and Imaging Michael S. Brown Brown - 1 Part 1 Overview Photography Preliminaries Traditional Film Imaging (Camera) Part 2 General Imaging 5D Plenoptic Function (McMillan)
More informationA Real-Time Distributed Light Field Camera
Thirteenth Eurographics Workshop on Rendering (2002) P. Debevec and S. Gibson (Editors) A Real-Time Distributed Light Field Camera Jason C. Yang Matthew Everett Chris Buehler Leonard McMillan Computer
More information5LSH0 Advanced Topics Video & Analysis
1 Multiview 3D video / Outline 2 Advanced Topics Multimedia Video (5LSH0), Module 02 3D Geometry, 3D Multiview Video Coding & Rendering Peter H.N. de With, Sveta Zinger & Y. Morvan ( p.h.n.de.with@tue.nl
More informationA SXGA 3D Display Processor with Reduced Rendering Data and Enhanced Precision
A SXGA 3D Display Processor with Reduced Rendering Data and Enhanced Precision Seok-Hoon Kim KAIST, Daejeon, Republic of Korea I. INTRODUCTION Recently, there has been tremendous progress in 3D graphics
More informationA New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction
A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction Jaemin Lee and Ergun Akleman Visualization Sciences Program Texas A&M University Abstract In this paper we present a practical
More informationDeformable Mesh Model for Complex Multi-Object 3D Motion Estimation from Multi-Viewpoint Video
Deformable Mesh Model for Complex Multi-Object 3D Motion Estimation from Multi-Viewpoint Video Shohei NOBUHARA Takashi MATSUYAMA Graduate School of Informatics, Kyoto University Sakyo, Kyoto, 606-8501,
More informationMultiple View Geometry
Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V
More informationView Synthesis for Multiview Video Compression
View Synthesis for Multiview Video Compression Emin Martinian, Alexander Behrens, Jun Xin, and Anthony Vetro email:{martinian,jxin,avetro}@merl.com, behrens@tnt.uni-hannover.de Mitsubishi Electric Research
More informationEnhanced Still 3D Integral Images Rendering Based on Multiprocessor Ray Tracing System
Journal of Image and Graphics, Volume 2, No.2, December 2014 Enhanced Still 3D Integral Images Rendering Based on Multiprocessor Ray Tracing System M. G. Eljdid Computer Sciences Department, Faculty of
More informationHIGH SPEED 3-D MEASUREMENT SYSTEM USING INCOHERENT LIGHT SOURCE FOR HUMAN PERFORMANCE ANALYSIS
HIGH SPEED 3-D MEASUREMENT SYSTEM USING INCOHERENT LIGHT SOURCE FOR HUMAN PERFORMANCE ANALYSIS Takeo MIYASAKA, Kazuhiro KURODA, Makoto HIROSE and Kazuo ARAKI School of Computer and Cognitive Sciences,
More informationModeling Light. Slides from Alexei A. Efros and others
Project 3 Results http://www.cs.brown.edu/courses/cs129/results/proj3/jcmace/ http://www.cs.brown.edu/courses/cs129/results/proj3/damoreno/ http://www.cs.brown.edu/courses/cs129/results/proj3/taox/ Stereo
More informationMorphable 3D-Mosaics: a Hybrid Framework for Photorealistic Walkthroughs of Large Natural Environments
Morphable 3D-Mosaics: a Hybrid Framework for Photorealistic Walkthroughs of Large Natural Environments Nikos Komodakis and Georgios Tziritas Computer Science Department, University of Crete E-mails: {komod,
More informationVirtual Reality Model of Koumokuten Generated from Measurement
Virtual Reality Model of Koumokuten Generated from Measurement Hiroki UNTEN Graduate School of Information Science and Technology The University of Tokyo unten@cvl.iis.u-tokyo.ac.jp Katsushi IKEUCHI Graduate
More information3D Editing System for Captured Real Scenes
3D Editing System for Captured Real Scenes Inwoo Ha, Yong Beom Lee and James D.K. Kim Samsung Advanced Institute of Technology, Youngin, South Korea E-mail: {iw.ha, leey, jamesdk.kim}@samsung.com Tel:
More informationEfficient Rendering of Glossy Reflection Using Graphics Hardware
Efficient Rendering of Glossy Reflection Using Graphics Hardware Yoshinori Dobashi Yuki Yamada Tsuyoshi Yamamoto Hokkaido University Kita-ku Kita 14, Nishi 9, Sapporo 060-0814, Japan Phone: +81.11.706.6530,
More informationIMAGE-BASED RENDERING TECHNIQUES FOR APPLICATION IN VIRTUAL ENVIRONMENTS
IMAGE-BASED RENDERING TECHNIQUES FOR APPLICATION IN VIRTUAL ENVIRONMENTS Xiaoyong Sun A Thesis submitted to the Faculty of Graduate and Postdoctoral Studies in partial fulfillment of the requirements for
More informationImage-Based Deformation of Objects in Real Scenes
Image-Based Deformation of Objects in Real Scenes Han-Vit Chung and In-Kwon Lee Dept. of Computer Science, Yonsei University sharpguy@cs.yonsei.ac.kr, iklee@yonsei.ac.kr Abstract. We present a new method
More informationJingyi Yu CISC 849. Department of Computer and Information Science
Digital Photography and Videos Jingyi Yu CISC 849 Light Fields, Lumigraph, and Image-based Rendering Pinhole Camera A camera captures a set of rays A pinhole camera captures a set of rays passing through
More informationFLY THROUGH VIEW VIDEO GENERATION OF SOCCER SCENE
FLY THROUGH VIEW VIDEO GENERATION OF SOCCER SCENE Naho INAMOTO and Hideo SAITO Keio University, Yokohama, Japan {nahotty,saito}@ozawa.ics.keio.ac.jp Abstract Recently there has been great deal of interest
More informationImage-based modeling (IBM) and image-based rendering (IBR)
Image-based modeling (IBM) and image-based rendering (IBR) CS 248 - Introduction to Computer Graphics Autumn quarter, 2005 Slides for December 8 lecture The graphics pipeline modeling animation rendering
More informationModeling Light. Michal Havlik
Modeling Light Michal Havlik 15-463: Computational Photography Alexei Efros, CMU, Spring 2010 What is light? Electromagnetic radiation (EMR) moving along rays in space R( ) is EMR, measured in units of
More informationShading of a computer-generated hologram by zone plate modulation
Shading of a computer-generated hologram by zone plate modulation Takayuki Kurihara * and Yasuhiro Takaki Institute of Engineering, Tokyo University of Agriculture and Technology, 2-24-16 Naka-cho, Koganei,Tokyo
More informationOcclusion Detection of Real Objects using Contour Based Stereo Matching
Occlusion Detection of Real Objects using Contour Based Stereo Matching Kenichi Hayashi, Hirokazu Kato, Shogo Nishida Graduate School of Engineering Science, Osaka University,1-3 Machikaneyama-cho, Toyonaka,
More informationA Survey of Light Source Detection Methods
A Survey of Light Source Detection Methods Nathan Funk University of Alberta Mini-Project for CMPUT 603 November 30, 2003 Abstract This paper provides an overview of the most prominent techniques for light
More informationView Synthesis for Multiview Video Compression
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com View Synthesis for Multiview Video Compression Emin Martinian, Alexander Behrens, Jun Xin, and Anthony Vetro TR2006-035 April 2006 Abstract
More informationDense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera
Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Tomokazu Satoy, Masayuki Kanbaray, Naokazu Yokoyay and Haruo Takemuraz ygraduate School of Information
More informationHybrid Textons: Modeling Surfaces with Reflectance and Geometry
Hybrid Textons: Modeling Surfaces with Reflectance and Geometry Jing Wang and Kristin J. Dana Electrical and Computer Engineering Department Rutgers University Piscataway, NJ, USA {jingwang,kdana}@caip.rutgers.edu
More informationSIMPLE ROOM SHAPE MODELING WITH SPARSE 3D POINT INFORMATION USING PHOTOGRAMMETRY AND APPLICATION SOFTWARE
SIMPLE ROOM SHAPE MODELING WITH SPARSE 3D POINT INFORMATION USING PHOTOGRAMMETRY AND APPLICATION SOFTWARE S. Hirose R&D Center, TOPCON CORPORATION, 75-1, Hasunuma-cho, Itabashi-ku, Tokyo, Japan Commission
More informationASIAGRAPH 2008 The Intermediate View Synthesis System For Soccer Broadcasts
ASIAGRAPH 2008 The Intermediate View Synthesis System For Soccer Broadcasts Songkran Jarusirisawad, Kunihiko Hayashi, Hideo Saito (Keio Univ.), Naho Inamoto (SGI Japan Ltd.), Tetsuya Kawamoto (Chukyo Television
More informationRe-live the Movie Matrix : From Harry Nyquist to Image-Based Rendering. Tsuhan Chen Carnegie Mellon University Pittsburgh, USA
Re-live the Movie Matrix : From Harry Nyquist to Image-Based Rendering Tsuhan Chen tsuhan@cmu.edu Carnegie Mellon University Pittsburgh, USA Some History IEEE Multimedia Signal Processing (MMSP) Technical
More informationTopics and things to know about them:
Practice Final CMSC 427 Distributed Tuesday, December 11, 2007 Review Session, Monday, December 17, 5:00pm, 4424 AV Williams Final: 10:30 AM Wednesday, December 19, 2007 General Guidelines: The final will
More informationConversion of free-viewpoint 3D multi-view video for stereoscopic displays Do, Q.L.; Zinger, S.; de With, P.H.N.
Conversion of free-viewpoint 3D multi-view video for stereoscopic displays Do, Q.L.; Zinger, S.; de With, P.H.N. Published in: Proceedings of the 2010 IEEE International Conference on Multimedia and Expo
More informationImage Transfer Methods. Satya Prakash Mallick Jan 28 th, 2003
Image Transfer Methods Satya Prakash Mallick Jan 28 th, 2003 Objective Given two or more images of the same scene, the objective is to synthesize a novel view of the scene from a view point where there
More informationShape as a Perturbation to Projective Mapping
Leonard McMillan and Gary Bishop Department of Computer Science University of North Carolina, Sitterson Hall, Chapel Hill, NC 27599 email: mcmillan@cs.unc.edu gb@cs.unc.edu 1.0 Introduction In the classical
More informationImage-Based Modeling and Rendering
Traditional Computer Graphics Image-Based Modeling and Rendering Thomas Funkhouser Princeton University COS 426 Guest Lecture Spring 2003 How would you model and render this scene? (Jensen) How about this
More informationMulti-view stereo. Many slides adapted from S. Seitz
Multi-view stereo Many slides adapted from S. Seitz Beyond two-view stereo The third eye can be used for verification Multiple-baseline stereo Pick a reference image, and slide the corresponding window
More informationContext based optimal shape coding
IEEE Signal Processing Society 1999 Workshop on Multimedia Signal Processing September 13-15, 1999, Copenhagen, Denmark Electronic Proceedings 1999 IEEE Context based optimal shape coding Gerry Melnikov,
More informationMR-Mirror: A Complex of Real and Virtual Mirrors
MR-Mirror: A Complex of Real and Virtual Mirrors Hideaki Sato 1, Itaru Kitahara 1, and Yuichi Ohta 1 1 Department of Intelligent Interaction Technologies, Graduate School of Systems and Information Engineering,
More informationMulti-View Imaging: Capturing and Rendering Interactive Environments
Multi-View Imaging: Capturing and Rendering Interactive Environments Cha Zhang Microsoft Research One Microsoft Way, Redmond WA 98052 Email: chazhang@microsoft.com Tsuhan Chen ECE, Carnegie Mellon University
More informationINTERACTIVE ENVIRONMENT FOR INTUITIVE UNDERSTANDING OF 4D DATA. M. Murata and S. Hashimoto Humanoid Robotics Institute, Waseda University, Japan
1 INTRODUCTION INTERACTIVE ENVIRONMENT FOR INTUITIVE UNDERSTANDING OF 4D DATA M. Murata and S. Hashimoto Humanoid Robotics Institute, Waseda University, Japan Abstract: We present a new virtual reality
More informationThe Light Field and Image-Based Rendering
Lecture 11: The Light Field and Image-Based Rendering Visual Computing Systems Demo (movie) Royal Palace: Madrid, Spain Image-based rendering (IBR) So far in course: rendering = synthesizing an image from
More informationAutomatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera
Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera Wolfgang Niem, Jochen Wingbermühle Universität Hannover Institut für Theoretische Nachrichtentechnik und Informationsverarbeitung
More informationDECODING COMPLEXITY CONSTRAINED RATE-DISTORTION OPTIMIZATION FOR THE COMPRESSION OF CONCENTRIC MOSAICS
DECODING COMPLEXITY CONSTRAINED RATE-DISTORTION OPTIMIZATION FOR THE COMPRESSION OF CONCENTRIC MOSAICS Eswar Kalyan Vutukuri, Ingo Bauerman and Eckehard Steinbach Institute of Communication Networks, Media
More informationLight Fields. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen
Light Fields Light Fields By Levoy and Hanrahan, SIGGRAPH 96 Representation for sampled plenoptic function stores data about visible light at various positions and directions Created from set of images
More informationMulti-View Image Coding in 3-D Space Based on 3-D Reconstruction
Multi-View Image Coding in 3-D Space Based on 3-D Reconstruction Yongying Gao and Hayder Radha Department of Electrical and Computer Engineering, Michigan State University, East Lansing, MI 48823 email:
More informationRigid Body Motion and Image Formation. Jana Kosecka, CS 482
Rigid Body Motion and Image Formation Jana Kosecka, CS 482 A free vector is defined by a pair of points : Coordinates of the vector : 1 3D Rotation of Points Euler angles Rotation Matrices in 3D 3 by 3
More informationCurved Projection Integral Imaging Using an Additional Large-Aperture Convex Lens for Viewing Angle Improvement
Curved Projection Integral Imaging Using an Additional Large-Aperture Convex Lens for Viewing Angle Improvement Joobong Hyun, Dong-Choon Hwang, Dong-Ha Shin, Byung-Goo Lee, and Eun-Soo Kim In this paper,
More informationVision-Based 3D Fingertip Interface for Spatial Interaction in 3D Integral Imaging System
International Conference on Complex, Intelligent and Software Intensive Systems Vision-Based 3D Fingertip Interface for Spatial Interaction in 3D Integral Imaging System Nam-Woo Kim, Dong-Hak Shin, Dong-Jin
More informationA Study of Medical Image Analysis System
Indian Journal of Science and Technology, Vol 8(25), DOI: 10.17485/ijst/2015/v8i25/80492, October 2015 ISSN (Print) : 0974-6846 ISSN (Online) : 0974-5645 A Study of Medical Image Analysis System Kim Tae-Eun
More informationPoint Cloud Streaming for 3D Avatar Communication
16 Point Cloud Streaming for 3D Avatar Communication Masaharu Kajitani, Shinichiro Takahashi and Masahiro Okuda Faculty of Environmental Engineering, The University of Kitakyushu Japan 1. Introduction
More informationVision Review: Image Formation. Course web page:
Vision Review: Image Formation Course web page: www.cis.udel.edu/~cer/arv September 10, 2002 Announcements Lecture on Thursday will be about Matlab; next Tuesday will be Image Processing The dates some
More informationPARALLEL VISUALIZATION OF LARGE-SCALE FINITE ELEMENT SOLUTIONS USING PC CLUSTER IN THE CABIN
Annual Report of ADVENTURE Project ADV-99-1 (1999) PARALLEL VISUALIZATION OF LARGE-SCALE FINITE ELEMENT SOLUTIONS Izuru. SHIRAI *1, Shinobu. YOSHIMURA *1, and Genki. YAGAWA *2 *1 School of Frontier Science,
More informationRate-distortion Optimized Streaming of Compressed Light Fields with Multiple Representations
Rate-distortion Optimized Streaming of Compressed Light Fields with Multiple Representations Prashant Ramanathan and Bernd Girod Department of Electrical Engineering Stanford University Stanford CA 945
More informationComputer Vision / Computer Graphics Collaboration for Model-based Imaging, Rendering, image Analysis and Graphical special Effects
Mirage 2003 Proceedings Computer Vision / Computer Graphics Collaboration for Model-based Imaging, Rendering, image Analysis and Graphical special Effects INRIA Rocquencourt, France, March, 10-11 2003
More informationPolyhedral Visual Hulls for Real-Time Rendering
Polyhedral Visual Hulls for Real-Time Rendering Wojciech Matusik Chris Buehler Leonard McMillan MIT Laboratory for Computer Science Abstract. We present new algorithms for creating and rendering visual
More information3D graphics, raster and colors CS312 Fall 2010
Computer Graphics 3D graphics, raster and colors CS312 Fall 2010 Shift in CG Application Markets 1989-2000 2000 1989 3D Graphics Object description 3D graphics model Visualization 2D projection that simulates
More informationCHAPTER 3. Single-view Geometry. 1. Consequences of Projection
CHAPTER 3 Single-view Geometry When we open an eye or take a photograph, we see only a flattened, two-dimensional projection of the physical underlying scene. The consequences are numerous and startling.
More informationReal-time Rendering System of Moving Objects
Real-time Rendering System o Moving Objects Yutaka Kunita Masahiko Inami Taro Maeda Susumu Tachi Department o Mathematical Engineering and Inormation Physics Graduate Shool o Engineering, The University
More informationPART A Three-Dimensional Measurement with iwitness
PART A Three-Dimensional Measurement with iwitness A1. The Basic Process The iwitness software system enables a user to convert two-dimensional (2D) coordinate (x,y) information of feature points on an
More informationExtensions of H.264/AVC for Multiview Video Compression
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Extensions of H.264/AVC for Multiview Video Compression Emin Martinian, Alexander Behrens, Jun Xin, Anthony Vetro, Huifang Sun TR2006-048 June
More informationPRE-PROCESSING OF HOLOSCOPIC 3D IMAGE FOR AUTOSTEREOSCOPIC 3D DISPLAYS
PRE-PROCESSING OF HOLOSCOPIC 3D IMAGE FOR AUTOSTEREOSCOPIC 3D DISPLAYS M.R Swash, A. Aggoun, O. Abdulfatah, B. Li, J. C. Fernández, E. Alazawi and E. Tsekleves School of Engineering and Design, Brunel
More information