Multi-projector-type immersive light field display

Similar documents
Self-calibration Three Dimensional Light Field Display Based on Scalable Multi-LCDs

Natural Viewing 3D Display

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

Reprint. from the Journal. of the SID

A new fractal algorithm to model discrete sequences

A three-step system calibration procedure with error compensation for 3D shape measurement

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN

Flexible Calibration of a Portable Structured Light System through Surface Plane

Research on QR Code Image Pre-processing Algorithm under Complex Background

Chapter 26 Geometrical Optics

dq dt I = Irradiance or Light Intensity is Flux Φ per area A (W/m 2 ) Φ =

Light and the Properties of Reflection & Refraction

Application of Geometry Rectification to Deformed Characters Recognition Liqun Wang1, a * and Honghui Fan2

Depth Estimation with a Plenoptic Camera

3D Autostereoscopic Display Image Generation Framework using Direct Light Field Rendering

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

3D Digitization of Human Foot Based on Computer Stereo Vision Combined with KINECT Sensor Hai-Qing YANG a,*, Li HE b, Geng-Xin GUO c and Yong-Jun XU d

PHYSICS 106. Assignment #10 Due by 10 pm Tuesday April 13, DISCUSSION SECTION: [ ] D1 W 9 am [ ] D2 W 10 am [ ] HS W 10 am

Dynamic three-dimensional sensing for specular surface with monoscopic fringe reflectometry

Efficient Path Finding Method Based Evaluation Function in Large Scene Online Games and Its Application

Pupil Localization Algorithm based on Hough Transform and Harris Corner Detection

Reduction of Blocking artifacts in Compressed Medical Images

Evaluating Measurement Error of a 3D Movable Body Scanner for Calibration

A DATA DRIVEN METHOD FOR FLAT ROOF BUILDING RECONSTRUCTION FROM LiDAR POINT CLOUDS

Coupling of surface roughness to the performance of computer-generated holograms

Curved Projection Integral Imaging Using an Additional Large-Aperture Convex Lens for Viewing Angle Improvement

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

Inventions on Three Dimensional GUI- A TRIZ based analysis

Omni Stereo Vision of Cooperative Mobile Robots

Visual Secret Sharing Scheme with Autostereogram*

Wide-viewing integral imaging system using polarizers and light barriers array

Efficient Rendering of Glossy Reflection Using Graphics Hardware

Uniform angular resolution integral imaging display with boundary folding mirrors

MEASUREMENT OF PERCEIVED SPATIAL RESOLUTION IN 3D LIGHT-FIELD DISPLAYS

2/26/2016. Chapter 23 Ray Optics. Chapter 23 Preview. Chapter 23 Preview

Physics for Scientists & Engineers 2

Optics. a- Before the beginning of the nineteenth century, light was considered to be a stream of particles.

CCNY Math Review Chapters 5 and 6: Trigonometric functions and graphs

CS 563 Advanced Topics in Computer Graphics Stereoscopy. by Sam Song

Three-dimensional directional display and pickup

Vision-Based 3D Fingertip Interface for Spatial Interaction in 3D Integral Imaging System

Enhanced two-frequency phase-shifting method

Panoramic Image Stitching

An Image Based 3D Reconstruction System for Large Indoor Scenes

Pattern Feature Detection for Camera Calibration Using Circular Sample

Kiangsu-Chekiang College (Shatin)

Jim Lambers MAT 169 Fall Semester Lecture 33 Notes

LED holographic imaging by spatial-domain diffraction computation of. textured models

Multimedia Technology CHAPTER 4. Video and Animation

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision

Optimizing Monocular Cues for Depth Estimation from Indoor Images

Experiment 9. Law of reflection and refraction of light

Study on Improving the Quality of Reconstructed NURBS Surfaces

The Law of Reflection

ÉCOLE POLYTECHNIQUE DE MONTRÉAL

Depth-Layer-Based Patient Motion Compensation for the Overlay of 3D Volumes onto X-Ray Sequences

Comparative Study of ROI Extraction of Palmprint

A Semi-Automatic 2D-to-3D Video Conversion with Adaptive Key-Frame Selection

Calculation on diffraction aperture of cube corner retroreflector

Index of Refraction and Total Internal Reflection

Optimizing the Deblocking Algorithm for. H.264 Decoder Implementation

Optics INTRODUCTION DISCUSSION OF PRINCIPLES. Reflection by a Plane Mirror


Journal of Chemical and Pharmaceutical Research, 2015, 7(3): Research Article

Operation Trajectory Control of Industrial Robots Based on Motion Simulation

Recap from Previous Lecture

!!!"#$%!&'()*&+,'-%%./01"&', Tokihiko Akita. AISIN SEIKI Co., Ltd. Parking Space Detection with Motion Stereo Camera applying Viterbi algorithm

Perceptual Quality Improvement of Stereoscopic Images

4. Refraction. glass, air, Perspex and water.

Count Vehicle Over Region of Interest Using Euclidean Distance

Qiqihar University, China *Corresponding author. Keywords: Highway tunnel, Variant monitoring, Circle fit, Digital speckle.

SOME stereo image-matching methods require a user-selected

Robust color segmentation algorithms in illumination variation conditions

Lecture 15: Shading-I. CITS3003 Graphics & Animation

AIM To determine the frequency of alternating current using a sonometer and an electromagnet.

Projection simulator to support design development of spherical immersive display

Design of Miniaturized Optical Image Stabilization and Autofocusing Camera Module for Cellphones

Spatial and Angular Resolution Measurement of a Tensor Light Field Display

LIGHT STRIPE PROJECTION-BASED PEDESTRIAN DETECTION DURING AUTOMATIC PARKING OPERATION

Virtual Interaction System Based on Optical Capture

Numerical investigation on the viewing angle of a lenticular three-dimensional display with a triplet lens array

Coherent digital demodulation of single-camera N-projections for 3D-object shape measurement: Co-phased profilometry

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

Quaternion-based color difference measure for removing impulse noise in color images

Geometry of Multiple views

Multiple View Geometry

Experiment 3: Reflection

For each question, indicate whether the statement is true or false by circling T or F, respectively.

Extracting Sound Information from High-speed Video Using Three-dimensional Shape Measurement Method

Capture and Displays CS 211A

Supplementary Information. Design of Hierarchical Structures for Synchronized Deformations

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Computational Photography

specular diffuse reflection.

Midterm Exam CS 184: Foundations of Computer Graphics page 1 of 11

Grasping Known Objects with Aldebaran Nao

The image is virtual and erect. When a mirror is rotated through a certain angle, the reflected ray is rotated through twice this angle.

Recognition of Human Body Movements Trajectory Based on the Three-dimensional Depth Data

NORTHERN ILLINOIS UNIVERSITY PHYSICS DEPARTMENT. Physics 211 E&M and Quantum Physics Spring Lab #7: Reflection & Refraction

Measurement of 3D Foot Shape Deformation in Motion

Transcription:

Multi-projector-type immersive light field display Qing Zhong ( é) 1, Beishi Chen (í ì) 1, Haifeng Li (Ó ô) 1, Xu Liu ( Ê) 1, Jun Xia ( ) 2, Baoping Wang ( ) 2, and Haisong Xu (Å Ø) 1 1 State Key Laboratory of Modern Optical Instrumentation, Zhejiang University, Hangzhou 310027, China 2 Southeast University, Nanjing 210096, China Corresponding author: lihaifeng@zju.edu.cn Received March 31, 2014; accepted April 4, 2014; posted online May 28, 2014 A light field three-dimensional (3D) display with multi-projectors and a concave screen is proposed. The system sets the viewing area at the center of the concave screen, making viewers enter the center of the system to watch 3D scene around them. The surrounded 3D scene provides viewers a feast of enhanced immersive experience. The light field principle, rendering algorithm, selection of viewing area and experimental results are discussed in the letter, showing the potential of being an all-around-type immersive 3D display by employing more projectors. OCIS codes: 110.0110, 120.2040. doi: 10.3788/COL201412.060009. 1. Introduction Light field displays have attracted numerous attentions since it was proposed, as it can provide real threedimensional (3D) images with smooth parallax. Parts of light field displays intend to set up a square viewing region for 3D scenes just like conventional two-dimensional (2D) displays [1 3]. Others try to build a certain central space for viewers to watch around [4 6]. An immersive 3D display that reconstructs 3D scene for things surrounding viewers will definitely increase the watching experience, as humans are living in 3D space seeing 3D scene all around everyday. Immersive displays with scene surrounded are widely used in conventional 2D displays to build more natural and vivid perception for viewers [7,8]. However, few 3D displays successfully reconstruct surrounded scene for immersive experience. In previous work, we have proposed several displays based on light field reconstruction principle [5,6,8]. The light field reconstruction algorithm has been analyzed [7]. Performance of the diffuser screen and arrangement of the projectors have been discussed [8]. Conventional multi-projector-type light field display with viewing position outside the system, suffers the problem that the projection angle of the projectors limits the viewing size of the system, leading to truncation to the whole reconstructed image. To widen the viewing size, a greater projection angle is necessary, which will increase the difficulty of manufacturing. In this letter, a novel immersive 3D display system using light field reconstruction principle is proposed. A concave screen is employed to set up a space in the center of the system for viewers to enter. Projectors are configured around the screen with projecting images to the center of the screen. The viewing area is designed inside the screen and with its viewing field depending on the configuration of projectors. Inside the designed viewing area, all the employed projectors can be seen at the same time without moving the viewing position. The jointed observing created inside the concave screen makes it possible to create 3D scene around the viewers, demonstrating an immersive experience with 3D perception. 2. System Overview A. System Structure The system, illustrated in Fig. 1, consists of multiple projectors and an arc diffuser screen. The projectors are distributed along a circle that is homocentric with the screen, so that panoramic view can be attained by Fig. 1. Illustration of system structure. 1671-7694/2014/060009(5) 060009-1 c 2014 Chinese Optics Letters

adding more projectors to complete the circle. The common viewing area is inside the screen. People stand near the central area to enjoy the 3D images reconstructed by the projectors and the screen, achieving an immersive feeling. The light field reconstruction algorithm has been introduced in many other papers [1,5,7,9]. In the certain system with two different spatial points A and B for instance, shown as Fig. 1, they are projected by different projectors to provide different depth perception. Projector P 1 and P 2 offer the binocular images of point A while P 2 and P 3 offer that of point B. To a single projector, its light rays projected pointing to different directions are used to reconstruct objective light emitted for different directions for observation, thus reconstructing the light field for some certain position and direction. The system offers only horizontal parallax. Therefore, a special diffuser screen is employed with light rays diffused a small angle in horizontal direction while a large angle in vertical direction. Viewers will see the aperture of the projector diffused to a vertical stripe image though the screen. When the apertures of the projectors are dense enough and the horizontal diffusing angle of the screen is appropriate, the stripe images will join together to form a complete image. The image changes at different viewing position, leading to binocular parallax. B. Rendering Algorithm As what in above subsection discussed, the diffuser screen diffuses different angle in horizontal and vertical directions, which means that different algorithms shall be applied in the two directions. For the horizontal diffusing angle is used to widen the pupils to join the stripes, it is equivalent to employing wider pupils and a diffuser without horizontal diffusing angle. Detail about the diffuser screen has been discussed in previous work [8]. The algorithm then can be discussed in two directions. Here, the rendering relationship for this system is shown in Fig. 2. From the relationship, the algorithm for single projector can be described by x = xr P z + R P y = R [ P (y hv )(R V R ) ] R + R P R V z + h V h P, (1) + (R + R P )h P z = z For other projectors with interval angle θ, it is equivalent to rotate the spatial point with angle θ, thus the new point A = AR(θ), where R(θ) is the inverse rotation matrix along z-axis. All the projectors are applied the same algorithm as the system is centrosymmetric. Detail about the similar deduction is shown in reference Ref. [7]. 3. Display Area For an immersive system, it is important that when viewers enter the viewing area, they can see scene around them as much as possible. As the projection angle of projectors is limited, meaning that the viewing area are limited. The screen shall be set around the viewing area to ensure all the scene provided is seen inside the screen. The red circle in Fig. 3 shows the common imaging area of the multi-projector system. The common area that light emitted from all the projectors can reach is a circle that is tangent to the boundaries of the projection, with its radius: R = D sinω. (2) Where D is the distance from the exit pupil of the projector to the center of the system, ω is the half of projection angle. It is significant that viewing area is truncated if viewers are watching images outside the common display area. As Fig. 3 shows, for certain viewing distance L, which calculated from the observing position to the center of the system, the viewing angle θ is determined by 1 D sin ω θ = 2 sin L. (3) The two edge-projectors viewers can see, called P L and P R for instance, are determined by the viewing distance L, shown as the blue and the green characters in Fig. 3. Generally, if a convex screen is set, or a concave screen is set but the system size is not large enough for viewers to get close enough, the viewing position is out of the common area. So that viewers can only see parts of all the projectors. However, if the system size is large enough for viewers to enter the public area, meaning Fig. 2. Rendering Algorithm for the system. Fig. 3. Illustration of the public viewing area. 060009-2

that L < D sin ω, all the projectors will be seen at the same time, shown as the red character in Fig. 3. The position L = R = D sinω yielding θ = 180, which is at the edge of the common viewing area, is the critical position. In this case, viewers can see an image with large view angle, even panoramic view field if the projectors configuration is able to complete a circle. By setting the screen just outside the public viewing area, viewers are limited inside the viewing area to ensure the largest viewing angle. 4. Experimental Results A. Prototype A prototype of the system is set up using 90 projectors. Specifications of the prototype are shown in Table 1. A snapshot of the prototype is shown in Fig. 4. The white grid displayed in Fig. 4 is approximate 7.0 4.5 (cm), and the blue lines are at the middle of each projector s picture. Viewers can see a complete image with size at least 0.8 0.4 (m). With 90 projectors applied, the prototype can offer an image with 90 viewing angle. When viewers are standing near the center of the system, distance from screen to eyes is 40 cm. A concave screen will make viewers feel themselves surrounded with 3D scene, leading to a strong immersive experience. B. Public Viewing Area Fig. 5. Experimental scheme for showing public viewing area. (a) shows positions chosen for the experiment, and (b) distinguishes different projectors by different colors. As Fig. 5 shows, 8 edge positions are chosen to show the public viewing area of the prototype, where A, B, C and D shows positions inside the area, while A, B, Table 1. Specifications of the Prototype Number of Projectors 90 Resolution of Single Projector (H V) 800 600 Projection angle (H V) 13 11 Angle of Adjacent Projectors Distance from Projectors to the Center of System Radius of Diffuser Screen Diffusing Angle (H V) Fig. 4. A picture of prototype. 1 1.6 m 0.4 m 1 60 Fig. 6. Illustration of viewing positions inside or outside the public viewing area. In each sub-figure, the picture upper is taken inside the public viewing area while the lower is taken outside. C, and D shows positions outside. 60 projectors in the center are chosen and project 3 different colors to distinguish each other, where blue is in the leftmost and the red is in the rightmost. Figure 6 shows the trend that edge projectors can be seen at the positions inside, while some of them start to disappeared at the positions outside. It shows that viewers can see all the projectors inside the public viewing area in any direction. C. Depth Range An experiment employing a 3D model of a small ball and the SSIM algorithm [10] is proposed to help figure out the depth range that the prototype reconstructs. The ball is put at different distances from the screen position then captured by a camera set at the center of the system. The images captured then is compared with the original image of the model (Fig. 7(a)) and the comparison is presented by SSIM index. We can see that the images that the ball is at the screen position (Fig. 7(b)) is more 060009-3

similar to the original one than which further from the camera (Fig. 7(c))or nearer to the camera (Fig. 7(d)). Finally a graph indicates the relationship of the distance of the ball from the screen position and the SSIM index is shown in Fig. 8. In Fig. 8, the x-axis indicates the distance from the screen position, where the positive direction directs to the camera, or the viewing position. The y-axis is the SSIM index, which the higher value indicates a more similar image pair. We can see the model put near the screen positions have the higher similarity, while the similarity falls when the model is away from the screen. What s more, the similarity falls even faster when the model is getting close to the viewing position than getting away from it. Six people are invited to point out the tolerance of the depth range. The experiment shows an average far distance and near distance of 120 and 70 mm respectively. It indicates that a threshold SSIM index=0.5 can be drew in Fig. 8, showing a total depth range approximately 180 to 200 mm. D. 3D Scene Figure 9 shows 3 different viewing positions of the same model. From the occlusion shown in the three pictures we can see that the prototype can display a large viewing angle. Fig. 7. Illustration of images of the (a) the original model, and the reconstructed model putting (b) at the screen position, (c) further from the camera, and (d) nearer to the camera. Fig. 8. SSIM index for the original image and images of the same object from different distances. Fig. 9. model. Three different viewing positions of the same 3D The experimental result shows that the prototype with 90 projectors can reconstruct 3D images with large viewing field and viewing angle. All the projectors can be seen at the same time in the public viewing area, providing a feast of immersive experience. Motion parallax is achieve and occlusion relationship is accurate. 5. Conclusion Compared to the previous work, here we achieve 3D immersive visual experience with employing higher-quality projectors and a concave designed diffuser screen. 90 projectors are employed with adjacent angle 1 to offer a viewing field of 90. The concave diffuser screen with radius 0.4 m makes common viewing area set inside the screen, which acts key role on enabling viewers perceiving whole reconstructed 3D scene around them at same time. The prototype successfully reconstructed 3D scene with large viewing angle and motion parallax, with a depth range about 180 to 200 mm. Stronger immersive experience will be achieved by adding more projectors to widen the viewing angle and perfect the display quality. 6. Acknowledgement This work was supported by the National High Technology Research and Development Program of China (Nos. 2012AA011902 and 2012AA03A302), the National Natural Science Foundation of China (No. 61177015), the National Basic Research Program of China (973 Program) (No. 2013CB328802), and the Fundamental Research Funds for the Central Universities of China (No. 2012XZZX013). Great thanks to the State Key Lab of Modern Optical Instrumentation, Zhejiang University, for providing the necessary support. References 1. J. A. Iglesias Guitián, E. Gobbetti, and F. Marton, The Visual Computer 26, 1037 (2010). 2. J. Hong, S.-W. Min, and B. Lee, Appl. Opt. 51, 4201 (2012). 060009-4

3. G. Wetzstein, D. Lanman, W. Heidrich, and R. Raskar, ACM Transactions on Graphics 30, 95 (2011). 4. A. Jones, I. McDowall, H. Yamada, M. Bolas, and P. Debevec, ACM Transactions on Graphics (TOG) 26, 40 (2007). 5. X. Xia, X. Liu, H. Li, Z. Zheng, H. Wang, Y. Peng, and W. Shen, Opt. Express 21, 11237 (2013). 6. Y. Peng, H. Li, R. Wang, Q. Zhong, X. Han, Z. Cao, and X. Liu, Journal of the Society for Information Display 20, 653 (2012). 7. Q. Zhong, Y. Peng, H. Li, C. Su, W. Shen, and X. Liu, Appl. Opt. 52, 4419 (2013). 8. Y.-F. Peng, H.-F. Li, Q. Zhong, X.-X. Xia, and X. Liu, Opt. Eng. 52, 017402 (2013). 9. M. Levoy and P. Hanrahan, (1996). 10. W. Zhou, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, Image Processing, IEEE Trans. on Image Processing 13, 600 (2004). 060009-5