Extended Fractional View Integral Photography Using Slanted Orthogonal Lenticular Lenses

Similar documents
Real-time Integral Photography Holographic Pyramid using a Game Engine

3.3 Implementation of a Lenticular 3D Display

3D Autostereoscopic Display Image Generation Framework using Direct Light Field Rendering

Reprint. from the Journal. of the SID

A Fast Image Multiplexing Method Robust to Viewer s Position and Lens Misalignment in Lenticular 3D Displays

Mobile 3D Display Technology to Realize Natural 3D Images

Natural Viewing 3D Display

Depth-fused display with improved viewing characteristics

PRE-PROCESSING OF HOLOSCOPIC 3D IMAGE FOR AUTOSTEREOSCOPIC 3D DISPLAYS

A SXGA 3D Display Processor with Reduced Rendering Data and Enhanced Precision

FEATURE. Measurement of Static Convergence and Accommodation Responses to Images of Integral Photography

Vision-Based 3D Fingertip Interface for Spatial Interaction in 3D Integral Imaging System

Color moiré pattern simulation and analysis in three-dimensional integral imaging for finding the moiré-reduced tilted angle of a lens array

Department of Photonics, NCTU, Hsinchu 300, Taiwan. Applied Electromagnetic Res. Inst., NICT, Koganei, Tokyo, Japan

Reduced Dual-Mode Mobile 3D Display Using Crosstalk

Directional Backlights for Time-multiplexed 3D Displays. As the advancement of image display from monochromic to color, each

A Liver Surgery Simulator Using Full HD Autostereoscopic Displays

Digital holographic display with two-dimensional and threedimensional convertible feature by high speed switchable diffuser

Three-dimensional integral imaging for orthoscopic real image reconstruction

Prospective Novel 3D Display Technology Development

Light and Lenses Notes

Multi-View Omni-Directional Imaging

Fast Response Fresnel Liquid Crystal Lens for 2D/3D Autostereoscopic Display

SID Display Week Yasuhiro Takaki. Institute of Symbiotic and Technology. Tokyo Univ. of Agri. & Tech.

Shading of a computer-generated hologram by zone plate modulation

(b) (a) 394 H. Liao et al. Fig.1 System configuration

Numerical investigation on the viewing angle of a lenticular three-dimensional display with a triplet lens array

Curved Projection Integral Imaging Using an Additional Large-Aperture Convex Lens for Viewing Angle Improvement

RV - AULA 07 - PSI3502/2018. Displays

PHYS 219 General Physics: Electricity, Light and Modern Physics

Holojackets. Semester Update: Fall 2017 December 4, 2017

This paper is published in the open archive of Mid Sweden University DIVA by permission of the publisher

CGH calculation with the ray tracing method for the Fourier transform optical system

MEASUREMENT OF PERCEIVED SPATIAL RESOLUTION IN 3D LIGHT-FIELD DISPLAYS

Head Mounted Display Optics I!

Friday, 22 August 14. Lenticular prints

Mahdi Amiri. May Sharif University of Technology

High-resistance liquid-crystal lens array for rotatable 2D/3D autostereoscopic display

Crosstalk in multiview 3-D images

Omni-directional Flat Tabletop Display Which Enables to View from Any Direction

Computational Photography

Efficient Rendering of Glossy Reflection Using Graphics Hardware

Three-dimensional directional display and pickup

An Introduction to 3D Computer Graphics, Stereoscopic Image, and Animation in OpenGL and C/C++ Fore June

Specialized Depth Extraction for Live Soccer Video

Enhanced Techniques 3D Integral Images Video Computer Generated

Self-calibration Three Dimensional Light Field Display Based on Scalable Multi-LCDs

Three dimensional Binocular Holographic Display Using Liquid Crystal Shutter

Invited Paper. Nukui-Kitamachi, Koganei, Tokyo, , Japan ABSTRACT 1. INTRODUCTION

Multi-View Isosurface Ray-casting

Victor S. Grinberg Gregg W. Podnar M. W. Siegel

Depth extraction from unidirectional integral image using a modified multi-baseline technique

Lecture 17: Recursive Ray Tracing. Where is the way where light dwelleth? Job 38:19

Challenge of 3D LCD Displays

Wide-viewing integral imaging system using polarizers and light barriers array

Live Video Integration for High Presence Virtual World

Rectification of distorted elemental image array using four markers in three-dimensional integral imaging

Convergence Point Adjustment Methods for Minimizing Visual Discomfort Due to a Stereoscopic Camera

Optical characterization of auto-stereoscopic 3D displays: interest of the resolution and comparison to human eye properties

Speed of light E Introduction

Announcements. Written Assignment 2 out (due March 8) Computer Graphics

Generating 5D Light Fields in Scattering Media for Representing 3D Images

Devices displaying 3D image. RNDr. Róbert Bohdal, PhD.

The image is virtual and erect. When a mirror is rotated through a certain angle, the reflected ray is rotated through twice this angle.

An Auto-Stereoscopic VRML Viewer for 3D Data on the World Wide Web

Enhanced Still 3D Integral Images Rendering Based on Multiprocessor Ray Tracing System

Head Mounted Display for Mixed Reality using Holographic Optical Elements

A SXGA 3D Display Processor with Reduced Rendering Data and Enhanced Precision. Seok-Hoon Kim MVLSI Lab., KAIST

Augmenting Reality with Projected Interactive Displays

Enhanced Reconstruction of 3D Shape and Texture from. Integral Photography Images

Interaxial Distance and Convergence Control for Efficient Stereoscopic Shooting using Horizontal Moving 3D Camera Rig

Prof. Feng Liu. Spring /27/2014

Multiview Generation for 3D Digital Signage

At the interface between two materials, where light can be reflected or refracted. Within a material, where the light can be scattered or absorbed.

Measurements using three-dimensional product imaging

PHYSICS 106. Assignment #10 Due by 10 pm Tuesday April 13, DISCUSSION SECTION: [ ] D1 W 9 am [ ] D2 W 10 am [ ] HS W 10 am

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

Research Article An Innovative Direct-Interaction-Enabled Augmented-Reality 3D System

Dispersion (23.5) Neil Alberding (SFU Physics) Physics 121: Optics, Electricity & Magnetism Spring / 17

Chapter 26 Geometrical Optics

A Representation of Human Visual Acuity on an Immersive Projection Display

Stereoscopic Presentations Taking the Difficulty out of 3D

Effect of fundamental depth resolution and cardboard effect to perceived depth resolution on multi-view display

I have a meeting with Peter Lee and Bob Cosgrove on Wednesday to discuss the future of the cluster. Computer Graphics

INTERACTIVE ENVIRONMENT FOR INTUITIVE UNDERSTANDING OF 4D DATA. M. Murata and S. Hashimoto Humanoid Robotics Institute, Waseda University, Japan

Algebra Based Physics

Frequency domain depth filtering of integral imaging

QUALITY MEASUREMENTS OF 3D LIGHT-FIELD DISPLAYS

F..\ Compression of IP Images for Autostereoscopic 3D Imaging Applications

Depth Estimation with a Plenoptic Camera

GLASSES FREE AUTO STEREOSCOPIC 3D GREETING COMPANY PRODUCT TECHNOLOGY

Volumetric Hyper Reality: A Computer Graphics Holy Grail for the 21st Century? Gavin Miller Apple Computer, Inc.

Experiment 6. Snell s Law. Use Snell s Law to determine the index of refraction of Lucite.

PHYS1004 Problem Sheet - Optics - with Solutions

Real-time Multiview Autostereoscopic Image Synthesis System

A new autostereoscopic display technology: The SynthaGram

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

An Abstraction Technique for Producing 3D Visual Contents

Victor S. Grinberg M. W. Siegel. Robotics Institute, School of Computer Science, Carnegie Mellon University 5000 Forbes Ave., Pittsburgh, PA, 15213

Depth Estimation for View Synthesis in Multiview Video Coding

Transcription:

Proceedings of the 2 nd World Congress on Electrical Engineering and Computer Systems and Science (EECSS'16) Budapest, Hungary August 16 17, 2016 Paper No. MHCI 112 DOI: 10.11159/mhci16.112 Extended Fractional View Integral Photography Using Slanted Orthogonal Lenticular Lenses Kazuhisa Yanaka, Takuya Nomura, Toshiaki Yamanouchi Kanagawa Institute of Technology 1030 Shimo-ogino, Atsugi-shi, Kanagawa-ken, 243-0292 Japan yanaka@ic.kanagawa-it.ac.jp; takuya.n.kns@gmail.com; yama@ic.kanagawa-it.ac.jp Abstract Integral photography (IP) is one of the best 3D image display systems because both horizontal and vertical parallaxes can be obtained without a need to wear stereo glasses. IP hardware can most easily be produced by placing a fly s eye lens on a high-definition flat panel display, such as a liquid crystal display (LCD). The price of high-definition LCDs is falling given the progress of its production technology. By contrast, the initial cost of producing a fly s eye lens remains very high because a very expensive metal mold has to be produced in most cases. This problem has been solved by introducing the extended fractional view method, which enables the combination of any ready-made LCD with any ready-made fly s eye lens. However, other problems persist, i.e., the limited types of ready-made fly s eye lens on the market and the lack of large fly s eye lenses. Without such lenses, the 3D display screen size cannot be enlarged. In this work, we solved this problem by replacing a fly s eye lens with two mutually perpendicular lenticular lenses based on the fact that two orthogonally stacked lenticular lenses work as if they were a single fly s eye lens. Unlike fly s eye lenses, large off-theshelf lenticular lenses are available in the market because of their applications in large 3D signboards, among others. Another issue to be addresses is the suppression of very obstructive moiré pattern caused by the interference between the LCD of pixels and small convex lenses. This problem was solved by slightly tilting the orthogonal fly s eye lenses against the LCD. In our experiment, a 3D scene data created with The Persistence of Vision Raytracer (POV-Ray) was rendered from 16 16 different viewpoints so that 256 still images were obtained. Our original software, which was developed in C# language, synthesized an IP image from the 256 images. When the image was displayed on a 4K 28-inch LCD and observed through the orthogonal lenticular lenses, a 3D image with both horizontal and vertical parallaxes was observed. Keywords: 3D display, autostereoscopic, integral photography, slanted lenticular, fractional view 1. Introduction The sense of depth caused by binocular parallax brings about a high sense of reality. Accordingly, various 3D displays have been proposed to date. Among them, the integral photography (IP), proposed by Lippmann in 1908 [1] is one of the best because both horizontal and vertical parallaxes are obtained without a need to wear stereo glasses. If an IP display is placed flat on a desk or on the floor, it can be viewed from any of the surrounding directions. However, compared with other autostereoscopic systems such as the lenticular system or the parallax barrier system (both of which cause parallax only on the horizontal direction), IP remains unpopular probably because IP of the following limitations: 1) the initial cost to produce a fly s eye lens is high (at least several tens of thousands of US dollars), 2) off-the-shelf large fly s lenses are unavailable; and 3) offensive moires sometimes occur. In this work, we present ways to settle these problems. 2. Technologies Used in Our System We developed a new system that combines the technologies of fractional view system, orthogonal lenticular system, and slanted lenticular system. 2.1. Fractional View System In autostereoscopic displays whose lenticular lenses are placed on this side of a liquid crystal display (LCD), the number of views is equal to the number of pixels per tiny semicylindrical lenses, as shown in Fig. 1(i). If the lenticular lens is designed and produced based on the LCD dimensions, the number of views can be made an integer precisely. Conversely, when an off-the-shelf LCD is combined with an off-the-shelf lenticular lens, the probability that the number of views becomes an integral by chance is almost zero. In most cases, the number of view becomes a non-integer, i.e., fractional, as shown in MHCI 112-1

Fig. 1(ii). Even in this case, the path of light emitted from each pixel of the LCD and refracted by the lenticular lens is exactly calculable by using an algorithm similar to the ray-tracing method. Therefore, such a combination can also be used to produce a 3D display. This idea is basic for the fractional view method [2][3] proposed by Ishii. Yanaka [4][5] extended Ishii s method so that it can be applied to the IP, in which a disparity exists in both vertical and horizontal directions. This method is called the extended fractional view (EFV) method, which has also been adopted in the system described in this paper. 3 pixels per lens pitch. 3 is an integer. 3.3 pixels per lens pitch. 3.3 is a non-integer. (i) Example of integer view. Fig. 1: Integer and fractional views. (ii) Example of fractional view. 2.2. Orthogonal Lenticular System Owing to the EFV method, inexpensive ready-made lenticular lenses can be used, which greatly reduces the initial cost. However, other problems persist, i.e., the limited types of ready-made fly s eye lens on the market and the lack of large fly s eye lenses. Without such lenses, the 3D display screen size cannot be enlarged. In this work, we solved this problem by replacing a fly s eye lens with two mutually perpendicular lenticular lenses, as shown in Fig. 1[4] based on the fact that two orthogonally stacked lenticular lenses work as if they were a single fly s eye lens. Unlike fly s eye lenses, large off-the-shelf lenticular lenses are available in the market because of their applications in large 3D signboards, among others. 2.3. Slanted Lenticular System In the lenticular system, which causes a parallax only in the horizontal direction, the number of views can be increased by tilting the lenticular lens against the LCD. A patent of Philips [6] is related to this technique, i.e., tilting to certain angles to obtain the certain number of viewpoints. Conversely, in the case of fractional view, the purpose of tilting the lenticular lens is not to increase the number of viewpoints because the number of viewpoints is infinite from the beginning. The purpose of tilting, as shown in Fig. 3, is to reduce moirés [7][8][9]. In this case, we have considerable latitude in the angle of tilt permitted in this case. We set the angle to 10 in our experiment, but wide-range angles can be chosen. Fig. 2: Normal orthogonal lenticular lenses (normal mutually perpendicular lenticular lenses). Fig. 3: Slanted orthogonal lenticular lenses (slanted mutually perpendicular lenticular lenses). MHCI 112-2

3. Experiment 3.1. System used for the experiment Fig. 4 shows the system configuration. 28-inch 4K LCD (Iiyama ProLite B2888UHSU) 10 degrees Fig. 4: System configuration. 20LPI lenticular lens 20LPI lenticular lens 3.2. Rendering of Multi-viewpoint Images A 3D scene data created with the Persistence of Vision Raytracer (POV-Ray) was rendered from 16 16 different viewpoints so that 256 still images are obtained, as shown in Fig. 5. 16 images Scene data POV-Ray 16 images Fig. 5: Rendering of Multi-viewpoint Images. 3.3. Synthesis of IP Image Our original software, which was developed in C# language and is shown in Fig. 6, was used to synthesize an IP image from the 256 images. First, the coordinates of all the subpixels on the screen are converted from the (x, y) into the (X, Y) based on the following formula, where (x, y) are the coordinates of a subpixel of the coordinate system of the LCD, and (X, Y) are the coordinates of the coordinate system of the lenticular lenses, as shown in Fig. 7. In both cases, units are subpixel. X = x cos θ + y sin θ Y = x sin θ + y cos θ Second, values of x and y are quantized by the width of the small cylindrical lens, thereby revealing the elemental image to which the subpixel belongs. At the same time, the coordinates of the center of the elemental image (X c, Y c) can be obtained. Accordingly, (X c, Y c) are converted to (x c, y c) based on the following formula. x c = X c cos θ Y c sin θ y c = X c sin θ + Y c cos θ Notably, both (x, y) and (x c, y c) are the coordinates of the screen coordinate system considering that the vector (x c -x, y c -y) denotes the vector from the subpixel to the center of the elemental image to which the subpixel belongs. This value denotes the direction of light emitted from the subpixel. Therefore, based on this value, an image is chosen among the 256 still images, and the value of the corresponding subpixel of the image is set to the subpixel of the IP image. This procedure is repeated for all of subpixels of the IP image until this image is completed. MHCI 112-3

Fig. 6: Screenshot of our original software. Y y Width Elemental image θ (Angle of tilt) Subpixel (x, y) Vector denoting the direction of light Center of elemental image (x c, y c) X x Fig. 7: Coordinate system of LCD (xy) and coordinate system of lenticular lens (XY). 4. Results Fig. 8 shows a synthesized IP image. When a part of it is magnified, elemental images become visible. Fig. 9 shows images, displayed with our experimental system, which are viewed from various positions. 5. Conclusion A new IP system combining three technologies (the EFV system, the orthogonal lenticular system, and the slanted lenticular system) was developed. Given that the EFV system was adopted, inexpensive off-the-shelf lens can be used. Moreover, considering the adoption of the orthogonal (or perpendicular) lenticular system, enlarging the screen size is easy because various ready-made large lenticular lenses are available in the market. Moreover, offensive moiré pattern is suppressed by slightly tilting the lenticular lenses to the LCD. Therefore, the method proposed in this paper is suitable for the single or small-scale production of large autostereoscopic displays using IP, which has been difficult up until now. The proposed scheme can also expand the use of digital signage with IP, for example. MHCI 112-4

Enlarged view Fig. 8: Synthesized image. Fig. 9: Image viewed from various positions. Acknowledgements This work was supported by JSPS KAKENHI Grant Number 16K00284. References [1] G. Lippmann, Épreuvestréversibles donnant la sensation du relief, J. Phys. Theor. Appl., vol. 7, pp.821-825, 1908. [2] M. Ishii, Fractional View 3D Display, 3D Image Conference 2004, pp.65-68, 2004 (in Japanese). [3] M. Ishii, Spacial Image by Fractional View Display, ITE Technical Report, vol. 30, no. 58, pp. 33-38, 2006 (in Japanese). [4] K. Yanaka, Integral photography suitable for small-lot production using mutually perpendicular lenticular sheets and fractional view, in Proc. SPIE 6490, Stereoscopic Displays and Virtual Reality Systems XIV, 2007, 9 pages. [5] Integral Photography using Hexagonal Fly s Eye Lens and Fractional View, in Proc. SPIE 6803, Stereoscopic Displays and Applications XIX, 2008, 8 pages. [6] C. V. Berkel and J. A. Clarke, Autostereoscopic display apparatus, U.S. Patent 6,064,424, May 16, 2000. [7] Y. Kim, G. Park, S.-W. Cho, J.-h. Jung, B. Lee, Y. Choi, and M.-G. Lee, Integral imaging with reduced color moire pattern by using a slanted lens array, in Proc. SPIE 6803, Stereoscopic Displays and Applications XIX, 8 pages, 2008. [8] Y. Kim, G. Park, J.-H. Jung, J. Kim, and B. Lee, Color moiré pattern simulation and analysis in three-dimensional integral imaging for finding the moiré-reduced tilted angle of a lens array, Applied Optics, vol. 48, no. 11, pp. 2178 2187, 2009. [9] K. Yanaka and K. Uehira, P-9: Extended Fractional View Integral Imaging Using Slanted Fly s Eye Lens, SID Symposium Digest of Technical Papers, vol. 42, no. 1, pp. 1124 1127, 2011. MHCI 112-5