HIGH SPEED 3-D MEASUREMENT SYSTEM USING INCOHERENT LIGHT SOURCE FOR HUMAN PERFORMANCE ANALYSIS

Similar documents
DEVELOPMENT OF REAL TIME 3-D MEASUREMENT SYSTEM USING INTENSITY RATIO METHOD

RECONSTRUCTION OF REGISTERED RANGE DATA USING GEODESIC DOME TYPE DATA STRUCTURE

A High Speed Face Measurement System

Structured Light II. Guido Gerig CS 6320, Spring (thanks: slides Prof. S. Narasimhan, CMU, Marc Pollefeys, UNC)

Handy Rangefinder for Active Robot Vision

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera

Measurement of Pedestrian Groups Using Subtraction Stereo

Optical Imaging Techniques and Applications

Multi-view reconstruction for projector camera systems based on bundle adjustment

Gesture Recognition using Temporal Templates with disparity information

Color-Stripe Structured Light Robust to Surface Color and Discontinuity

EASY PROJECTOR AND MONOCHROME CAMERA CALIBRATION METHOD USING PLANE BOARD WITH MULTIPLE ENCODED MARKERS

Registration of Moving Surfaces by Means of One-Shot Laser Projection

Surround Structured Lighting for Full Object Scanning

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

AUTOMATED CALIBRATION TECHNIQUE FOR PHOTOGRAMMETRIC SYSTEM BASED ON A MULTI-MEDIA PROJECTOR AND A CCD CAMERA

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

3D Scanning Method for Fast Motion using Single Grid Pattern with Coarse-to-fine Technique

Active Stereo Vision. COMP 4900D Winter 2012 Gerhard Roth

Project 3 code & artifact due Tuesday Final project proposals due noon Wed (by ) Readings Szeliski, Chapter 10 (through 10.5)

A PHOTOGRAMMETRIC METHOD FOR ENHANCING THE DETECTION OF BONE FRAGMENTS AND OTHER HAZARD MATERIALS IN CHICKEN FILETS

Measurements using three-dimensional product imaging

Measurement of 3D Foot Shape Deformation in Motion

A 3-D Scanner Capturing Range and Color for the Robotics Applications

Projector Calibration for Pattern Projection Systems

Project 2 due today Project 3 out today. Readings Szeliski, Chapter 10 (through 10.5)

Multi-View Stereo for Community Photo Collections Michael Goesele, et al, ICCV Venus de Milo

Stereo vision. Many slides adapted from Steve Seitz

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe

3D Computer Vision. Structured Light I. Prof. Didier Stricker. Kaiserlautern University.

Overview of Active Vision Techniques

Multiple View Geometry

Robust and Accurate One-shot 3D Reconstruction by 2C1P System with Wave Grid Pattern

Construction of an Immersive Mixed Environment Using an Omnidirectional Stereo Image Sensor

Range Sensors (time of flight) (1)

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Three-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera

Modeling, Combining, and Rendering Dynamic Real-World Events From Image Sequences

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Trapezoidal phase-shifting method for threedimensional

Color-Stripe Structured Light Robust to Surface Color and Discontinuity

A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera

A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision

Projection Center Calibration for a Co-located Projector Camera System

Recap from Previous Lecture

Structured light 3D reconstruction

Acquiring 3D Model of Object by Motion-Stereo

CS5670: Computer Vision

3D Corner Detection from Room Environment Using the Handy Video Camera

AN AUTOMATIC 3D RECONSTRUCTION METHOD BASED ON MULTI-VIEW STEREO VISION FOR THE MOGAO GROTTOES

: Easy 3D Calibration of laser triangulation systems. Fredrik Nilsson Product Manager, SICK, BU Vision

A Coded Structured Light Projection Method for High-Frame-Rate 3D Image Acquisition

Sensing Deforming and Moving Objects with Commercial Off the Shelf Hardware

Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement

AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor

Distortion Correction for 3D Scan of Trunk Swaying Human Body Segments

The main problem of photogrammetry

A COMPREHENSIVE SIMULATION SOFTWARE FOR TEACHING CAMERA CALIBRATION

3D-OBJECT DETECTION METHOD BASED ON THE STEREO IMAGE TRANSFORMATION TO THE COMMON OBSERVATION POINT

Introduction to Computer Vision. Introduction CMPSCI 591A/691A CMPSCI 570/670. Image Formation

Multimedia Technology CHAPTER 4. Video and Animation

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION

10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1.

Dynamic 3-D surface profilometry using a novel color pattern encoded with a multiple triangular model

APPROACH TO ACCURATE PHOTOREALISTIC MODEL GENERATION FOR COMPLEX 3D OBJECTS

A 100Hz Real-time Sensing System of Textured Range Images

Compact and Low Cost System for the Measurement of Accurate 3D Shape and Normal

RECENT PROGRESS IN CODED STRUCTURED LIGHT AS A TECHNIQUE TO SOLVE THE CORRESPONDENCE PROBLEM: A SURVEY

Dedicated Software Algorithms for 3D Clouds of Points

Laser Eye a new 3D sensor for active vision

L2 Data Acquisition. Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods

Skeleton Cube for Lighting Environment Estimation

Auto-focusing Technique in a Projector-Camera System

ENGN 2911 I: 3D Photography and Geometry Processing Assignment 2: Structured Light for 3D Scanning

Introduction to 3D Machine Vision

Using temporal seeding to constrain the disparity search range in stereo matching

Camera Calibration Utility Description

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg

Occlusion Detection of Real Objects using Contour Based Stereo Matching

PHOTOGRAMMETRIC TECHNIQUE FOR TEETH OCCLUSION ANALYSIS IN DENTISTRY

Tecnologie per la ricostruzione di modelli 3D da immagini. Marco Callieri ISTI-CNR, Pisa, Italy

A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Stereo and Epipolar geometry

3D graphics, raster and colors CS312 Fall 2010

Rectification and Disparity

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR

Spacetime Stereo: A Unifying Framework for Depth from Triangulation

Development of a video-rate range finder using dynamic threshold method for characteristic point detection

Agenda. DLP 3D scanning Introduction DLP 3D scanning SDK Introduction Advance features for existing SDK

FOREGROUND DETECTION ON DEPTH MAPS USING SKELETAL REPRESENTATION OF OBJECT SILHOUETTES

Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera

3D Modeling of Objects Using Laser Scanning

Understanding Variability

3D MODELING OF OUTDOOR SCENES BY INTEGRATING STOP-AND-GO AND CONTINUOUS SCANNING OF RANGEFINDER

Transcription:

HIGH SPEED 3-D MEASUREMENT SYSTEM USING INCOHERENT LIGHT SOURCE FOR HUMAN PERFORMANCE ANALYSIS Takeo MIYASAKA, Kazuhiro KURODA, Makoto HIROSE and Kazuo ARAKI School of Computer and Cognitive Sciences, Chukyo University. 101, Tokodachi, Kaizu-cho, Toyota-shi, 470-0393 Japan miyasaka@grad.sccs.chukyo-u.ac.jp KEY WORDS: 3D measurement, High speed, Incoherent light source, Motion analysis ABSTRACT This paper presents a newly developed high speed 3-D measurement system, which enable us to obtain continuously range image of 768x480 spatial resolution at video rate. It is based on a light pattern projection method and uses incoherent light source instead of laser beam. So it is not dangerous to environment. To obtain one range image, it needs to project only two types of light patterns. Each light pattern has different intensity profile toward specific direction. Then, the target space is encoded to many thin slit-like regions by the intensity ratio of those two light patterns. So, taking video images corresponding to each light pattern, we can determine the slit-like region corresponding to each pixel by the density ratio of it in those two video images. As a result, we can obtain a range image by using triangulation. Trial system has ensured the performance shown in the beginning. 1 INTRODUCTION Recently, a 3-D measurement system which acquires range image under non-contact condition has been in actuality by advancement of computer and the peripherals. Most systems, however, can measure only motionless object. Then, it is difficult that they apply to areas of measurement of a moving target, mobile robot and so on. The fast 3-D measurement system has highly expected to application of above mentioned areas. To perform fast 3-D measurement, various researches have been carried out. 3-D measurement methods can be classified into two category, passive method and active method. A typical passive method is stereo vision. It has been most actively studied. An example of high speed stereo vision is the multi stereo method presented by Kanade et al (Kanade and Okutomi, 1992). In this method, the moving target is taken by a number of video cameras. Successive range images can be obtained from those captured video images. It, however, requires huge computing time in the post-processing and its accuracy is not sufficient, so it is difficult to apply it to the area which needs real time operation, such as mobile robot vision and so on. The slit-ray projection method is the most popular one of active methods. In this method, it needs to project a slit like beam onto the target and take the reflection light image. It realizes to obtain a range image in smaller computing time than a stereo vision. However, it needs a number of video images corresponding to the spatial resolution. So, it needs further improvement to perform a high speed range finding. To resolve the problem, the pattern light projection method is proposed. In the method, the light which is structured to have 2-D pattern is projected onto the target space. The target space is 2-dimensionally encoded by the structured light. To encode the target space, some features of light pattern are used, for example, color (Boyer and Kak, 1980) (Tajima and Iwakawa, 1990) (Caspi et al., 1998) and light intensity (Carrihill and Hummel, 1985). The one common method of light pattern projection method is a gray coded light projection method (Inokuchi et al., 1984) (Sato and Inokuchi, 1985). The method encodes the target space by the gray-coded light pattern. It is robust method because the error is decreased by using gray code. In addition, the method can measure faster than a simple slit-ray projection method. For example, number of video images needed to complete one measurement are 8 frames in case of 256 horizontal resolution. However, the measurement speed of it is not fast enough for measurement of moving target. In such situation, Araki et al (Araki et al., 1991) (Araki et al., 1992) (Araki et al., 1995) proposed a new slit-ray projection method which uses non-scanning type of image plane. The key component is the image plane constructed by a positionsensitive device(psd) array that is horizontally unsegmented and linear, but vertically segmented. By virtue of this configuration of the image plane, one range image is acquired during only one scan of the slit-ray at high speed, This system enables us to obtain range image of 128x128 spatial resolution continuously at 1 / 30 seconds per scene with an International Archives of Photogrammetry and Remote Sensing. Vol. XXXIII, Part B5. Amsterdam 2000. 547

(a) Flat pattern (b) Linear pattern Figures 1: Projected light patterns error within ± 0.3 %. By using this system, we can obtain not only 3-D shapes but also 3-D movement information of the target. The system, however, uses a laser beam for a light source, so it has problem in safety, and it is difficult to apply to areas of human face measurement and mobile robot and so on. Furthermore, it has problems in the cost, and its spatial resolution is not enough for many applications, caused by using the particular element for a image plane. In this paper, we propose a safe and inexpensive 3-D measurement system that can continuously obtain high resolution range images at high speed. It is based on a intensity ratio method which is one branch of light projection methods. Trial system has ensured the performance of 768x480 spatial resolution and continuous measurement at video rate. 2 INTENSITY RATIO RANGE FINDING METHOD The intensity ratio method is a kind of pattern projection methods. The key concept of it is to encode the target space by light intensity of projected light patterns. In our method, two types of light patterns are projected alternately onto the target. They are shown in Figures 1. The first pattern of light is designed to such that light intensity profile is flat in the target space and we call it flat pattern (Figures 1 (a)). The light intensity of the second pattern increases linearly toward the specific direction in the target space(figures 1 (b)). That is called linear pattern. As a result, the area, at which the ratio of light intensity by two light patterns is equivalent, belongs to a thin slit-like area in the target space. That is, the target space is encoded by the ratio of light intensity of each light pattern. So, the target space is distinguished by that ratio as a slit-like region. Therefore, we designate the ratio of intensity of flat and linear pattern lights as the slit-code. In current system, the light pattern is produced on a computer, and it is projected onto the target space through a digital projector. Next, the video image of the target which corresponds to each light pattern is taken by a video camera. Then, the ratio of density on each pixel gives the slit-code, so we can decide in which slit area the point on the object imaging on that pixel exists. So, we can calculate range data by the triangulation. As the unique slit-code is decided to each pixel, the spatial resolution of obtained range images is equivalent to the resolution of the video camera used. So, our measurement method can give range images of extremely high resolution. Our method can compute range image of one scene from 2 successive frames of video images. In detail, the first frame of range image is calculated from in the first and the second video images. The second range image is obtained from the second and the third video images. By such way, the measurement speed becomes 30 scenes per second in case of using normal video camera and 120 scenes per second in case of using quadruple speed video camera which developed in a recent, although there is time lag of only one frame at the beginning. As we can obtain a color 2-D video image at the same time of measurement, highly realistic color 3-D model can be reconstructed easily. Furthermore, since our method uses incoherent light, the intensity of light can be increased. In consequence, we can measure a longer distance. 3 THE TRIAL SYSTEM We developed the trial system based on the preceding method. The configuration of this system is shown in Figure 2. The system uses a color 3 CCD video camera (Sony XC-003) for the image grabber. It outputs color video image sequence as NTSC signal. For the projection system, we use a color LCD projector(epson ELP-5100) on the market. It projects above mentioned flat and linear light patterns which are generated by a computer. The computer is also used for control of the camera and the projector. Video images of the target space are taken by the video camera, then, captured by the computer through frame grabber. The measurement process is performed in 1 / 30 seconds. 548 International Archives of Photogrammetry and Remote Sensing. Vol. XXXIII, Part B5. Amsterdam 2000.

target object digital projector video camera Figure 2: Outline of the measurement system Since the system is constructed by mass production component, we must calibrate the configuration of some parameters, for example, lens position, lens distortion and so on. Which is performed by Tsai s method (Tsai, 1986)(Tsai, 1987) in this paper. Ratio of light intensity on same pixel in successive two video images is computed on the computer. Then, the coordinate of the respective point on the target is calculated by the triangulation principle. The system can obtain a range image in 1 / 30 seconds. By successively executing the process, the range image sequence can be obtained. Thus the system enables us to measure a moving target and to calculate information on movement of it. 4 IMPROVEMENT OF ACCURATE In intensity ratio method, it is important how to recover the intensity ratio of projected light patterns from respective video images. In other words, if the slit-code corresponding to each pixel is correctly detected, the measurement accuracy is improved. To eliminate influence of the environment, two types of light patterns are projected and the density ratio of same pixel in respective video images is used as label of the corresponding slit-code. But conventional intensity ratio method have not succeeded in accurate measurement. The reason, we think, is that they don t consider dependence of reflectance to light frequence. That is, the surface of the target has different reflectance for each frequency of light consisting incident light patterns. So, in case that the target space is encoded by gray-scale light as in conventional intensity ratio method, intensity ratio of incident light doesn t correspond to density ratio of same pixel in respective video images. So, to improve accuracy of measurement, we must use intensity and density of monochromatic light instead of gray-scale light. The 3 CCD camera used in this research picks up separately light intensities of 3 monochromatic colors, i.e, red, green and blue. So, we can use intensity and density ratio of red, green and blue, independently. Thus we have overcome the problem mentioned above. 5 RESULT We tried 3-D measurement by using above trial system. The result of labeling of slit-code is shown in Figure 3. This image shown slit-code number as pixel brightness. A smaller slit-code is shown as darker pixel, and larger is brighter. Figure 4 shows the range image of the target. In this figure, the depth of each pixel is shown as brightness. A closer point is expressed brighter, and far one is darker. In the result of experiment, we can obtain successive range images at video rate with an error within ± 1%. International Archives of Photogrammetry and Remote Sensing. Vol. XXXIII, Part B5. Amsterdam 2000. 549

Figure 3: Labeling Image Figure 4: Range images 6 CONCLUSION By using proposed intensity ratio method, we achieved high speed and successive range finding. Using informations of captured video images for red, blue and green monochromatic colors separately, we can improve the system accuracy. As the result, the measurement system can obtain stable range images without influence of the environment. In addition, this measurement system is safe to environment and can be constructed in low cost. We think that the measurement system can be applied various fields, for example the vision system in mobile robot, a movement analysis of gymnastics player and so on. In particular, we are planing to apply it to a human motion analysis. The human motion analysis system will be constructed by a number of this measurement system. The analysis system may obtain 3-D human motion informations automatically. ACKNOWLEDGEMENT This work was partially supported by Kayamori Foundation of Information Science Advancement and Foundation of Chukyo University for Advanced Research. REFERENCES Araki, K., Sato, M., Noda, T., Chiba, Y. and Shimizu, M., 1991. High speed and continuous rangefinding system. IEICE Transactions Special Issue on Computer Vision and Its Applications E 74, pp. 3400 3406. Araki, K., Shimizu, M., Noda, T., Chiba, Y., Tsuda, Y., Ikeg aya, K. and Sannomiya, K Gomi, M., 1992. High speed and continuous 3-d measurement system. 11th IAPR International Conference on Pattern Recognition 4, pp. 62 65. Araki, K., Shimizu, M., Noda, T., Chiba, Y., Tsuda, Y., Ikegaya, K., sannnomiya, K. and Gomi, M., 1995. A high-speed and continuous 3d measurement system. Machine Vision and Applications 8, pp. 79 84. Boyer, K. and Kak, A., 1980. Color-encoded structural light for rapid active ranging. IEEE Trans Pattern Analysis and Machine Intelligence 9, pp. 14. Carrihill, B. and Hummel, R., 1985. Experiments with the intensity ratio depth sensor. Comput. Vis. Graph. and Image Process 32, pp. 337 358. Caspi, D., Kiryati, N. and Shamir, J., 1998. Range imaging with adaptive color structured light. IEEE Trans. Pattern Analysis and Machine Intelligence 20(5), pp. 470 480. Inokuchi, S., Sato, K. and Matsuda, F., 1984. Range-imaging system for 3-d range imaging. Proc. 7th ICPR pp. 806 808. Kanade, T. and Okutomi, M., 1992. A multiple base-line stereo method. DARPA pp. 409 429. Sato, K. and Inokuchi, S., 1985. Three-dimensional surface measurement by space encoding range imaging. J. Robotic Systems 2, pp. 27039. 550 International Archives of Photogrammetry and Remote Sensing. Vol. XXXIII, Part B5. Amsterdam 2000.

Tajima, J. and Iwakawa, M., 1990. 3-d data acquisition by rainbow range finder. Proc. 10th ICPR pp. 309 313. Tsai, R. Y., 1986. An efficient and accurate camera calibration technique for 3d machine vision. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition pp. 364 374. Tsai, R. Y., 1987. A versatile camera calibration technique for high-accuracy 3d machine vision metrology using off-theshelf tv cameras and lenses. IEEE of Robotics and Automation RA-3(4), pp. 323 244. International Archives of Photogrammetry and Remote Sensing. Vol. XXXIII, Part B5. Amsterdam 2000. 551