3D-2D Laser Range Finder calibration using a conic based geometry shape

Similar documents
Image Transformations & Camera Calibration. Mašinska vizija, 2018.

Reconstruction of complete 3D object model from multi-view range images.

Project Title: Welding Machine Monitoring System Phase II. Name of PI: Prof. Kenneth K.M. LAM (EIE) Progress / Achievement: (with photos, if any)

LATEST TRENDS on APPLIED MATHEMATICS, SIMULATION, MODELLING

AUTOMATIC ORIENTATION AND MERGING OF LASER SCANNER ACQUISITIONS THROUGH VOLUMETRIC TARGETS: PROCEDURE DESCRIPTION AND TEST RESULTS

LUMS Mine Detector Project

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit

Projector Calibration for Pattern Projection Systems

#65 MONITORING AND PREDICTING PEDESTRIAN BEHAVIOR AT TRAFFIC INTERSECTIONS

Processing 3D Surface Data

Processing 3D Surface Data

Construction and Calibration of a Low-Cost 3D Laser Scanner with 360º Field of View for Mobile Robots

3D Reconstruction of Real World Scenes Using a low-cost 3D range scanner

An Efficient Algorithm for Extrinsic Calibration between a 3D Laser Range Finder and a Stereo Camera for Surveillance

Introduction to 3D Machine Vision

Calibration of a rotating multi-beam Lidar

REFINEMENT OF COLORED MOBILE MAPPING DATA USING INTENSITY IMAGES

CREATING 3D WRL OBJECT BY USING 2D DATA

CLASSIFICATION FOR ROADSIDE OBJECTS BASED ON SIMULATED LASER SCANNING

Structured light 3D reconstruction

Estimation of Camera Motion with Feature Flow Model for 3D Environment Modeling by Using Omni-Directional Camera

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,

MERGING POINT CLOUDS FROM MULTIPLE KINECTS. Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia

Automatic camera to laser calibration for high accuracy mobile mapping systems using INS

SIMPLE ROOM SHAPE MODELING WITH SPARSE 3D POINT INFORMATION USING PHOTOGRAMMETRY AND APPLICATION SOFTWARE

Stereoscopic Vision System for reconstruction of 3D objects

A NEW AUTOMATIC SYSTEM CALIBRATION OF MULTI-CAMERAS AND LIDAR SENSORS

Interactive Collision Detection for Engineering Plants based on Large-Scale Point-Clouds

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

A FAST LOW-COST 3D SCANNER FOR NAVIGATION AND MODELLING. Miguel Matos, Vítor Santos, Paulo Dias*

Single-shot Extrinsic Calibration of a Generically Configured RGB-D Camera Rig from Scene Constraints

PHOTOGRAMMETRIC TECHNIQUE FOR TEETH OCCLUSION ANALYSIS IN DENTISTRY

3D Object Representations. COS 526, Fall 2016 Princeton University

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion

Registration of Moving Surfaces by Means of One-Shot Laser Projection

An Acquisition Geometry-Independent Calibration Tool for Industrial Computed Tomography

ENGN D Photography / Spring 2018 / SYLLABUS

Multiple View Geometry

calibrated coordinates Linear transformation pixel coordinates

Scan-point Planning and 3-D Map Building for a 3-D Laser Range Scanner in an Outdoor Environment

Multiple View Geometry in Computer Vision Second Edition

Stereo Image Rectification for Simple Panoramic Image Generation

Ray Tracer I: Ray Casting Due date: 12:00pm December 3, 2001

High Altitude Balloon Localization from Photographs

Dedicated Software Algorithms for 3D Clouds of Points

Chapters 1 7: Overview

EECS 442: Final Project

Iterative Closest Point Algorithm in the Presence of Anisotropic Noise

A Desktop 3D Scanner Exploiting Rotation and Visual Rectification of Laser Profiles

Rigid ICP registration with Kinect

Flexible Calibration of a Portable Structured Light System through Surface Plane

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006

Computer Vision cmput 428/615

Algorithm of correction of error caused by perspective distortions of measuring mark images

Absolute Scale Structure from Motion Using a Refractive Plate

AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor

MULTIPLE-SENSOR INTEGRATION FOR EFFICIENT REVERSE ENGINEERING OF GEOMETRY

REGISTRATION OF AIRBORNE LASER DATA WITH ONE AERIAL IMAGE

Advances in 3D data processing and 3D cameras

Stereo Vision. MAN-522 Computer Vision

3D Point Cloud Processing

BIM-based Indoor Positioning Technology Using a Monocular Camera

Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming

MOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE

3D Reconstruction of Outdoor Environments from Omnidirectional Range and Color Images

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Shape Model-Based 3D Ear Detection from Side Face Range Images

3D Digitization of Human Foot Based on Computer Stereo Vision Combined with KINECT Sensor Hai-Qing YANG a,*, Li HE b, Geng-Xin GUO c and Yong-Jun XU d

Automatic Reconstruction of Colored 3D Models

Study on Gear Chamfering Method based on Vision Measurement

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences

Pattern Feature Detection for Camera Calibration Using Circular Sample

Construction Progress Management and Interior Work Analysis Using Kinect 3D Image Sensors

LIDAR and Panoramic Camera Extrinsic Calibration Approach Using a Pattern Plane

Stereo Vision Inside Tire

Jo-Car2 Autonomous Mode. Path Planning (Cost Matrix Algorithm)

arxiv: v1 [cs.cv] 28 Sep 2018

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism

3D MODELING OF OUTDOOR SCENES BY INTEGRATING STOP-AND-GO AND CONTINUOUS SCANNING OF RANGEFINDER

3D Models from Range Sensors. Gianpaolo Palma

3D Modeling of Objects Using Laser Scanning

3D Modeling of Outdoor Environments by Integrating Omnidirectional Range and Color Images

Automatic Generation of Indoor VR-Models by a Mobile Robot with a Laser Range Finder and a Color Camera

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

1 Projective Geometry

A 100Hz Real-time Sensing System of Textured Range Images

A Fast Linear Registration Framework for Multi-Camera GIS Coordination

Extrinsic Calibration of Camera and 2D Laser Sensors without Overlap

Shape and Appearance from Images and Range Data

Cut-And-Sew: A Distributed Autonomous Localization Algorithm for 3D Surface Sensor Networks

Freehand Laser Scanning Using Mobile Phone

Computer Vision Projective Geometry and Calibration. Pinhole cameras

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

L2 Data Acquisition. Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods

DERIVING PEDESTRIAN POSITIONS FROM UNCALIBRATED VIDEOS

Processing 3D Surface Data

Transcription:

3D-2D Laser Range Finder calibration using a conic based geometry shape Miguel Almeida 1, Paulo Dias 1, Miguel Oliveira 2, Vítor Santos 2 1 Dept. of Electronics, Telecom. and Informatics, IEETA, University of Aveiro, Portugal {m.almeida,paulo.dias}@ua.pt 2 Dept of Mechanical Engineering, University of Aveiro, Portugal {mriem,vitor}@ua.pt Abstract. The AtlasCar is a prototype that is being developed at the University of Aveiro to research advanced driver assistance systems. The car is equipped with several sensors: 3D and 2D laser scanners, a stereo camera, inertial sensors and GPS. The combination of all these sensor data in useful representations is essential. Therefore, calibration is one of the first problems to tackle. This paper focuses on 3D/2D laser calibration. The proposed method uses a 3D Laser Range Finder (LRF) to produce a reference 3D point cloud containing a known calibration object. Manual input from the user and knowledge of the object geometry are used to register the 3D point cloud with the 2D Lasers. Experimental results with simulated and real data demonstrate the effectiveness of the proposed calibration method. Keywords: laser range finder; laser to laser calibration; sensor fusion 1 INTRODUCTION Mobile platforms typically combine several data acquisition systems such as lasers, cameras and inertial systems. However the geometrical combination of the different sensors requires their calibration, at least, through the definition of the extrinsic parameters, i.e., the transformation matrices that register all sensors in the same coordinate system. Geometric calibration is an active field of study due to its importance in sensor fusion. Several methods use cameras combined with laser range finders (LRF) and a chessboard pattern for calibration [1, 2]. Other approaches require the user to select key points in laser data in order to perform the calibration [3]. A more similar solution to the one proposed in this paper can be found in Lisca, Pangyu and Nedevshi [4], where the calibration between a multi-line LRF and a stereo camera is obtained from the information of a known calibration object in the scene. This paper focuses on 3D/2D laser calibration for the AtlasCar project, which is a prototype for research on Advanced Driver Assistance Systems [5]. The modified car includes several sensors as illustrated on Fig. 1.The 3D laser scanner is based on a modified 2D laser scanner originally proposed in Dias, Matos and Santos [6] and further modified to its actual configuration [7]. Besides the 3D laser, two 2D SICK

LMS 151 lasers are mounted on the front bumper of the car. Other sensors are also available, although they will not be considered in this work. The actual calibration process of the AtlasCar sensors is based on manual measurements that are time consuming and prone to errors. Besides this, any accidental or deliberated change in the position of the sensors or the addition of a new laser forces a new time consuming manual calibration process. 3D laser 2D laser Fig. 1. The AtlasCar. 2D laser 2 PROPOSED CALIBRATION PROCESS This section presents the calibration procedure used to register the 3D/2D lasers available. The proposed approach uses a static calibration object measured in the acquired scene. However, the calibration process may use several calibration objects in the scene. The point cloud acquired by the 3D laser scanner is used as reference to calibrate the 2D laser scanners. A 3D point cloud simulator was also developed in order to evaluate the calibration algorithm. 2.1 Calibration object The calibration object was selected having in mind the unambiguous determination of the position and orientation of the 2D laser footprints regarding the 3D point cloud. According to these requirements, a conic geometry is interesting since it removes ambiguity in height. The combination with a second cone provides enough information to determine the orientation around the vertical axis. An intermediate plane was also added to keep the cones at a fixed distance and may be used in future work to allow camera calibration by adding visual patterns. A virtual representation of the calibration object is presented in section 3.2. 2.2 Point cloud simulator The limited availability of the AtlasCar (demonstrations, repairs, mechanical modifications) led to the development of a point cloud simulator module that allows the

creation of 2D and 3D point clouds by providing several parameters. The simulator also proved to be useful to evaluate the precision of the calibration process. The simulator uses a 3D polygonal mesh that can be either computer generated or based on real data acquired from the 3D sensors. Additional 3D objects can be added to the model, for example a calibration object model can be added to the scene at a given position and orientation. The user can then specify the parameters of the laser sensor and the simulator generates a new point cloud by intercepting laser rays with the configured polygonal mesh. Fig. 2 presents a simulated point cloud. 2D Laser Fig. 2. - (left) 2D (in red) and 3D simulated point cloud (points in grey); (right) detail of the 2D laser 2.3 Calibration model detection and processing The process of calibration requires the user to specify the approximate position of the calibration object in the 3D point cloud. This is done by selecting a point in the center of the calibration object (red sphere in Fig. 3) and an additional point in the plane that is between the two cones (blue sphere in Fig. 3). Fig. 3. Calibration object detection The selected points define approximately the center and orientation around the Z axis of the calibration object with respect to the 3D laser. The rotation around the other axes is not considered in this step since it is assumed that the calibration object is in a flat surface. The 3D data is then filtered using a bounding box based on the known object dimensions. With this process the 3D points belonging to the calibration object are extracted. Then a virtual calibration object is fit to the extracted data using the Iterative Closest Point [8]. The result of this fitting is a transformation matrix that moves the calibration object from the origin of the 3D point cloud reference system to the calculated position. 2.4 Ellipse localization and reconstruction Having detected the position of the calibration object in the 3D data, it is now necessary to do the same in the 2D data, in order to compute the transformation from the 2D laser to the 3D laser. The cones in the calibration object produce two ellipse sections in the 2D laser data. The user selects two points per ellipse section as shown in

Fig. 4. The selection of these points allows the extraction of the points from the 2D laser data that belong to the ellipses. Fig. 4. - Points selection for ellipse reconstruction An ellipse fitting method is used to find the ellipse parameters that best fit the extracted points according to minimum square errors [9]. For each ellipse, the algorithm returns the center, major and minor axis length and the angle between the major axis and the Y axis. An additional step is used to get the angle between the calibration object direction and the ellipse s major axis as shown in Fig. 5. The final calibration angle is the one between the calibration object and the Y axis. For the calculation of the calibration object direction, opencv is used to determine the vector that best fits the points on the plane of the calibration object. These points are automatically selected based on the previous ellipses point selection, shown in Fig. 5. Fig. 5. Ellipses obtained in the 2D laser 2.5 Fitting the ellipse to the cone Geometrically, an ellipse has only one way to fit into a cone in terms of height and inclination. The only indetermination would be in the rotation angle around the vertical axis of the cone, which is solved by the angle given by the ellipse reconstruction algorithm described above. Fig. 6 (left) presents a 2D triangle corresponding to a cut in the cone along the ellipse s major axis. From this view, it is possible to retrieve the

height and angle of the ellipse in the cone. Fig. 6 (right) shows a view of the ellipse that is used to retrieve the major and minor axis, both figures are related by the distance. Fig. 6. (left) cone s cut along the major axis; (right) projection of the Ellipse From Fig. 6 we find: ( ) (1) ( ( )) (2) A numeric approximation is used to solve equations (1) and (2), retrieving the height and inclination angle of the ellipse. These values combined with the ellipse orientation (section 2.4) define univocally the ellipse fitting in the cone. 2.6 2D laser transform calculation From the analytical model of each ellipse, four points are computed in the intersections of the ellipse with the major and minor axes, as shown in Fig. 7. Hence, a total of eight points are obtained from the fitting of the ellipses to the 2D laser data (section 2.4). Also, another eight points are obtained from the fitting of the ellipse to the cones in the calibration object (section 2.5). Since this process is done for each ellipse, each calibration will provide 8 points in the 3D point cloud and the 3D laser cloud. Rigid body transform functionalities offered by VTK [10] are then used to compute the transformation matrix that best transforms those eight points from the 2D laser data into the eight transformed points in the 3D laser data. This transformation matrix corresponds to the 2D laser calibration matrix. Fig. 7. Ellipse fitting in a cone in a tridimensional space

3 EXPERIMENTAL RESULTS Several experiments were performed using simulated and real data to validate our approach. In these tests, the number of calibration objects used varies from one to three. In simulated data, the results were validated by comparing the known transformation with the computed one. In real data, since no ground truth is available, a visual evaluation was performed comparing the proposed calibration with the manual method that was used previously. 3.1 Using the point cloud simulator The angular resolution of the 2D and 3D laser were defined to 0.25º, the same value used on the real lasers of the AtlasCar. All simulated results are based on a triangulated model of a scene. Fig. 8 shows a simulation using three calibration objects. Fig. 8. Simulated point cloud with three calibration objects included in a scene. With one calibration object on the scene, the mean distance between the calibration and ideal points was approximately 5 cm. With two calibration objects, the distance decreased to 0.17 cm and with three calibration objects, the mean distance drops to 0.05 cm. In this last case, a visual analysis of the data cannot find differences from the ground truth and calibrated data. As expected, the presence of more calibration objects increases the accuracy of the proposed method. Fig. 9 shows a detail of the ground truth and calibrated data of a calibration using a single calibration object. Calibrated data Ground truth data Fig. 9. Ground truth and calibrated 2D laser data comparison using one calibration object 3.2 Using real data from the AtlasCar A real scale prototype of the calibration object was made to perform real data acquisition (Fig. 10(left)). Fig. 10(right) shows the real acquisition site. To avoid the

construction of several calibration objects, we simulate various object locations by capturing several times the same scene with the calibration object in different positions. The several acquisitions were then merged into a single point cloud with the required number of calibration objects. Tests were made using three and four of these objects. Fig. 10. (left) virtual calibration object; (center) - Calibration object; (right) - acquisition location The calibration process is analogous to the one described using simulated data. The results of the calibration using four calibration objects are shown in Fig. 11. In red we present the actual calibration based on manual measurements and in yellow the result using the calibration process. Visually we notice a notorious improvement in alignment with the 3D laser data in the case of the calibration method here presented (see Fig. 11(a) and (b)). The main limitation of the proposed process appears in Fig. 11(c). The resulting calibration (yellow) is rotated around the X axis. This is due to errors in the 2D fitting of the ellipses, related to the limited number of points available in the 2D point cloud (in some cases the fitting was computed with only 8 points per ellipse). (a) (b) Fig. 11. Results of the proposed calibration method in real data. (a) Detail of the scene; (b) Detail on the calibration objects; (c) Overview of the scene. (c)

4 CONCLUSION AND FUTURE WORK The proposed calibration process is capable of calibrating 2D to 3D laser with results that are clearly better than the previously existing manual calibration as verified in the real data experimental results. The main actual limitation is the estimation of rotation around the X axis related to the resolution of 2D laser (providing few points for ellipse fitting). Possible solutions for this problem are increasing the resolution of the 2D lasers to get more points for the ellipse fitting algorithm, and/or defining restrictions on the position, size and shape of the ellipse/object, minimizing the impact of estimation errors on the final calibration (for example considering that the car and calibration objects are all in the same plane and disregarding the angle around X axis). Future work also involves an automatic identification of the calibration object(s) either in the 2D or the 3D scene to reduce user interaction. Finally, the integration of camera calibration is planned, since the proposed calibration object allows the addition of camera calibration tools like the use of patterns (chessboard). All these modules can be added to the existing application that is a starting point for a fully integrated calibration of all sensors available on the AtlasCar. 5 REFERENCES 1. Qilong, Z. and R. Pless. Extrinsic calibration of a camera and laser range finder (improves camera calibration). Intelligent Robots and Systems (IROS 2004). International Conference on, Vol. 3 (2 October 2004), pp. 2301-2306 2. Tingxiang Jia; Mengyin Fu; Yi Yang; Hao Zhu;, "Calibration of a 3D laser range finder and a camera based on stereoscopic checkerboard," Vehicular Electronics and Safety (ICVES), 2011 IEEE International Conference on, vol., no., pp.92-96, 10-12 July 2011. 3. Huijing Zhao; Yuzhong Chen; Shibasaki, R.;, "An efficient extrinsic calibration of a multiple laser scanners and cameras' sensor system on a mobile platform," Intelligent Vehicles Symposium, 2007 IEEE, vol., no., pp.422-427, 13-15 June 2007. 4. Lisca, G.; Pangyu Jeong; Nedevschi, S.;, "Automatic one step extrinsic calibration of a multi layer laser scanner relative to a stereo camera," Intelligent Computer Communication and Processing (ICCP), IEEE International Conference on, vol., no., pp.223-230, 26-28 Aug. 2010. 5. Santos, V., Almeida, J., Ávila, E., Gameiro, D., Oliveira, M., Pascoal, R., Sabino, R., Stein, P. ATLASCAR. Technologies for a computer assisted driving system on board a common automobile. Conference on Intelligent Transportation Systems (ITSC), 2010. p. 1421-1427. 6. Dias, P., M. Matos, and V. Santos, 3D Reconstruction of Real World Scenes Using a Low-Cost 3D Range Scanner. Computer-Aided Civil and Infrastructure Engineering, 2006. 21(7): p. 486-497. 7. R. Pascoal, V. Santos. Compensation of Azimuthal Distortions on a Free Spinning 2D Laser Range Finder for 3D Data Set Generation, 10th Int. Conference on Mobile Robots and Competitions ROBOTICA2010, pp. 41-46, Leiria, Portugal 8. Besl, P.J. and H.D. McKay, A method for registration of 3-D shapes. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 1992. 14(2): p. 239-256. 9. Bradski, G. and A. Kaehler, Learning OpenCV: computer vision with the OpenCV library, O'Reilly. 10. W. Schroeder, K. Martin, B. Lorensen, The Visualization Toolkit - an object oriented approach to 3D graphics, 4th ed. Kitware; 2006.