Model-based segmentation and recognition from range data

Similar documents
Detecting Geometric Primitives in 3D Data

Triangular Mesh Segmentation Based On Surface Normal

British Machine Vision Conference 2 The established approach for automatic model construction begins by taking surface measurements from a number of v

OBJECT RECOGNITION: OBTAINING 2-D RECONSTRUCTIONS FROM COLOR EDGES. G. Bellaire**, K. Talmi*, E. Oezguer *, and A. Koschan*

Extraction of Geometric Primitives from Point Cloud Data

Robotics Programming Laboratory

Absolute Scale Structure from Motion Using a Refractive Plate

Corner Detection. Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology

Semantic Object Recognition in Digital Images

Estimating normal vectors and curvatures by centroid weights

Cover Page. Abstract ID Paper Title. Automated extraction of linear features from vehicle-borne laser data

MULTIVIEW REPRESENTATION OF 3D OBJECTS OF A SCENE USING VIDEO SEQUENCES

Stereo and Epipolar geometry

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

SUBDIVISION ALGORITHMS FOR MOTION DESIGN BASED ON HOMOLOGOUS POINTS

Unwrapping of Urban Surface Models

Three-Dimensional Computer Vision

Short Survey on Static Hand Gesture Recognition

Micro-scale Stereo Photogrammetry of Skin Lesions for Depth and Colour Classification

from Range Data Edvaldo M. Bispo, Andrew W. Fitzgibbon and Robert B. Fisher Dept. of Articial Intelligence, University of Edinburgh,

Algorithm research of 3D point cloud registration based on iterative closest point 1

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

DRAFT: Analysis of Skew Tensegrity Prisms

Interpretation of Urban Surface Models using 2D Building Information Norbert Haala and Claus Brenner Institut fur Photogrammetrie Universitat Stuttgar

COMPUTER AND ROBOT VISION

A 3D Point Cloud Registration Algorithm based on Feature Points

3D object recognition used by team robotto

Texture Image Segmentation using FCM

3D Complex Scenes Segmentation from a Single Range Image Using Virtual Exploration

A 3-D Scanner Capturing Range and Color for the Robotics Applications

Flexible Calibration of a Portable Structured Light System through Surface Plane

Motion Detection Algorithm

Patch-based Object Recognition. Basic Idea

The Novel Approach for 3D Face Recognition Using Simple Preprocessing Method

3D Models and Matching

Segmentation of Range Data for the Automatic Construction of Models of Articulated Objects

Human Motion Detection and Tracking for Video Surveillance

Camera Calibration for a Robust Omni-directional Photogrammetry System

MORPHOLOGICAL BOUNDARY BASED SHAPE REPRESENTATION SCHEMES ON MOMENT INVARIANTS FOR CLASSIFICATION OF TEXTURES

Training-Free, Generic Object Detection Using Locally Adaptive Regression Kernels

Scale Space and PDE methods in image analysis and processing. Arjan Kuijper

Robust Pose Estimation using the SwissRanger SR-3000 Camera

Component-based Face Recognition with 3D Morphable Models

Using Geometric Constraints to Capture. design intent

A Survey of Light Source Detection Methods

Texture Mapping using Surface Flattening via Multi-Dimensional Scaling

Introduction to Computer Graphics. Modeling (3) April 27, 2017 Kenshi Takayama

Marcel Worring Intelligent Sensory Information Systems

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Simultaneous surface texture classification and illumination tilt angle prediction

AUTOMATED CALIBRATION TECHNIQUE FOR PHOTOGRAMMETRIC SYSTEM BASED ON A MULTI-MEDIA PROJECTOR AND A CCD CAMERA

Categorization by Learning and Combining Object Parts

Image Mining: frameworks and techniques

PART-LEVEL OBJECT RECOGNITION

GENERAL AUTOMATED FLAW DETECTION SCHEME FOR NDE X-RAY IMAGES

MULTIPLE-SENSOR INTEGRATION FOR EFFICIENT REVERSE ENGINEERING OF GEOMETRY

Component-based Face Recognition with 3D Morphable Models

VIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD. Ertem Tuncel and Levent Onural

Structure from Motion. Prof. Marco Marcon

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Towards the completion of assignment 1

Fingerprint Classification Using Orientation Field Flow Curves

Feature Based Registration - Image Alignment

Image processing and features

arxiv: v1 [cs.cv] 28 Sep 2018

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation

An Intelligent Clustering Algorithm for High Dimensional and Highly Overlapped Photo-Thermal Infrared Imaging Data

NIST. Support Vector Machines. Applied to Face Recognition U56 QC 100 NO A OS S. P. Jonathon Phillips. Gaithersburg, MD 20899

Bioimage Informatics

Object Recognition. The Chair Room

Compu&ng Correspondences in Geometric Datasets. 4.2 Symmetry & Symmetriza/on

Development of Reverse Engineering System for Machine Engineering Using 3D Bit-map Data. Tatsuro Yashiki* and Tarou Takagi*

DATA EMBEDDING IN TEXT FOR A COPIER SYSTEM

Impact of Intensity Edge Map on Segmentation of Noisy Range Images

In-plane principal stress output in DIANA

Operators-Based on Second Derivative double derivative Laplacian operator Laplacian Operator Laplacian Of Gaussian (LOG) Operator LOG

EE 701 ROBOT VISION. Segmentation

Occlusion Robust Multi-Camera Face Tracking

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Nearest Clustering Algorithm for Satellite Image Classification in Remote Sensing Applications

International Journal of Computer Science Trends and Technology (IJCST) Volume 3 Issue 1, Jan-Feb 2015

Experiments with Edge Detection using One-dimensional Surface Fitting

Automated Extraction of Buildings from Aerial LiDAR Point Cloud and Digital Imaging Datasets for 3D Cadastre - Preliminary Results

Similarity Image Retrieval System Using Hierarchical Classification

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries

Recognizing Deformable Shapes. Salvador Ruiz Correa Ph.D. UW EE

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Three-dimensional nondestructive evaluation of cylindrical objects (pipe) using an infrared camera coupled to a 3D scanner

3-D MRI Brain Scan Classification Using A Point Series Based Representation

3D from Photographs: Automatic Matching of Images. Dr Francesco Banterle

Classification and Generation of 3D Discrete Curves

Structured light 3D reconstruction

Advanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation

SEGMENTATION AND CLASSIFICATION OF POINT CLOUDS FROM DENSE AERIAL IMAGE MATCHING

arxiv: v1 [cs.cv] 28 Sep 2018

Transcription:

Model-based segmentation and recognition from range data Jan Boehm Institute for Photogrammetry Universität Stuttgart Germany Keywords: range image, segmentation, object recognition, CAD ABSTRACT This paper aims at developing a model-based system for the object recognition of three-dimensional objects with curved surfaces using range images. The model data is represented using a CAD-model, providing a mathematical precise and reliable description of arbitrary shapes. The proposed method is based on modelbased range image segmentation, using curvature as invariant features. By integrating model information into the segmentation stage, the segmentation process is guided to provide a partitioning corresponding to that of the CAD-model. The work provides a way to detect objects in arbitrary positions and derive the transformation onto a CAD-model. Thereby it contributes to the development of automated systems in the areas of inspection, manufacturing and robotics. 1. INTRODUCTION The availability of highly capable three-dimensional sensor systems creates the need for adequate systems for automated processing of three-dimensional data. Contrary to conventional intensity images, which contain information on the intensity of an objects surface, range images store the distance of each point on a surface to the sensor. In order to process these data on a high level of abstraction, new concepts need to be developed, exceeding two-dimensional image analysis. In the context of developing automated three-dimensional computer vision systems, this paper brings two aspects into focus: partitioning the image data and establishing a correspondence in-between image- and model-data. The stage of partitioning the sensed data, known as segmentation is crucial to the image understanding process. We employ a model-based strategy, where curvature information extracted from the model is used as a priori information for a classification of input pixels, which are grouped into regions. The features extracted during segmentation stage are brought into correspondence with the features of the model using a tree-search strategy. The number of possible pairings is reduced due to the fact that model information has been utilized during segmentation. After successful pairing, the three-dimensional transformation of the model onto the object in the scene is computed, completing the object recognition process. An overview of the complete process is given in figure 1. Our own range image data in the examples of this work were acquired with a triangulation sensor using a coded light approach. The system is described in Brenner et al. (1999). 1 However the application of the proposed method is not limited to this sensor type but can be adapted to any range acquisition device or measurement system, such as described in Boehm et al. (2001). 2 2. SEGMENTATION Segmentation is the process of partitioning an image into meaningful entities. What exactly meaningful means, depends on the specific application. In the context of object recognition segmentation is typically used as feature extraction. It shall therefore be robust and derive a partitioning sufficiently similar to that of the model. When range images are used it is common to compute a grouping of the pixels into regions, where each region represents a surface of the object. Many segmentation algorithms have been proposed in the literature, which work solely data-driven, see for example Hoover et al. (1996) 4 and Powell et al. (1998). 5 These approaches do not need any a priori knowledge

Datadescription Dataprocessing Dataacquisition on-line Matching Model Modelanalysis Modeldescription off-line Figure 1. Overview of the object recognition process in reference to Arman and Aggarwal (1993). 3 model analysis is incorporated into the data processing. Information of the of the objects shape. These approaches can be successful in cases where the partitioning of the boundary representation of an object is unique. This is particularly true for polyhedral objects. Also for objects which have sharp edges at the surface boundaries or for objects with a limited variety of surface types, such as scenes solely composed of cylinders, data driven approaches can be employed. However in the general case the partitioning of the boundary representation of an object is not unique (see for example Mortenson (1997) 6 ). For such a case a purely data driven approach is less suitable. Especially when segmentation is used as a feature extraction procedure for object recognition purposes, it is crucial that the partitioning of the range data coincides with that of the model. The information of the specific partitioning, which has been chosen for the model, is stored in its CAD model. It is therefore desirable to incorporate this information into the segmentation process. Since at the segmentation stage no relation is established in-between the coordinate system of the model and that of the sensed data, no information of the model can be used which depends on the chosen coordinate system. Features invariant to orientation and translation have to be used. This leads to the use of features based on differential geometry. We follow the work of Besl (1988) 7 for using mean and Gaussian curvature to describe the local properties of a surface. We compute the mean curvature as H = (1 + f 2 y )f xx +(1+f 2 x)f yy 2f x f y f xy 2(f 2 x + f 2 y +1) 3 2 and the Gaussian curvature as K = f xxf yy f 2 xy (f 2 x + f 2 y +1) 2 from the first and second derivatives of the surface function f. The surface function is a second order polynomial which is fit to the range data in a least squares estimation process. While Besl proposed to use the curvature information to classify each pixel of a range image into one of eight fundamental surface types, we extend the scheme to classify the pixels according to the model information. To extract this information we use the CAD system Pro/Engineer to extract the curvature information for each surface patch of the CAD model. This information is stored in a table and is used to define the classes for the classification process. The interface to the CAD system and the output for the displayed part are given in figure 2.

Figure 2. The interface to the Pro/Engineer CAD system and the corresponding curvature output. Surface type H K Plane 0.0000 0.0000 Cylinder 0.0200 0.0000 Sphere 0.0200 0.0004 Table 1. The table that holds the curvature information of the test scene. We use a simple nearest neighbor classification scheme in order to assign each pixel to its corresponding class. Pixels which have no corresponding class are assigned the background label. After initial classification each pixel is either labeled according to its corresponding surface or remains unlabeled. 2.1. Test Scene The proposed algorithm is tested on a simple test scene. The scene consists of a plane, a cylinder and a sphere, three basic surface types commonly encountered. An image of the test scene is given in figure 3. The model curvature information is given in the table 1. The information of this table graphed in the two-dimensional feature space spanned by mean and Gaussian curvature corresponds well to the curvature histogram of the acquired range data as is shown in figure 4. The results show a considerable amount of points which were misclassified, i.e. were assigned a wrong label or were not assigned a label but should have. These misclassifications are caused by false curvature estimation. Especially on the cylinder it becomes evident that not all of the pixels could be classified correctly. These classification errors have to be removed in a refined processing step. The most dominant regions, i.e. regions above a certain size threshold, are selected as seed regions for a region growing process. Region growing is implemented as a morphological operation. A 3 3 mask is moved over the dataset. When a neighbor to the point of interest (the center of the mask) has a label assigned, the point of interest is checked for compatibility to that region. In case it is found to be compatible it is assigned the label of the corresponding region. If there are conflicting regions, i.e. there are different regions adjacent to the point of interest, the largest region is preferred. This is also the case if the center pixel is already labeled. The results of the pure classification and of the refined processing are shown in figure 5. 2.2. SAMPL Test Data Additional tests of the classification algorithm were made based on data from the SAMPL database maintained by P. Flynn. 8 The database provides CAD models and range images for various simple parts. Most of them contain only cylindrical and planar surfaces and have few surfaces (less than ten). The data set we present here

Figure 3. Intensity image of the test scene consisting of three basic surface types. K 0.0004 0-0.02 0 H (a) Figure 4. Left: the model curvature information as a two-dimensional plot. Right: the curvature histogram of the corresponding range image. The points of accumulation in the histogram correspond well to the points in the model s feature space.

Figure 5. The results of the classification (left) and of the refined processing (right). (a) (b) (c) (d) Figure 6. (a) The CAD model of bigwye. (b)-(d) Segmentation results on the dataset. Segmentation of the surfaces is independent of the pose of the object. contains five cylinders and the according planar surfaces. Figure 6(a) shows this dataset called bigwye. Figure 6(b)-(d) show the result of classification. The overall results show that the surfaces can be reliably detected independent of the pose of the object. 3. MATCHING Once the surfaces have been segmented from the range data the correspondence of surfaces in the scene to surfaces of the model can be established. For this matching the set Ω of possible correspondences for each feature / region f i of the range image is established. We note that since we know the curvature characteristics of each of the surfaces, we can eliminate those correspondences which connect surfaces of different curvature characteristics. Therefore the set Ω is typically smaller than the set of all features / surfaces F i of the model. In figure 7 we show the results for one range image of the bigwye dataset. The size of the set of possible pairings ranges from one to three. This limits the size of the search space. The final search for correspondences is carried out by a tree search algorithm following the principles of Grimson. 9 The tree search results in several hypotheses for the pose of the object in the scene. The final verification of the match is performed by computing the rigid body transformation of the centroid p i of each region of the range image to the centroid P i of the corresponding model surface. The coordinates of all centroids

f 6 f 2 f 1 f 3 F 6 F 2 F 1 F 3 F 7 f 4 F 4 f 5 F 5 F 8 Ω(f 1 )={F 1 } Ω(f 2 )={F 2,F 3 } Ω(f 3 )={F 2,F 3 } Ω(f 4 )={F 4 } Ω(f 5 )={F 5 } Ω(f 6 )={F 6,F 7,F 8 } Figure 7. Deriving the set of possible correspondences from the segmentation result. The segmentation of the dataset bigwye resulted in six regions f 1 to f 6. The CAD model consists of eight surfaces F 1 to F 8. The set of possible correspondences for each of the six regions are given. are reduced by their mean value, which gives the centroids p i and P i respectively. The computation of the translation and rotation follows the approach of Horn. 10 The matrix M = N i=1 p ip T i = S xx S xy S xz S yx S yy S yz S zx S zy S zz gives the elements S xx = N i=1 p i x P i x, S xy = N i=1 p i x P i y and so on. These elements are used to form the symmetric 4 4 matrix N = (S xx + S yy + S zz ) S yz S zy S zx S xz S xy S yx S yz S zy (S xx S yy S zz ) S xy + S yx S zx + S xz S zx S xz S xy + S yx ( S xx + S yy S zz ) S yz + S zy S xy S yx S zx + S xz S yz + S zy ( S xx S yy + S zz ) The Eigenvector of this matrix corresponding to the largest Eigenvalue can be interpreted as a quaternion which contains the desired rotation R. The translation t is simply computed as t = p R P The RMS of this estimation process is used as criterion to select the best match. Figure 8 shows four instances of the bigwye dataset in different poses. For each pose the three best matches sorted by the RMS of the computed transformation are shown. We observe that the transformation computed with the smallest RMS corresponds to the correct pose for each of the cases.

rms =0.368 rms =0.790 rms =1.126 rms =0.307 rms =1.482 rms =2.245 rms =0.293 rms =0.929 rms =1.662 rms =0.320 rms =1.724 rms =2.266 Figure 8. Result of four different range images of the bigwye dataset. Each row shows one pose of the object. The first column shows the range image. The second, third and forth column show the CAD model in the computed transformation. The transformations are sorted by their RMS value. The best transformation is in the second column.

4. SUMMARY We have presented a method to detect objects in arbitrary positions and derive the transformation onto a corresponding CAD-model. The method employs an efficient technique for the model-based segmentation of range images. The proposed method is aimed at inspection and measurement task of industrial objects, but also has potential for a wide variety of applications in the areas of inspection, manufacturing and robotics. The topic of curvature estimation has not been mentioned within this paper, however it is still is a crucial part of the process and remains a topic of research. REFERENCES 1. C. Brenner, J. Böhm, and J. Gühring, Photogrammetric calibration and accuracy evaluation of a crosspattern stripe projector, Proc. SPIE Videometrics VI 3641, pp. 164 172, January 1999. 2. J. Böhm, J. Gühring, and C. Brenner, Implementation and calibration of a multi sensor measuring system, in Videometrics and Optical Methods for 3D Shape Measurement, S. El-Hakim and A. Grün, eds., Proceedings of SPIE 4309, pp. 154 161, SPIE, 2001. 3. F. Arman and J. K. Aggarwal, Model-based object recognition in dense range images, ACM Computing Surveys 25, pp. 5 43, March 1993. 4. A. Hoover, G. Jean-Baptiste, X. Jiang, P. Flynn, H. Bunke, D. Goldgof, K. Bowyer, D. Eggert, A. Fitzgibbon, and R. Fisher, An experimental comparison of range image segmentation algorithms, IEEE Transactions on Pattern Analysis and Machine Intelligence 18, pp. 673 689, July 1996. 5. M. W. Powell, K. W. Bowyer, X. Jiang, and H. Bunke, Comparing curved-surface range image segmenters, in International Conference on Computer Vision 1998, pp. 286 291, 1998. 6. M. E. Mortenson, Geometric Modeling, John Wiley & Sons, 1997. 7. P. J. Besl, Surfaces in Range Image Understanding, Springer series in perception engineering, Springer- Verlag, 1988. Revised ed. of Ph.D. thesis, University of Michigan. 8. R. J. Campbell and P. J. Flynn, A www-accessible database for 3d vision research, in Proc. IEEE Workshop on Empirical Evaluation Methods in Computer Vision, pp. 148 152, 1998. 9. W. E. L. Grimson, Object Recognition by Computer, MIT Press, 1990. 10. B. K. P. Horn, Closed-form solution of absolut orientation using unit quaternions, Journal of the Optical Society of America A 4, pp. 629 642, April 1987.