Methods for LiDAR point cloud classification using local neighborhood statistics

Similar documents
Monterey, CA, USA ABSTRACT 1. INTRODUCTION. phone ; fax ; nps.edu/rsc

Detecting trails in LiDAR point cloud data

ABSTRACT 1. INTRODUCTION

Terrain categorization using LIDAR and multi-spectral data

APPLICATION OF SOFTMAX REGRESSION AND ITS VALIDATION FOR SPECTRAL-BASED LAND COVER MAPPING

SIMULATED LIDAR WAVEFORMS FOR THE ANALYSIS OF LIGHT PROPAGATION THROUGH A TREE CANOPY

Automated Extraction of Buildings from Aerial LiDAR Point Cloud and Digital Imaging Datasets for 3D Cadastre - Preliminary Results

FOOTPRINTS EXTRACTION

AUTOMATIC BUILDING DETECTION FROM LIDAR POINT CLOUD DATA

SYNERGY BETWEEN AERIAL IMAGERY AND LOW DENSITY POINT CLOUD FOR AUTOMATED IMAGE CLASSIFICATION AND POINT CLOUD DENSIFICATION

Improvement of the Edge-based Morphological (EM) method for lidar data filtering

Lab 9. Julia Janicki. Introduction

REGISTRATION OF AIRBORNE LASER DATA TO SURFACES GENERATED BY PHOTOGRAMMETRIC MEANS. Y. Postolov, A. Krupnik, K. McIntosh

Data: a collection of numbers or facts that require further processing before they are meaningful

ABSTRACT 1. INTRODUCTION 2. DATA

An Introduction to Lidar & Forestry May 2013

Integrated analysis of Light Detection and Ranging (LiDAR) and Hyperspectral Imagery (HSI) data

[Youn *, 5(11): November 2018] ISSN DOI /zenodo Impact Factor

A Method to Create a Single Photon LiDAR based Hydro-flattened DEM

Airborne Laser Scanning: Remote Sensing with LiDAR

The Effect of Changing Grid Size in the Creation of Laser Scanner Digital Surface Models

EVOLUTION OF POINT CLOUD

Multi-ray photogrammetry: A rich dataset for the extraction of roof geometry for 3D reconstruction

Digital Image Classification Geography 4354 Remote Sensing

Chapters 1 7: Overview

Tools, Tips and Workflows Geiger-Mode LIDAR Workflow Review GeoCue, TerraScan, versions and above

Files Used in This Tutorial. Background. Feature Extraction with Example-Based Classification Tutorial

APPENDIX E2. Vernal Pool Watershed Mapping

HEURISTIC FILTERING AND 3D FEATURE EXTRACTION FROM LIDAR DATA

AUTOMATIC GENERATION OF DIGITAL BUILDING MODELS FOR COMPLEX STRUCTURES FROM LIDAR DATA

A QUALITY ASSESSMENT OF AIRBORNE LASER SCANNER DATA

Obstacle Detection From Roadside Laser Scans. Research Project

Classify Multi-Spectral Data Classify Geologic Terrains on Venus Apply Multi-Variate Statistics

Classification (or thematic) accuracy assessment. Lecture 8 March 11, 2005

DIGITAL IMAGE ANALYSIS. Image Classification: Object-based Classification

Overview. 1. Aerial LiDAR in Wisconsin (20 minutes) 2. Demonstration of data in CAD (30 minutes) 3. High Density LiDAR (20 minutes)

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

FOR 274: Surfaces from Lidar. Lidar DEMs: Understanding the Returns. Lidar DEMs: Understanding the Returns

CRF Based Point Cloud Segmentation Jonathan Nation

COMBINING HIGH SPATIAL RESOLUTION OPTICAL AND LIDAR DATA FOR OBJECT-BASED IMAGE CLASSIFICATION

Terrestrial Laser Scanning: Applications in Civil Engineering Pauline Miller

Spectral Classification

Figure 1: Workflow of object-based classification

A DATA DRIVEN METHOD FOR FLAT ROOF BUILDING RECONSTRUCTION FROM LiDAR POINT CLOUDS

WAVELET AND SCALE-SPACE THEORY IN SEGMENTATION OF AIRBORNE LASER SCANNER DATA

The Reference Library Generating Low Confidence Polygons

(Refer Slide Time: 0:51)

The Gain setting for Landsat 7 (High or Low Gain) depends on: Sensor Calibration - Application. the surface cover types of the earth and the sun angle

Ground and Non-Ground Filtering for Airborne LIDAR Data

2010 LiDAR Project. GIS User Group Meeting June 30, 2010

Copyright 2005 Society of Photo-Optical Instrumentation Engineers.

CO-REGISTERING AND NORMALIZING STEREO-BASED ELEVATION DATA TO SUPPORT BUILDING DETECTION IN VHR IMAGES

IMPROVED TARGET DETECTION IN URBAN AREA USING COMBINED LIDAR AND APEX DATA

BUILDING DETECTION AND STRUCTURE LINE EXTRACTION FROM AIRBORNE LIDAR DATA

Reality Check: Processing LiDAR Data. A story of data, more data and some more data

AUTOMATIC EXTRACTION OF LARGE COMPLEX BUILDINGS USING LIDAR DATA AND DIGITAL MAPS

Hyperspectral Processing II Adapted from ENVI Tutorials #14 & 15

doi: /

EXTRACTING SURFACE FEATURES OF THE NUECES RIVER DELTA USING LIDAR POINTS INTRODUCTION

Outline of Presentation. Introduction to Overwatch Geospatial Software Feature Analyst and LIDAR Analyst Software

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

An Approach for Combining Airborne LiDAR and High-Resolution Aerial Color Imagery using Gaussian Processes

Alaska Department of Transportation Roads to Resources Project LiDAR & Imagery Quality Assurance Report Juneau Access South Corridor

APPLICABILITY ANALYSIS OF CLOTH SIMULATION FILTERING ALGORITHM FOR MOBILE LIDAR POINT CLOUD

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

A COMPETITION BASED ROOF DETECTION ALGORITHM FROM AIRBORNE LIDAR DATA

ADVANCED TERRAIN PROCESSING: ANALYTICAL RESULTS OF FILLING VOIDS IN REMOTELY SENSED DATA TERRAIN INPAINTING

COMPONENTS. The web interface includes user administration tools, which allow companies to efficiently distribute data to internal or external users.

GEON Points2Grid Utility Instructions By: Christopher Crosby OpenTopography Facility, San Diego Supercomputer Center

Principal Component Image Interpretation A Logical and Statistical Approach

CHAPTER 5 OBJECT ORIENTED IMAGE ANALYSIS

Automatic DTM Extraction from Dense Raw LIDAR Data in Urban Areas

Remote Sensing Introduction to the course

FOR 474: Forest Inventory. Plot Level Metrics: Getting at Canopy Heights. Plot Level Metrics: What is the Point Cloud Anyway?

AUTOMATED 3-D FEATURE EXTRACTION FROM TERRESTRIAL AND AIRBORNE LIDAR

Lab 5: Image Analysis with ArcGIS 10 Unsupervised Classification

Object-Based Classification & ecognition. Zutao Ouyang 11/17/2015

Presented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey

Data Representation in Visualisation

N.J.P.L.S. An Introduction to LiDAR Concepts and Applications

Masking Lidar Cliff-Edge Artifacts

CLASSIFICATION FOR ROADSIDE OBJECTS BASED ON SIMULATED LASER SCANNING

COMPARISON OF TREE EXTRACTION FROM INTENSITY DROP AND FROM MULTIPLE RETURNS IN ALS DATA

Color Image Segmentation

Images Reconstruction using an iterative SOM based algorithm.

Introduction to digital image classification

LP360 (Advanced) Planar Point Filter

Identifying man-made objects along urban road corridors from mobile LiDAR data

POINT CLOUD SEGMENTATION FOR URBAN SCENE CLASSIFICATION

A MAXIMUM NOISE FRACTION TRANSFORM BASED ON A SENSOR NOISE MODEL FOR HYPERSPECTRAL DATA. Naoto Yokoya 1 and Akira Iwasaki 2

Generate Digital Elevation Models Using Laser Altimetry (LIDAR) Data

Terrain Modeling and Mapping for Telecom Network Installation Using Scanning Technology. Maziana Muhamad

Spatial Density Distribution

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Robotics Programming Laboratory

PIXEL VS OBJECT-BASED IMAGE CLASSIFICATION TECHNIQUES FOR LIDAR INTENSITY DATA

Light Detection and Ranging (LiDAR)

INCREASING CLASSIFICATION QUALITY BY USING FUZZY LOGIC

Using Statistical Techniques to Improve the QC Process of Swell Noise Filtering

Transcription:

Methods for LiDAR point cloud classification using local neighborhood statistics Angela M. Kim, Richard C. Olsen, Fred A. Kruse Naval Postgraduate School, Remote Sensing Center and Physics Department, 833 Dyer Road, Monterey, CA, USA ABSTRACT LiDAR data are available in a variety of publicly-accessible forums, providing high-resolution, accurate 3- dimensional information about objects at the Earth s surface. Automatic extraction of information from LiDAR point clouds, however, remains a challenging problem. The focus of this research is to develop methods for point cloud classification and object detection which can be customized for specific applications. The methods presented rely on analysis of statistics of local neighborhoods of LiDAR points. A multi-dimensional vector composed of these statistics can be classified using traditional data classification routines. Local neighborhood statistics are defined, and examples are given of the methods for specific applications such as building extraction and vegetation classification. Results indicate the feasibility of the local neighborhood statistics approach and provide a framework for the design of customized classification or object detection routines for LiDAR point clouds. Keywords: LiDAR, point cloud, classification, statistics, local neighborhood 1. INTRODUCTION It is clear from a visual inspection of LiDAR point cloud data that the 3-dimensional arrangement of LiDAR points is useful for distinguishing materials and objects in a scene. Characteristics of the distribution of points such as smoothness, regularity, and vertical scatter may help determine what type of surface or object the points represent. In this work, local neighborhoods of LiDAR points were examined and a series of statistics were defined to quantify these characteristics. Methods for defining local neighborhoods are discussed in Section 4, and a selection of local neighborhood statistics is discussed in Section 5. The statistical attributes were combined into feature vectors that were then used to segment the point cloud data into classes using traditional data classification routines. Not all attributes were useful for each scene classification task. The utility of individual statistical measures depends upon the characteristics of the particular LiDAR point cloud being examined, and the goal of the classification scheme. To determine which features were most useful for creating separable classes, the histograms of values for each statistic within a training region of interest were examined. Features that maximize the separability of classes were chosen for inclusion in the classification routine (Section 6). Two methods for classifying the LiDAR feature vectors, one supervised and one unsupervised, are discussed in Section 7. Results using both classification approaches and different combinations of statistical features are compared in Section 8. A summary of results and directions for future work are given in the final two sections (Sections 9 and 10). Further author information: A.M.K.: E-mail: amkim@nps.edu, Telephone: 1 401 647 3536 Laser Radar Technology and Applications XVIII, edited by Monte D. Turner, Gary W. Kamerman, Proc. of SPIE Vol. 8731, 873103 2013 SPIE CCC code: 0277-786X/13/$18 doi: 10.1117/12.2015709 Proc. of SPIE Vol. 8731 873103-1

2. PREVIOUS WORK Chehata et al. (2009) present a variety of statistics-based point cloud features and perform feature selection from LiDAR data using a random forests classification. 1 The ISPRS Commission III, Working Group 4, on 3D scene analysis provided a benchmark dataset for comparing methods of urban object classification and 3D building reconstruction. An accompanying report summarizes the success of the participants in classifying buildings and trees. 2 Using the benchmark dataset, Niemeyer et al. (2012) used an approach similar to Chehata et. al (2009), incorporating point cloud statistics, but instead made use of a conditional random field classification scheme to incorporate contextual information. 1, 3 Brodu and Lague (2012) present a multi-scale dimensionality approach for classifying terrestrial laser scanning data. 4 The method relies on calculating the principal components of the matrix of xyz-coordinates of local neighborhoods of points. By analyzing the percentage of variance explained by each principal component, the local neighborhood of points can be classified as being linear, planar, or globular. Looking at the dimensionality over varying spatial scales allows points to be classified. The approach for classifying LiDAR point cloud data using point cloud statistics based features presented in this paper is a very similar approach to that presented in Chehata et. al (2009) or Niemeyer et. al (2012). 1, 3 We use slightly different statistics, however, apply the multi-scale approach of Brodu and Lague (2012) to additional statistics, and use different methods for feature selection and classification. 4 3. THE STUDY AREA AND LIDAR DATASET LiDAR data were acquired for selected areas of California during 2010 by the Association of Monterey Bay Area Governments (AMBAG), via a USGS grant through the American Reinvestment and Recovery Act of 2009. The primary dataset used for this paper was collected for a portion of Camp Roberts (near Paso Robles, California) over multiple days in September, 2010, with a point density of 2.5 points per square meter. Bare-earth Digital Elevation Models (DEMs) with a resolution of 3-meter pixels were delivered with the data. Regions of interest were selected in 3D-space using the open source CloudCompare utility developed by Daniel Girardeau-Montaut. 5 This tool enables selection of tree points, while excluding LiDAR returns from ground underneath the tree. Similarly, the building ROI only contains rooftop points. Returns that may have come from the side of the building were excluded (Figure 1). Proc. of SPIE Vol. 8731 873103-2

4 At-, /%, :k".ª ª/; /. Figure 1. The sample dataset of a portion of Camp Roberts, California used in this research. The Google Earth image (Top Left) shows the approximate area covered by the LiDAR data. (Top Right) The LiDAR data colored according to point intensity (darker regions are shown in blue; brighter regions in green). (Bottom) A grayscale display of the LiDAR data with regions of interest points in color: building (red), tree (cyan), road (blue), and grass (green). Regions of interest were defined in 3D-space using the open source CloudCompare utility. 5 4. LOCAL NEIGHBORHOODS One of the challenges of working with LiDAR point cloud data is the irregularly-gridded nature of the data. In this work, we made use of a fast nearest neighbor searching algorithm provided by Merkwirth et. al. 6 This tool enables the rapid determination of local neighborhoods of points without requiring the data to be gridded. This approach maintains the accuracy of the original point cloud, and avoids the introduction of interpolation errors that must occur if the point cloud data is gridded into a traditional raster-type image. Two approaches were used for defining local neighborhoods of points. The first approach defines the neighborhood based on a set number of nearest neighbors. A minimum of three points is required for each neighborhood to ensure the local surface planar orientation can be calculated. Increasing the number of points included in the nearest neighbor search increases the computation time. In the results presented in this paper, we restricted the number of nearest neighbors to three. The second approach for defining a local neighborhood depends upon a user-defined radial distance. The radius defines a circular area in the x-y plane, and the neighborhood incorporates the vertical column of points at all heights. The radius must be large enough to incorporate a sufficient number of points for calculation of meaningful statistics, but not so large that spatial details are lost. In this work, with an average point density of 2.5 points per square meter, we considered multiple sizes of neighborhoods with the radii ranging from 1.2 to 3.7 meters. 5. POINT CLOUD ATTRIBUTES A series of LiDAR point attributes were examined, including the inherent LiDAR properties of intensity and height above ground level, and a set of statistics was designed to characterize the 3-dimensional arrangement of local neighborhoods of LiDAR points. Each of the individual attributes was recorded for all LiDAR point neighborhoods and then normalized so that each of the attribute had the same weight in the classification scheme. Proc. of SPIE Vol. 8731 873103-3

5.1 Height Above Ground Level and Intensity It is straightforward to make use of the height above ground level and return intensity values to classify LiDAR point cloud data. While the return intensity values are not typically normalized between flight lines, the relative intensity of points may be useful in distinguishing different materials. A DEM or some other estimate of the bare earth surface is required to determine a point s height above ground level. In this work, we made use of the 3-meter resolution DEMs that were delivered by the vendor. The fast nearest neighbor searching algorithm of Merkwirth et. al. was used to determine nearest DEM pixels to each LiDAR point. 6 The elevation of the nearest DEM pixel was subtracted from the elevation of the LiDAR point to determine the height above ground level of each point. 5.2 Differences of Elevations This statistic captures the average differences in elevations of neighborhoods of points. The statistic is calculated by subtracting the minimum elevation within the neighborhood from the mean of elevations of all points within a neighborhood. 5.3 Standard Deviations of Elevations This statistic is a useful measure of smoothness. This statistic is simply the standard deviation of elevations of all points within a neighborhood. 5.4 Standard Deviations of Intensity Values This statistic gives a measure of visual texture, and is calculated by finding the standard deviation of intensity values of all points within a neighborhood. 5.5 Angle of Planar Surface Normal In this paper, this statistic was only calculated for neighborhoods defined using a point s three nearest neighbors. The three nearest neighbors of each LiDAR point were used to define a plane. The deviation of the surface normal of the plane from vertical is recorded. 5.6 Principal Component Percentage of Variance This statistic is adapted from Brodu and Lague (2012). 4 For each neighborhood of points, the principal components of the matrix of xyz-coordinates is calculated. The percentage of the total variance explained by each principal component gives an indication of the dimensionality of the object. For example, to explain the location of a point on a linear object (such as a telephone wire), only a single measurement is needed. The distance in the direction of the axis aligned with the telephone wire is sufficient for defining the point s location. Therefore, a set of points representing a linear feature will have most of the variance in its xyz-coordinates explained by the first principal component. A planar surface, where points are spread in two dimensions, will require two components to define the location of points. The first two principal components will be necessary for explaining most of the variance in the data. Finally, if all three principal components are needed to represent a significant percentage of the variance, the object is 3-dimensional. Figures 2 and 3 show an example point cloud created to demonstrate this concept. The dimensionality of an object also depends upon the spatial scale at which it is examined. An object that looks planar close up (at a small spatial scale) may be 3-dimensional at a larger spatial scale. Further, this multi-scale behavior will vary for different materials. Calculating this statistic at multiple spatial scales is useful for distinguishing different types of materials. 5.7 Point Density The point density is determined by dividing the total number of points in a neighborhood by the square of the neighborhood radius. Proc. of SPIE Vol. 8731 873103-4

I I I I I Perspective Views of Sample Point Cloud with Linear, Planar, and 3D Features 1D Linear Feature 2D Planar Feature 3D Feature 100 50 o 15-10- 5- N 0-._'. 'Z: ;. : F1t. k 1= -20 100 50-50 X -100 50-50 -50 Y -100-100 -50 50 100-5- -10- -15 60. ' 40 20 0-20 -40-60 Figure 2. An example point cloud with linear (red), planar (green), and 3D objects (blue). 100 Percentage of Variance Explained by Each Principal Component 7, 80 ÉF c 8 60 d q 60 0 as 1... ÿÿ40 d a. 0.8 ', 0.6 lr 20 - o 1 2 Principal Component Number Variability of Feature Vectors f 1D Linear Feature e 2D Planar Feature 3D Feature Figure 3. The percentage of variance explained by each principal component for a linear (red), planar (green), and 3D object (blue). The linear feature requires only one principal component to explain all of the variance in the xyz-coordinates. 6. FEATURE SELECTION Many of the features described in Section 5 are similar for multiple classes of materials. In Figure 4, the shaded regions represent the variability in the feature vectors for each of the four regions of interest. This figure illustrates that the intraclass variability often exceeds the differences between classes. It is therefore necessary to choose a subset of features that maximize the class separability. 3 0.4 _ Gras Buil g 0.2 Mi Boad. 7i Tree z o `D 9 A h 0 4 6 0 10 15-1 19 1A 1h 1 ti4 4' 10 ry0 Attribute tig h4' ti 4'## 9h 9\ # # Figure 4. The mean feature vectors for each region of interest, with the variability of each class represented by the shaded region. The intraclass variability of the features often exceeds the differences between classes. Proc. of SPIE Vol. 8731 873103-5

The features as plotted in Figure 4 are as follows: 1. Differences of elevations (1.2-m neighborhood) 2. Differences of elevations (1.8-m neighborhood) 3. Differences of elevations (2.4-m neighborhood) 4. Differences of elevations (3-m neighborhood) 5. Differences of elevations (3.7-m neighborhood) 6. Differences of elevations (3-nearest neighbors) 7. Standard deviation of elevations (1.2-m) 8. Standard deviation of elevations (1.8-m) 9. Standard deviation of elevations (2.4-m) 10. Standard deviation of elevations (3-m) 11. Standard deviation of elevations (3.7-m) 12. Standard deviation of elevations (3-nearest neighbors) 13. Standard deviation of intensities (1.2-m) 14. Standard deviation of intensities (1.8-m) 15. Standard deviation of intensities (2.4-m) 16. Standard deviation of intensities (3-m) 17. Standard deviation of intensities (3.7-m) 18. Standard deviation of intensities (3-nearest neighbors) 19. Angle (3-nearest neighbors) 20. Percentage of variance - 1st Principal Component (PC) (1.2-m) 21. Percentage of variance - 1st PC (1.8-m) 22. Percentage of variance - 1st PC (2.4-m) 23. Percentage of variance - 1st PC (3-m) 24. Percentage of variance - 1st PC (3.7-m) 25. Percentage of variance - 2nd PC (1.2-m) 26. Percentage of variance - 2nd PC (1.8-m) 27. Percentage of variance - 2nd PC (2.4-m) 28. Percentage of variance - 2nd PC (3-m) 29. Percentage of variance - 2nd PC (3.7-m) 30. Percentage of variance - 3rd PC (1.2-m) 31. Percentage of variance - 3rd PC (1.8-m) 32. Percentage of variance - 3rd PC (2.4-m) 33. Percentage of variance - 3rd PC (3-m) 34. Percentage of variance - 3rd PC (3.7-m) 35. Point density (1.2-m) 36. Point density (1.8-m) 37. Point density (2.4-m) 38. Point density (3-m) 39. Point density (3.7-m) 40. Height above ground level 41. Intensity To select the most useful features for scene classification, the average distance between class means and the average standard deviation of values were measured for each attribute. The most useful attributes were chosen as those for which the average distance between class means exceeds the average standard deviation. For the sample dataset used in this paper, the attributes meeting this criterium are the following: 4 and 5: Differences of elevations (3-m and 3.7-m radius neighborhoods) 9-11: Standard deviation of elevations (2.4, 3, and 3.7-m radius neighborhoods) 16-18: Standard deviation of intensity values (3 and 3.7-m radius neighborhoods and 3-nearest neighbors neighborhood) 37-39: Point density (2.4, 3, 3.7-m radius neighborhoods) 41: Intensity Although in this case the point density features meet the criteria for being useful for separating the classes, it is unlikely that these features would be useful for scene classification, as the point density has more to do with collection geometry than any characteristic of the materials in the scene. A second subset of features was chosen using all of the separable features except for the point density features (statistics 4, 5, 9-11, 16-18 and 41). A third subset of features was defined subjectively by choosing features that appear to have distinct class means (statistics 4, 5, 8-11, 13, 15-18, 22, 25, 31, 32 and 41). Classification results using these subsets are compared to results of classification based on only the height above ground level and intensity information, and classification using all available features in Section 7. Proc. of SPIE Vol. 8731 873103-6

7. CLASSIFICATION METHODS Statistics of local neighborhoods were used to define a series of attributes for each LiDAR point. These attributes comprise a feature vector for each LiDAR point. The feature vectors were classified using traditional data processing techniques. 7 Both a supervised Spectral Angle Mapper (SAM) and an unsupervised KMEANS classification method were considered here. For the supervised SAM classification technique, training regions of interest (as shown in Figure 1) are used to design classification decision criteria. The SAM classifier assigns feature vectors to a class based on minimizing the angle between the feature vector and the mean feature vector of the training class. 8 The angle between vectors is calculated using the well known formula: ( ) A B Angle = arccos (1) A B A minimum angle threshold can be defined so that points that are not close to any of the training classes remain unclassified. A threshold of 0.05 radians was used for the work presented in this paper. For the unsupervised KMEANS classification method, the user must define the number of clusters to be used. Four clusters were used for this paper. The KMEANS algorithm randomly chooses cluster centers, and then assigns vectors to the nearest cluster center. The cluster centers are then updated, and the process is repeated until cluster centers stabilize, or a certain number of iterations have been completed. Because the initial cluster centers are randomly assigned, running the classifier more than once may produce different results. 9 8. COMPARISON OF RESULTS The KMEANS classifier is an unsupervised routine, with classes being randomly assigned. In the figures showing KMEANS classification results below, the colors do not necessarily correspond to particular classes, although some effort was made to match the KMEANS class colors to the SAM classes for purposes of visual comparison. For the SAM classifier, class colors correspond to the regions of interest as shown in Figure 1. Unclassified points are shown in black. 8.1 Classification Using Height Above Ground Level and Intensity It is a fairly standard practice to classify LiDAR data using the height above ground level and intensity information. Results of a supervised SAM classification and an unsupervised KMEANS classification are shown in Figure 5. The SAM classification results are very poor. The KMEANS method does a slightly better job of distinguishing the road from the dry grass covered ground, but trees and buildings are confused. s; Figure 5. Spectral angle mapper (SAM) supervised classification results (L) and KMEANS unsupervised classification results (R) using only height above ground level and return intensity information....17.% Proc. of SPIE Vol. 8731 873103-7

8.2 Classification Using All Available Attributes Results of the classification using all of the available LiDAR point attributes - statistical measures based on neighborhoods of points, along with the height above ground level and intensity information - are shown in Figure 6. In this case, the SAM and KMEANS classifier produce similar results, with a slightly cleaner result from the SAM classifier. For both classification methods, there is some confusion in the grass and road classes. Also, while buildings are mostly classified correctly, the edges of the buildings are confused with the tree class. %S 0. A C v } :+id '?:',i';.`.i : sn.im''+357' '..-t;. SA-;ïitaS.`Ì.P' :z1;41 11:4". jj L.r,`.? st 4. a.4ts, %- (, 14' vte'.- %tor Figure 6. Spectral angle mapper (SAM) supervised classification results (L) and KMEANS unsupervised classification results (R) using all available local neighborhood statistics, height above ground level, and return intensity information. 8.3 Classification Using Selected Subsets of Attributes It is more efficient, and with an appropriate selection - more accurate, to use a subset of attributes. As discussed in Section 6, an automated method of choosing a feature subset is to select those attributes for which the average distance between class means is greater than the average standard deviation of the class. Using this automated attribute selection method, twelve features were selected (4, 5, 9-11, 16-18, and 37-40). These features include attributes based on the differences in elevations, the standard deviation of elevations, standard deviation of intensity values, point density, and height above ground level. Results of classification using this subset of attributes are shown in Figure 7. While the point density attributes meet the criteria for the automated feature selection, they are not expected to be useful for scene classification in general, since the point density is dependent upon sensor collection geometry more than on any material characteristics. By comparing classification results incorporating the point density features (Figure 7), and classification results excluding these features (Figure 8), we can see that for this sample dataset, inclusion of the point density attributes improved the distinction of buildings and trees, but caused confusion among the road and grass classes. The automated method used for feature selection in this paper is very simple and not overly successful. A subjective choice of a subset of attributes that appear to have distinct class means produced better results (Figure 9). The subset includes statistics 4, 5, 8-11, 13, 15-18, 22, 25, 31, 32, and 41. These features include attributes based on differences in elevations, standard deviation of elevations, standard deviation of intensities, return intensity, and some of the principal component based features. In this case, the SAM routine successfully distinguished the four classes. Some confusion remained along the edges of buildings, with these points being incorrectly classified as belonging to the tree class. The KMEANS routine was also successful in distinguishing buildings from trees, but did a much poorer job of separating the road and grass classes. SV Proc. of SPIE Vol. 8731 873103-8

' ry;. ß1?5. r, gemdds,..4,0, ;P. Figure 7. Spectral angle mapper (SAM) supervised classification results (L) and KMEANS unsupervised classification results (R) using attributes chosen by automatic threshold, including some attributes based on point density. 40'4 0:,4111 o". Figure 8. Spectral angle mapper (SAM) supervised classification results (L) and KMEANS unsupervised classification results (R) using attributes chosen by automatic threshold without any point density attributes. +4* V-Vcre. 4.1.1"..e V. 1/4... Ktl..4 Figure 9. Spectral angle mapper (SAM) supervised classification results (L) and KMEANS unsupervised classification results (R) using a subjective choice of attributes that appear to have distinct class means. i 9. SUMMARY AND CONCLUSIONS A series of LiDAR point cloud attributes based on statistics of local neighborhoods of points were presented. These attributes were combined into feature vectors for each LiDAR point. The feature vectors were then Proc. of SPIE Vol. 8731 873103-9

classified using traditional image processing techniques, including a supervised spectral angle mapper (SAM) classification and an unsupervised KMEANS classification. A simple method was presented for automated feature selection based on finding features that have an average distance between class means greater than the average standard deviation of the classes. This simple method of feature selection does not lead to a very successful classification result. A subjectively chosen subset of features based on choosing features that appear to have distinct class means was found to provide the best classification result. Results indicate that inclusion of the local neighborhood statistical attributes leads to an improved discrimination of materials over a classification using only height above ground level and intensity information. 10. FUTURE WORK The methods presented here are sensitive to local neighborhood size. The best method for choosing an appropriate neighborhood size is an ongoing area of research. It is dependent upon the size of the features to be distinguished, the density of the LiDAR point cloud, and the desired speed of processing. Feature selection affects the efficiency of processing and the accuracy of the classification result. Subjectively choosing the features which had the most distinct training class means worked well in this case, but an automated method for feature selection would be preferable. Application of the methods presented here to a dataset with ground truth, such as the benchmark dataset provided by the ISPRS Commission III, Working Group 4, will allow for a quantitative assessment of the method. 2 REFERENCES [1] Chehata, N., Guo, L., and Mallet, C., Airborne LiDAR feature selection for urban classification using random forests, Laser Scanning 2009, IAPRS 38(Part 3/W8), 207 212 (2009). [2] Rottensteiner, F., Sohn, G., Jung, J., Gerke, M., Baillard, C., Benitez, S., and Breitkopf, U., The ISPRS benchmark on urban object classification and 3D building reconstruction, ISPRS Commission III, WGIII/4, 1 6 (2012). [3] Niemeyer, J., Rottensteiner, F., and Soergel, U., Conditional random fields for LiDAR point cloud classification in complex urban areas, ISPRS Annals of Photogrammetry, Remote Sensing and Spatial Information Sciences I-3, 263 268 (2012). [4] Brodu, N. and Lague, D., 3D terrestrial lidar data classification of complex natural scenes using a multi-scale dimensionality criterion: Applications in geomorphology, ISPRS Journal of Photogrammetry and Remote Sensing 68(C), 121 134 (2012). [5] Girardeau-Montaut, D., CloudCompare Open Source Project, http://www.danielgm.net/cc/. [6] Merkwirth, C., Parlitz, U., Wedekind, I., Engster, D., and Lauterborn, W., OpenTSTOOL User Manual, www.physik3.gwdg.de/tstool/. [7] Richards, J. A. and Jia, X., [Remote Sensing Digital Image Analysis An Introduction], Springer, 4 ed. (2006). [8] Kruse, F. A., Lefkoff, A. B., Boardman, J. B., Shapiro, A. T., Barloon, P. J., and Goetz, A. F. H., The Spectral Image Processing System (SIPS) Interactive Visualization and Analysis of Imaging Spectrometer Data, Remote Sensing of Environment 44, 144 163 (1993). [9] Seber, G. A. F., [Multivariate Observations], John Wiley & Sons, Inc. (1984). Proc. of SPIE Vol. 8731 873103-10