3D Convolutional Neural Networks for Landing Zone Detection from LiDAR

Similar documents
Deep Learning with Tensorflow AlexNet

Multi-View 3D Object Detection Network for Autonomous Driving

Know your data - many types of networks

LiDAR Data Processing:

Deep Learning for Robust Normal Estimation in Unstructured Point Clouds. Alexandre Boulch. Renaud Marlet

PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. Charles R. Qi* Hao Su* Kaichun Mo Leonidas J. Guibas

Statistical Techniques in Robotics (16-831, F12) Lecture#05 (Wednesday, September 12) Mapping

POINT CLOUD DEEP LEARNING

Large-Scale Point Cloud Classification Benchmark

Object Localization, Segmentation, Classification, and Pose Estimation in 3D Images using Deep Learning

ECE 5470 Classification, Machine Learning, and Neural Network Review

Airborne Laser Scanning: Remote Sensing with LiDAR

Volumetric and Multi-View CNNs for Object Classification on 3D Data Supplementary Material

Statistical Techniques in Robotics (16-831, F10) Lecture#06(Thursday September 11) Occupancy Maps

Using Machine Learning for Classification of Cancer Cells

Generate Digital Elevation Models Using Laser Altimetry (LIDAR) Data

Deconvolution Networks

Improving 3-D 3 D processing: an efficient data structure and scale selection. Jean-François Lalonde Vision and Mobile Robotics Laboratory

3D Object Classification via Spherical Projections

Dynamic Routing Between Capsules

Deep Learning for Computer Vision II

An Introduction to Lidar & Forestry May 2013

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Keras: Handwritten Digit Recognition using MNIST Dataset

NATIONWIDE POINT CLOUDS AND 3D GEO- INFORMATION: CREATION AND MAINTENANCE GEORGE VOSSELMAN

Perceptron: This is convolution!

Scan Matching. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics

2 OVERVIEW OF RELATED WORK

Index. Springer Nature Switzerland AG 2019 B. Moons et al., Embedded Deep Learning,

Terrestrial GPS setup Fundamentals of Airborne LiDAR Systems, Collection and Calibration. JAMIE YOUNG Senior Manager LiDAR Solutions

Terrain Modeling and Mapping for Telecom Network Installation Using Scanning Technology. Maziana Muhamad

Where s the Boss? : Monte Carlo Localization for an Autonomous Ground Vehicle using an Aerial Lidar Map

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

Lecture 37: ConvNets (Cont d) and Training

LIDAR and Terrain Models: In 3D!

Simultaneous Localization and Mapping

PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017

Intro to Deep Learning. Slides Credit: Andrej Karapathy, Derek Hoiem, Marc Aurelio, Yann LeCunn

LIDAR MAPPING FACT SHEET

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.

Object Recognition II

Vehicle Localization. Hannah Rae Kerner 21 April 2015

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting

High Resolution Tree Models: Modeling of a Forest Stand Based on Terrestrial Laser Scanning and Triangulating Scanner Data

Supplementary A. Overview. C. Time and Space Complexity. B. Shape Retrieval. D. Permutation Invariant SOM. B.1. Dataset

2-4 April 2019 Taets Art and Event Park, Amsterdam CLICK TO KNOW MORE

Fuzzy Set Theory in Computer Vision: Example 3, Part II

CS 2750: Machine Learning. Neural Networks. Prof. Adriana Kovashka University of Pittsburgh April 13, 2016

Advanced Machine Learning

Introduction to Mobile Robotics Techniques for 3D Mapping

Recurrent Convolutional Neural Networks for Scene Labeling

Classification of objects from Video Data (Group 30)

Keras: Handwritten Digit Recognition using MNIST Dataset

Object Detection. CS698N Final Project Presentation AKSHAT AGARWAL SIDDHARTH TANWAR

ENY-C2005 Geoinformation in Environmental Modeling Lecture 4b: Laser scanning

Accelerating Convolutional Neural Nets. Yunming Zhang

INTRODUCTION TO DEEP LEARNING

08 An Introduction to Dense Continuous Robotic Mapping

2. POINT CLOUD DATA PROCESSING

FOR 274: Surfaces from Lidar. Lidar DEMs: Understanding the Returns. Lidar DEMs: Understanding the Returns

Automatic DTM Extraction from Dense Raw LIDAR Data in Urban Areas

Pedestrian Detection and Tracking in Images and Videos

Practical Methodology. Lecture slides for Chapter 11 of Deep Learning Ian Goodfellow

Overview. 1. Aerial LiDAR in Wisconsin (20 minutes) 2. Demonstration of data in CAD (30 minutes) 3. High Density LiDAR (20 minutes)

Algorithms for GIS csci3225

MULTI-LEVEL 3D CONVOLUTIONAL NEURAL NETWORK FOR OBJECT RECOGNITION SAMBIT GHADAI XIAN LEE ADITYA BALU SOUMIK SARKAR ADARSH KRISHNAMURTHY

3D-CNN and SVM for Multi-Drug Resistance Detection

Statistical Techniques in Robotics (STR, S15) Lecture#05 (Monday, January 26) Lecturer: Byron Boots

SIMULATED LIDAR WAVEFORMS FOR THE ANALYSIS OF LIGHT PROPAGATION THROUGH A TREE CANOPY

Deep Learning Cook Book

Light Detection and Ranging (LiDAR)

Automated Extraction of Buildings from Aerial LiDAR Point Cloud and Digital Imaging Datasets for 3D Cadastre - Preliminary Results

Lidar Sensors, Today & Tomorrow. Christian Sevcik RIEGL Laser Measurement Systems

1. INTRODUCTION 2. OPAL LIDAR OVERVIEW

HAWAII KAUAI Survey Report. LIDAR System Description and Specifications

Hide-and-Seek: Forcing a network to be Meticulous for Weakly-supervised Object and Action Localization

Spring Localization II. Roland Siegwart, Margarita Chli, Martin Rufli. ASL Autonomous Systems Lab. Autonomous Mobile Robots

Spring Localization II. Roland Siegwart, Margarita Chli, Juan Nieto, Nick Lawrance. ASL Autonomous Systems Lab. Autonomous Mobile Robots

Deep Learning Explained Module 4: Convolution Neural Networks (CNN or Conv Nets)

Machine Learning 13. week

Lecture 10 Dense 3D Reconstruction

Tutorial on Machine Learning Tools

Lecture: Autonomous micro aerial vehicles

Computer Graphics Ray Casting. Matthias Teschner

VEHICLE CLASSIFICATION And License Plate Recognition

CS468: 3D Deep Learning on Point Cloud Data. class label part label. Hao Su. image. May 10, 2017

Photo-realistic Renderings for Machines Seong-heum Kim

Ryerson University CP8208. Soft Computing and Machine Intelligence. Naive Road-Detection using CNNS. Authors: Sarah Asiri - Domenic Curro

Plankton Classification Using ConvNets

Artificial Intelligence Introduction Handwriting Recognition Kadir Eren Unal ( ), Jakob Heyder ( )

S7348: Deep Learning in Ford's Autonomous Vehicles. Bryan Goodman Argo AI 9 May 2017

EE795: Computer Vision and Intelligent Systems

Presented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey

Dynamic Routing Between Capsules. Yiting Ethan Li, Haakon Hukkelaas, and Kaushik Ram Ramasamy

Detection and Tracking of Moving Objects Using 2.5D Motion Grids

Light Field Super Resolution with Convolutional Neural Networks

Deep Learning For Video Classification. Presented by Natalie Carlebach & Gil Sharon

Transcription:

3D Convolutional Neural Networks for Landing Zone Detection from LiDAR Daniel Mataruna and Sebastian Scherer Presented by: Sabin Kafle

Outline Introduction Preliminaries Approach Volumetric Density Mapping Volumetric Convolutional Neural Nets Experiments Datasets Results

Introduction A system to detect a Landing Zone for autonomous helicopters is developed using 3D convolutional neural network and processed LiDAR point cloud as input. The system outperforms various benchmarks for small and potentially obscured obstacles detection in vegetated terrain.

System

Key challenges Engineering a system capable of reliably detecting landing safe site is difficult Appearance of vegetations, obstacles is heterogeneous and difficult to capture with hand build rules Noisiness and sparsity of point cloud data Machine learning system also has issues Maintenance of relevant information from large stream of data Obtaining Labelled data (is an expensive and laborious process)

Solutions offered Use 3D CNN with volumetric occupancy map which can perform real time safety evaluation once trained Use semi-synthetic point clouds to generate labelled data

Preliminaries LiDAR Convolutional Neural Network

LiDAR A remote sensing technology Measures distance by illuminating a target with laser and analyzing the reflected light Popular used technology to make highresolution maps A narrow laser-beam can map physical features with high resolutions; An aircraft can map terrain at 30cm resolution or better.

Fully-connected Neural Network (NN)

Convolutional Neural Networks

CNNs

Example Architecture INPUT [32x32x3] will hold the raw pixel values of the image CONV layer will compute the output of neurons that are connected to local regions in the input RELU layer will apply an elementwise activation function, such as the max (0,x)max(0,x) thresholding at zero. POOL layer will perform a downsampling operation along the spatial dimensions FC (i.e. fully-connected) layer will compute the class scores

Example

A single layer example An example input volume in red (e.g. a 32x32x3 image), and an example volume of neurons in the first Convolutional layer. Each neuron in the convolutional layer is connected only to a local region in the input volume spatially.

Examples A CNN Weight Update Weight Update A CNN demo on MNIST data MNIST

Approach Volumetric Density Mapping Volumetric CNNs

Volumetric Density Mapping The volumetric map covers a horizontal area of 100m 2 to 200m 2 and a height of 2 to 4m. CNN is used to independently predict the safety of 1m 3 subvolume within this map Space containing vegetation is porous Probabilistic formulation of this idea is used For a given grid cell coordinate (i,j,k), T sequence of point data that either hit (z = 1) or pass through the grid cell (z = 0) is used A voxel size of 0.05m 3 is used For each cell state, Beta Process is used with parameters (alpha and beta) initialized to 1 and updated as Alpha = Alpha + z Beta = Beta + (1 - z) Posterior mean is calculated based on (Alpha) / (Alpha + Beta)

Volumetric Density Mapping (contd..) Non-overlapping tiles of 1m 2 along XY-plane are generated Ground surface height for each tile is obtained using minima calculation of height of all occupied cells within Then 1m 3 volume is placed on each tile with center along ground surface height. Each grid has resolution of 0.05m 3, so each subvolume has dimension of 20 x 20 x 20. TO SUMMARIZE Mean calculated using Beta variable parameters are considered as pixel intensity of each individual sub-volume A pixel for this mapping is of volume 0.05m 3 The dimension of resulting CNN will have input of size 20 x 20 x 20.

Volumetric Convolutional Neural Network

Model Parameters Hyperparameters Number of convolutional layers Number of feature maps Dimension of features (dimension of local connectivity) Pooling Dimension Parameters Weight and Bias of each neurons (including convoluted and fully-connected) Theano Library is used for computing gradients and Training

Experiments Datasets Results

Datasets Synthetic Datasets Scanning pattern of LiDAR sensor is simulated by its pulse repetition rate, angular resolution and sensor mount behavior Gaussian Noise is added to range (based on manufacturer's specification) Scenes were scanned until point density of 3000 points/m 2 was reached First Synthetic Dataset Ground surface is a mesh with height perturbed using noise Grass blades are simulated using three triangle strip Grass placement is generated using Poisson process Box obstacles are placed Second Synthetic Dataset Box obstacles are replaced with 3D model selection 11 models of rock are used along with a tire and cinder blocks For each dataset, various parameters are swept, with 20 (first) instances and 40 (second) per parameter setting For each case, only half have an obstacle. Finally, generated data are augmented after being perturbed (to learn invariant classifier)

Datasets (contd..) Semi-synthetic Data Consists of real point cloud data for vegetation and ground combined with simulated data for obstacles Can be achieved simply by inserting virtual obstacles into the world frame and altering rays if they interfere Noise are added into rays in case of interference

Results Evaluation Metric Baselines Receiver Operating Characteristic (ROC) curves are used to compare different algorithms Random Forest Classifier Residuals of a robust plane fit Feature Bag of Words (BoW) [until recently considered state of art] Experiments The first synthetic data set is reserved to chose hyper-parameters for CNN. Two CNNs were chosen for final comparison C7-F64-P4-D512 C7-F32-P2-C5-F64-P2-D512

Results

Results

Misclassification

Future works Extend model to include more information including intensity, other source data etc. More extensive evaluations Current model might be overfitting to surface with grass

Questions? THANK YOU

References Stanford Neural network class http://cs231n.github.io/convolutionalnetworks/ Voxnet https://github.com/dimatura/voxnet