Gas Distribution Modeling Using Sparse Gaussian Process Mixture Models

Similar documents
Introduction to Mobile Robotics

08 An Introduction to Dense Continuous Robotic Mapping

Introduction to Mobile Robotics

A Statistical Approach to Gas Distribution Modelling with Mobile Robots The Kernel DM+V Algorithm

Robot Mapping. A Short Introduction to the Bayes Filter and Related Models. Gian Diego Tipaldi, Wolfram Burgard

Clustering Lecture 5: Mixture Model

Practical Course WS12/13 Introduction to Monte Carlo Localization

CS 229 Midterm Review

arxiv: v2 [stat.ml] 5 Nov 2018

IROS 05 Tutorial. MCL: Global Localization (Sonar) Monte-Carlo Localization. Particle Filters. Rao-Blackwellized Particle Filters and Loop Closing

Autonomous Mobile Robot Design

Particle Filter in Brief. Robot Mapping. FastSLAM Feature-based SLAM with Particle Filters. Particle Representation. Particle Filter Algorithm

Gaussian Processes for Robotics. McGill COMP 765 Oct 24 th, 2017

Introduction to Mobile Robotics SLAM Grid-based FastSLAM. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello

Probabilistic Robotics

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

Particle Filters. CSE-571 Probabilistic Robotics. Dependencies. Particle Filter Algorithm. Fast-SLAM Mapping

Introduction to Machine Learning CMU-10701

Acoustic to Articulatory Mapping using Memory Based Regression and Trajectory Smoothing

10-701/15-781, Fall 2006, Final

PSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D.

A Bayesian Nonparametric Approach to Modeling Motion Patterns

Humanoid Robotics. Monte Carlo Localization. Maren Bennewitz

Introduction to Mobile Robotics SLAM Landmark-based FastSLAM

Comparing different interpolation methods on two-dimensional test functions

Probabilistic Robotics. FastSLAM

Applying Supervised Learning

Probabilistic Robotics

Expectation Maximization (EM) and Gaussian Mixture Models

Simultaneous Localization and Mapping

Mixture Models and EM

3. Data Structures for Image Analysis L AK S H M O U. E D U

An Artificial Potential Field based Sampling Strategy for a Gas-Sensitive Micro-Drone

Probabilistic Robotics

Domain Adaptation For Mobile Robot Navigation

Monte Carlo for Spatial Models

Spatial Interpolation & Geostatistics

Nearest Neighbor Predictors

Adaptive Sampling and Online Learning in Multi-Robot Sensor Coverage with Mixture of Gaussian Processes

Machine Learning Lecture 3

An Introduction To Automatic Tissue Classification Of Brain MRI. Colm Elliott Mar 2014

Spatial Interpolation - Geostatistics 4/3/2018

What is the SLAM problem?

Lecture 11: E-M and MeanShift. CAP 5415 Fall 2007

Probabilistic Robotics

Introduction to Objective Analysis

Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013

Pressure Drop Evaluation in a Pilot Plant Hydrocyclone

Artificial Intelligence for Robotics: A Brief Summary

Package gpr. February 20, 2015

Gradient Descent. Wed Sept 20th, James McInenrey Adapted from slides by Francisco J. R. Ruiz

Note Set 4: Finite Mixture Models and the EM Algorithm

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]

Generative and discriminative classification techniques

Uncertainties: Representation and Propagation & Line Extraction from Range data

ICRA 2016 Tutorial on SLAM. Graph-Based SLAM and Sparsity. Cyrill Stachniss

Evaluation of Moving Object Tracking Techniques for Video Surveillance Applications

What is machine learning?

Machine Learning Lecture 3

Introduction to ANSYS DesignXplorer

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

CS787: Assignment 3, Robust and Mixture Models for Optic Flow Due: 3:30pm, Mon. Mar. 12, 2007.

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History

Time Series Analysis by State Space Methods

CS249: ADVANCED DATA MINING

Discovery of the Source of Contaminant Release

Gas Source Declaration With a Mobile Robot

Introduction to Computational Mathematics

K-Means and Gaussian Mixture Models

This chapter explains two techniques which are frequently used throughout

Generative and discriminative classification techniques

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Challenges motivating deep learning. Sargur N. Srihari

The EM Algorithm Lecture What's the Point? Maximum likelihood parameter estimates: One denition of the \best" knob settings. Often impossible to nd di

Robot Mapping. TORO Gradient Descent for SLAM. Cyrill Stachniss

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves

Model learning for robot control: a survey

Gaussian Process Latent Variable Models for Visualisation of High Dimensional Data

ELEC Dr Reji Mathew Electrical Engineering UNSW

Learning from Data: Adaptive Basis Functions

Surface Smoothing Using Kriging

Chap.12 Kernel methods [Book, Chap.7]

Advanced Techniques for Mobile Robotics Graph-based SLAM using Least Squares. Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz

Random Forest A. Fornaser

Localization and Map Building

Functional Discretization of Space Using Gaussian Processes for Road Intersection Crossing

AUTONOMOUS SYSTEMS. PROBABILISTIC LOCALIZATION Monte Carlo Localization

CSC 2515 Introduction to Machine Learning Assignment 2

Last time... Bias-Variance decomposition. This week

Locally Weighted Learning for Control. Alexander Skoglund Machine Learning Course AASS, June 2005

Recap: Gaussian (or Normal) Distribution. Recap: Minimizing the Expected Loss. Topics of This Lecture. Recap: Maximum Likelihood Approach

Computational Fluid Dynamics (CFD) Simulation in Air Duct Channels Using STAR CCM+

Robot Mapping. Grid Maps. Gian Diego Tipaldi, Wolfram Burgard

Mixture of Gaussian Processes for Combining Multiple Modalities

Adaptive Dropout Training for SVMs

CHAPTER 1 INTRODUCTION

Machine Learning Lecture 3

Knowledge Discovery and Data Mining

Transcription:

Gas Distribution Modeling Using Sparse Gaussian Process Mixture Models Cyrill Stachniss, Christian Plagemann, Achim Lilienthal, Wolfram Burgard University of Freiburg, Germany & Örebro University, Sweden

Applications of Gas Distribution Modeling Oil refinery surveillance Garbage dump site surveillance Pollution monitoring in cities Air quality monitoring Garbage detection Rescue Robotics Disaster Prevention

GDM with Autonomous Sensor Networks Integration of mobile sensors positioned by human operators...... or by mobile robots accurate positioning useful for other tasks central integration into one consistent model model should allow to make prediction about observed and unobserved locations

Gas Dispersal (FIX ME) Turbulent Flow Characteristics turbulent transport is much faster than molecular diffusion gaseous ethanol at 25 C and 1 atm: diffusion constant: 0.119 cm 2 /s diffusion velocity: 20.7 cm/h turbulent flow is chaotic/unpredictable instantaneous velocity/concentration at some instant of time is generally insufficient to predict the velocity some time late high degree of vortical motion large-scale eddies cause a meandering dispersal small scale eddies stretch and twist the gas distributi resulting in a complicated patchy structure P.J.W. Roberts and D.R. Webster, "Turbulent Diffusion". In "Environmental Fluid Mechanics Theories and Application". ASCE Press, Reston, Virginia, 2002

Potential Environments

Statistical Gas Distribution Modeling Simulation of Turbulent Flow Characteristics? no general solution to the non-linear fluid dynamics equations numerical simulations computationally expensive and depend sensitively on the boundary conditions boundary conditions not known in typical scenarios model gas distribution statistically from a large number of measurements

Statistical Gas Distribution Modeling Interpret measurements in a statistical sense Build a representation of the observed gas distribution from a sequence of measurements Aimed at a probabilistic representation gas sensor measurements treated as random variables Is it a good model? Allows to infer concentration levels "explains observations best and accurately predict new ones" Allows to infer hidden parameters average concentrations gas source locations

Problem Definition Learn predictive model gas prediction query location measurement locations gas measurements Or: estimate a posterior over gas distribution models

Related Approaches Average concentration on a grid [Ishida et al. 98] Grid with Bi-cubic interpolation [Pyk et al. 06] Peak concentration estimates [Purnamadjaja & Russel 05] Kernel Extrapolation Algorithm [Lilienthal & Duckett 04] FastSLAM1 on grids combined with GDM [Lilienthal et al. 07] So far: no predictive uncertainty

Gaussian Processes (GPs) GPs are a framework for non-parametric regression Provides a predictive mean and variance A covariance function is used to specify the influence of neighboring data points The covariance function requires hyperparameters that need to be learned Cost of learning: matrix inversion O(n3) (n training samples) mean Cov(test,train) Cov(train) obs noise targets

GDM with Gaussian Processes Using all measurements is too expensive Use a subsampled set of observations (40 locations & gas measurements out of ~2500) Learn hyperparameters via cross-validation data collection mean prediction predictive variance

Observations Issues(?) in gas distribution rather smooth gas distributions away from hotspots localized packages of high concentration diffusion (?) GP: predictive uncertainty independent of the gas measurements TODO: ADD PLOT!

GP Mixture Models Each component is an individual GP A gating function specifies the influence of the individual models for each position In GDM: different components have the ability to represent the individual physical properties One GP component for the hotspots One GP component for the background GP mixture models first introduced by Tresp in 2000

GP Mixture Models Math x TODO

Learning the Mixture Model Selecting data points for training Initialization of the mixture components Learning the components Finding hyperparameters

Subsampling & Error GP Randomly draw data points Initialize the first component with drawn data Learn an Error GP Draw data points in areas of high error Initialize 2 nd component with these values

Learning the Components Learning is done in an EM-style procedure Determine which data point belongs to which components Recompute the GP components based on the new assignment/weight Used to model the influence of training data normal GP data w infinite noise

Learning the Components M-step E-step

Finding Good Hyperparameters Squared exponential covariance function noise: Optimization inside the M-Stpe of EM with Rasmussens s minimize Overfitting problems Sampling hyperparameters outside the EM Start with a good guess [Snelson & Ghahramani 06] Even iteration: sample random parameters Odd iteration: refine the best solution so far Cross-validation to avoid overfitting

1D Example IS THIS NEEDED?

Experiments 5 real world datasets TODO: ADD OTHER FOTOS!

Experiments (A) GP mean prediction GP predictive variance Mixture components Mixture gating function Mixture mean prediction Mixture predictive variance

Experiments (B) initial guess error GP 2 nd component After EM: mixture mean mixture predictive variance

Experiments (B 2D view) GP mean prediction GP predictive variance Mixture mean prediction Mixture predictive variance

Comparison GP mixture models Lilienthal & Duckett Comparable mean estimates but GPs provide also a predictive uncertainty

Comparison 100 data points used for training 1000 data points used to evaluate the sampled hyperparameter Rest of the data for testing Average neg. log likelihood TODO: ADD NLPD PLOT OVER TIME!

Runtime 100 training points per model 2 model components Pentium 2.4 GHz, Matlab code Learning one mixture model (all steps) takes < 1s 60 hyperparameter sampling runs overall runtime < 1 min Complexity O(training points^3 + test points^2)

Conclusion Gas distribution modeling with Gaussian process mixtures Uses a sparse set of training data Provides a predictive model Most alternative approaches do not provide predictive variances, GP models do so Outperforms standard GP model Efficiency/Complexity? <- Achim how fast is your stuff approx. Efficiency & Complexity

Future Work Non-stationary covariance functions Heteroscedastic GPs Robot action selection based on the model