Discovery of the Source of Contaminant Release
|
|
- Garry Merritt
- 5 years ago
- Views:
Transcription
1 Discovery of the Source of Contaminant Release Devina Sanjaya 1 Henry Qin Introduction Computer ability to model contaminant release events and predict the source of release in real time is crucial in various applications, especially in environmental safety monitoring and homeland security. In the event of unintentional industrial accidents or biological attacks in urban environments, immediate, accurate response is required. A real-time computer program that provides information about the identification of the contaminant, the source of the release, and the prediction of the subsequent path of contamination can be used to assist the decision making process for evacuation and countermeasures. The contaminant source inversion problem involves intricate geometry, uncertain flow conditions, and limited, noisy sensor readings. Moreover, the problem is generally ill-conditioned in the sense that small changes in the sensor readings can cause large changes in the calculated source of release [2]. This makes single-point deterministic calculations not robust. The lack of robustness is due to the fact that some inputs might produce nearly the same outputs, especially when measurement error is taken into account. To increase robustness, statistical approaches are used, but it often requires large sampling and numerous forward simulations which can quickly become computationally expensive. Multiple previous studies have been conducted to reduce the computational cost, such as grid coarsening, reduced-order modeling, and stochastic expansions. Previous studies have also considered applying uncertainty quantification methods to analyze the propagation of input uncertainties. In this paper, we combine machine learning models and computational fluid dynamics to discover the source of release for large-scale problems in real time. Various machine learning models are evaluated for robustness to noisy sensor readings and limited training data. We will also compare our results with statistical results from Markov Chain Monte Carlo (MCMC) with single walker [3] and ensemble walkers [1]. 2 Data Format Our training and test data are obtained using the computational fluid dynamics software (XFLOW) developed by Dr. Fidkowski at the University of Michigan: Ann Arbor. For the 2D case, we simulate a contaminant release around cross sections of buildings (see figure 1 (left)). Five sensors are placed around the buildings in a pseudorandom fashion without iteration or tuning. Each sensor takes three readings spaced equally in time, for a total of 15 sensor readings. A spatial approximation order of p = 2 is used and the Peclet number for the simulations, based on the mean velocity and domain size in the x-direction, is P e = 100. We use the 15 sensor readings as input features to our machine learning models, and X and Y as the output features. Each forward simulation used to obtain sensor readings was completed in less than 1 minute when parallelized with 8 processors. Figure 1: Mesh and setup used during CFD simulation for two (left) and 3D case (right). Sensor readings from both cases are used as our training and test sets. 1
2 For the 3D case, we simulate contaminant release in a realistic urban area (see figure 1 (right)); this domain is the same as in Lieberman et al [4]. There are 10 sensors placed around the buildings chosen in a pseudorandom fashion without iteration or tuning. Each sensor provides 4 readings for a total of 40 sensor readings. A spatial approximation order of p = 1 was used and the Peclet number for the simulations, based on the mean velocity and the domain extent in the direction of velocity, was Pe = 50. Our input features are the 40 sensor readings, and our outputs are the X, Y, Z, and Amplitude. Each forward simulation takes about 8 minutes on 100 processors. 3 Implementation & Discussion In this section, we will discuss how we implement machine learning models to our problem. All of our models are trained using the statistical programming language R or Matlab. For the purpose of the discussion below, we assume that we are only attempting to predict X, since predicting other variables (Y in 2D case and Y, Z, and Amplitude in 3D case) is symmetric. Moreover, to create realistic test cases, sensor errors are considered; we consider both uniform and Gaussian error distributions. Due to our time constraint, we will mainly discuss the results from the 2D case. 3.1 Test Error Definition All reported test errors are defined as a percentage of the interval over the parameter which we are making predictions. Equation 1) shows how we compute the test error when predicting X. %Test Error = x ˆx (max(x) min(x)) where x is the true value of X, ˆx is the predicted value of X, max(x) is the maximum of the interval of X, and min(x) is the minimum of the interval of X.In the 2D case, the X [0, 1], and in the 3D case, X [0, 4.71]. We believe this definition of test error percentage, rather than the standard definition of (Expected - Actual) / Actual, allows us to fairly compare predictions across examples with different true locations of the source of release. Intuitively, since we are trying to predict a location rather than a quantity, an error of 0.01 model units should be interpreted the same way whether it is an error from of a ground truth of 0.1 or a ground truth of 0.2, and our definition of error reflects this. Furthermore, this percentage error deifinition enables us to compare the test errors between 2D and 3D cases. 3.2 Perfect Sensor Readings First, we consider the case where all sensor readings are perfect. For our 2D test case, we have 144 examples in total: 72 for training and 72 for testing. We found that ordinary linear regression, which directly models the output values as a linear combination of the raw input feature values, does not perform well, with a mean error of 24%. However, linear regression with logarithmic feature mapping (see equation 2) performs quite well, with a mean error of 1%. This suggests that there is a log relationship between the sensor readings and the location of contaminant release. X = β 0 + β 1 log r 1 + β 2 log r β 15 log r 15 (2) To use log-transformed model, we first replace any readings that are less than equal to zero with the fixed constant , take the log of each sensor reading, and then apply multiple linear regression. Figure 2 (left) shows the residuals plot from our testing. This figure shows the mean error of 1% and standard deviation of 1% in predicting the source of release. Next, we consider the 3D case. Here, we have 256 examples in total: 200 examples for training and 56 examples for test. As with the 2D case, linear regression does not perform well, while linear regression with logarithmic feature mapping works well, with a mean error of 22.5%. Performing the same steps as before, we found a mean error of 0.26% and standard deviation of 0.26% as shown in Figure 2 (right). We acknowledge that these errors are unusually low, but do not claim that these errors will generalize to other 3D simulations, even assuming perfect sensors. (1) 2
3 Figure 2: Test residuals of linear regression with logarithmic feature mapping from the 2D (left) and 3D (right) cases with perfect sensor readings 3.3 Sensor Readings with Uniform Error Distribution Having successfully modeled the simple case, we moved onto a more complex case: uniform, substantial sensor error. Based on knowledge of the field, is a reasonable sensor error. To apply the uniform error, we add the same fixed constant to each of our sensor readings, and treat these perturbed sensor readings as the new raw features. During training, we naturally assume that the constant is unknown, as it would be in practice. Our training and test sets are the same as in the previous case. Unfortunately, this more complex case clearly demonstrated that our previous method was not robust against uniform sensor error, as our test errors increased by an order of magnitude. This behavior is consistent with the ill-conditioned nature of the inverse problem. In retrospect, we could have anticipated this. When systematic sensor error is added, the true model starts to look like the function in equation 3, while we were still trying to model it using equation 2. X = β 0 + β 1 log (r 1 + ɛ) + β 2 log (r 2 + ɛ) β 15 log (r 15 + ɛ) (3) We tried to model equation 3 by using R s nonlinear least squares nls package, but we ran into singularity problem. Since direct model fitting did not pan out, we implemented a hill-climbing algorithm in an attempt to greedily discover the value of the hidden constant ɛ. More specifically, our hill-climbing algorithm varies the value of ɛ to find the maximum R 2 statistic for a least squares fit against log (r i ɛ). The procedure is as follows: 1. Initialize a step size s to a constant Choose a random starting value ɛ 0 [1 10 2, ]. 3. Fit a least squares model using the features log (r i ɛ 0 ). 4. Fit two more least squares models using ɛ = ɛ 0 s and ɛ = ɛ 0 + s. 5. Set ɛ 0 to be equal the ɛ in the model above which produced the highest R 2 statistic. 6. Half the step size s, so s s/2. 7. If R 2 > 0.99, terminate the algorithm and report ɛ 0. Otherwise, return to Step 3. The above algorithm is able to pinpoint ɛ in the 2D case well, and thus, we can substitute ɛ to 3 before modeling the data. However, this hill-climbing algorithm does not work well in 3D case due to multiple local maxima. 3.4 Mixed Sensor Readings Now, we will consider the case where some sensor readings might happen to contain no errors while others have uniform or Gaussian-distributed errors. To simulate these errors, we first replicate the original set of examples three times, creating a new data set with three times the number of examples. Next, we add a constant error term to one full replica, add Gaussian-distributed errors (mean 0, stddev 1E 2) to the second full replica, and then randomize the order of the data 3
4 set. From this mixed data set, we randomly select half the examples for use in training, and hold out the other half as a test set. Multiple machine learning models are trained, such as linear regression, linear regression with logarithmic feature mapping, locally weighted linear regression with logarithmic feature mapping, decision tree, boosting, random forest, and K-nearest neighbors. Figure 3 compares the various models based on different testing error metrics (mean, standard deviation, median, and 90th percentile) for our 2D case. Here, we can see random forest gives us the lowest mean testing error, which is about 5% and the lowest 90th percentile error, which is about 10%. Figure 4 shows the test residuals of random forest for our 2D and 3D cases Linear Linear with Log Loc. Weighted Linear with Log Decision Tree Boosting Random Forest K nearest Neighbor 0.3 Testing Errors Mean StdDev Median 90% Error Metrics Figure 3: Testing error metrics of all machine learnings method applied to the 2D case with mixed sensor readings Figure 4: Test residuals of random forest applied on the 2D (left) and 3D (right) cases with mixed sensor readings. Note that these figures are not on the same scale because the 2D and 3D cases have different ranges for their dimensions. 4
5 3.5 Comparison with MCMC We compare our results with statistical results from MCMC with single and ensemble walkers presented in[5]. Using MCMC and noisy sensor readings of , we obtained less than 1% in predicting X for both 2D and 3D cases. Although the results from MCMC are far more accurate, it takes substantial time to obtain a single prediction from MCMC because generating a set of sensor readings at multiple proposed location of the MCMC walker(s) is time consuming. For instance, the 2D case converges in about 4 minutes using 32 processors and the 3D case converges in about 6 minutes using 100 processors, On the other hand, we can train a Random Forest and evaluate hundreds of examples in less than 1.5 seconds on a single processor. 4 Conclusion To summarize, we make the following contributions in this paper. With perfect sensor data, the relationship between sensor readings and the contaminant source is a simple log-linear one. Out of all the models we experimented with, Random Forest proved to be the most robust against noisy data. Comparing to MCMC, supervised learning requires far less computational power, but is less accurate and less robust to noisy data. To increase the robustness of supervised learning to noisy data, more research is required. To the best of our knowledge, predicting the location of contaminant release in a realistic setting remains an open problem. 5 Acknowledgement We gratefully acknowledge Dr. Fidkowski at the University of Michigan: Ann Arbor for the use of his computing resources and simulation software (XFLOW) in generating our training and test data. References [1] J. Goodman and J. Weare. Ensemble samplers with affine invariance. Communications in Applied Mathematics and Computational Science, 5(1):65 80, [2] J. Hadamard. Lectures on the Cauchy Problem in Linear Partial Differential Equations. Yale University Press, [3] W.K. Hastings. Monte Carlo sampling methods using Markov chains and their applications. Biometrika, 57(1):97 109, [4] C. Lieberman, K. Fidkowski, K. Willcox, and B. van Bloemen Waanders. Hessian-based model reduction: large-scale inversion and prediction. International Journal for Numerical Methods in Fluids, [5] D. Sanjaya, I. Tobasco, and K. Fidkowski. Adjoint-accelerated statistical and deterministic inversion of atmospheric contaminant transport. Unpublished. 5
Reddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011
Reddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011 1. Introduction Reddit is one of the most popular online social news websites with millions
More informationAdaptive spatial resampling as a Markov chain Monte Carlo method for uncertainty quantification in seismic reservoir characterization
1 Adaptive spatial resampling as a Markov chain Monte Carlo method for uncertainty quantification in seismic reservoir characterization Cheolkyun Jeong, Tapan Mukerji, and Gregoire Mariethoz Department
More informationRobustness analysis of metal forming simulation state of the art in practice. Lectures. S. Wolff
Lectures Robustness analysis of metal forming simulation state of the art in practice S. Wolff presented at the ICAFT-SFU 2015 Source: www.dynardo.de/en/library Robustness analysis of metal forming simulation
More informationInclusion of Aleatory and Epistemic Uncertainty in Design Optimization
10 th World Congress on Structural and Multidisciplinary Optimization May 19-24, 2013, Orlando, Florida, USA Inclusion of Aleatory and Epistemic Uncertainty in Design Optimization Sirisha Rangavajhala
More informationCSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM
CSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM 1 Introduction In this assignment you will implement a particle filter to localize your car within a known map. This will
More informationFMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu
FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)
More informationClustering Reduced Order Models for Computational Fluid Dynamics
Clustering Reduced Order Models for Computational Fluid Dynamics Gabriele Boncoraglio, Forest Fraser Abstract We introduce a novel approach to solving PDE-constrained optimization problems, specifically
More informationOverview. Monte Carlo Methods. Statistics & Bayesian Inference Lecture 3. Situation At End Of Last Week
Statistics & Bayesian Inference Lecture 3 Joe Zuntz Overview Overview & Motivation Metropolis Hastings Monte Carlo Methods Importance sampling Direct sampling Gibbs sampling Monte-Carlo Markov Chains Emcee
More informationMulti-Mesh CFD. Chris Roy Chip Jackson (1 st year PhD student) Aerospace and Ocean Engineering Department Virginia Tech
Multi-Mesh CFD Chris Roy Chip Jackson (1 st year PhD student) Aerospace and Ocean Engineering Department Virginia Tech cjroy@vt.edu May 21, 2014 CCAS Program Review, Columbus, OH 1 Motivation Automated
More informationarxiv: v1 [cs.cv] 2 May 2016
16-811 Math Fundamentals for Robotics Comparison of Optimization Methods in Optical Flow Estimation Final Report, Fall 2015 arxiv:1605.00572v1 [cs.cv] 2 May 2016 Contents Noranart Vesdapunt Master of Computer
More informationIntroduction to ANSYS DesignXplorer
Lecture 4 14. 5 Release Introduction to ANSYS DesignXplorer 1 2013 ANSYS, Inc. September 27, 2013 s are functions of different nature where the output parameters are described in terms of the input parameters
More informationTheoretical Concepts of Machine Learning
Theoretical Concepts of Machine Learning Part 2 Institute of Bioinformatics Johannes Kepler University, Linz, Austria Outline 1 Introduction 2 Generalization Error 3 Maximum Likelihood 4 Noise Models 5
More informationExperiments with Edge Detection using One-dimensional Surface Fitting
Experiments with Edge Detection using One-dimensional Surface Fitting Gabor Terei, Jorge Luis Nunes e Silva Brito The Ohio State University, Department of Geodetic Science and Surveying 1958 Neil Avenue,
More informationImproving Positron Emission Tomography Imaging with Machine Learning David Fan-Chung Hsu CS 229 Fall
Improving Positron Emission Tomography Imaging with Machine Learning David Fan-Chung Hsu (fcdh@stanford.edu), CS 229 Fall 2014-15 1. Introduction and Motivation High- resolution Positron Emission Tomography
More informationThe Comparative Study of Machine Learning Algorithms in Text Data Classification*
The Comparative Study of Machine Learning Algorithms in Text Data Classification* Wang Xin School of Science, Beijing Information Science and Technology University Beijing, China Abstract Classification
More informationHigh-Resolution Ocean Wave Estimation
DISTRIBUTION STATEMENT A. Approved for public release; distribution is unlimited. High-Resolution Ocean Wave Estimation David Walker SRI International 2100 Commonwealth Boulevard, Third Floor Ann Arbor,
More informationVariability in Annual Temperature Profiles
Variability in Annual Temperature Profiles A Multivariate Spatial Analysis of Regional Climate Model Output Tamara Greasby, Stephan Sain Institute for Mathematics Applied to Geosciences, National Center
More informationQuantitative Biology II!
Quantitative Biology II! Lecture 3: Markov Chain Monte Carlo! March 9, 2015! 2! Plan for Today!! Introduction to Sampling!! Introduction to MCMC!! Metropolis Algorithm!! Metropolis-Hastings Algorithm!!
More information1.2 Numerical Solutions of Flow Problems
1.2 Numerical Solutions of Flow Problems DIFFERENTIAL EQUATIONS OF MOTION FOR A SIMPLIFIED FLOW PROBLEM Continuity equation for incompressible flow: 0 Momentum (Navier-Stokes) equations for a Newtonian
More informationLocally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling
Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling Moritz Baecher May 15, 29 1 Introduction Edge-preserving smoothing and super-resolution are classic and important
More informationChapter 1. Introduction
Chapter 1 Introduction A Monte Carlo method is a compuational method that uses random numbers to compute (estimate) some quantity of interest. Very often the quantity we want to compute is the mean of
More informationParameter Estimation in Differential Equations: A Numerical Study of Shooting Methods
Parameter Estimation in Differential Equations: A Numerical Study of Shooting Methods Franz Hamilton Faculty Advisor: Dr Timothy Sauer January 5, 2011 Abstract Differential equation modeling is central
More information10703 Deep Reinforcement Learning and Control
10703 Deep Reinforcement Learning and Control Russ Salakhutdinov Machine Learning Department rsalakhu@cs.cmu.edu Policy Gradient I Used Materials Disclaimer: Much of the material and slides for this lecture
More informationExpectation-Maximization Methods in Population Analysis. Robert J. Bauer, Ph.D. ICON plc.
Expectation-Maximization Methods in Population Analysis Robert J. Bauer, Ph.D. ICON plc. 1 Objective The objective of this tutorial is to briefly describe the statistical basis of Expectation-Maximization
More informationLouis Fourrier Fabien Gaie Thomas Rolf
CS 229 Stay Alert! The Ford Challenge Louis Fourrier Fabien Gaie Thomas Rolf Louis Fourrier Fabien Gaie Thomas Rolf 1. Problem description a. Goal Our final project is a recent Kaggle competition submitted
More informationSamuel Coolidge, Dan Simon, Dennis Shasha, Technical Report NYU/CIMS/TR
Detecting Missing and Spurious Edges in Large, Dense Networks Using Parallel Computing Samuel Coolidge, sam.r.coolidge@gmail.com Dan Simon, des480@nyu.edu Dennis Shasha, shasha@cims.nyu.edu Technical Report
More informationDriven Cavity Example
BMAppendixI.qxd 11/14/12 6:55 PM Page I-1 I CFD Driven Cavity Example I.1 Problem One of the classic benchmarks in CFD is the driven cavity problem. Consider steady, incompressible, viscous flow in a square
More informationResearch on the New Image De-Noising Methodology Based on Neural Network and HMM-Hidden Markov Models
Research on the New Image De-Noising Methodology Based on Neural Network and HMM-Hidden Markov Models Wenzhun Huang 1, a and Xinxin Xie 1, b 1 School of Information Engineering, Xijing University, Xi an
More informationModeling Plant Succession with Markov Matrices
Modeling Plant Succession with Markov Matrices 1 Modeling Plant Succession with Markov Matrices Concluding Paper Undergraduate Biology and Math Training Program New Jersey Institute of Technology Catherine
More informationBootstrapping Method for 14 June 2016 R. Russell Rhinehart. Bootstrapping
Bootstrapping Method for www.r3eda.com 14 June 2016 R. Russell Rhinehart Bootstrapping This is extracted from the book, Nonlinear Regression Modeling for Engineering Applications: Modeling, Model Validation,
More informationCOSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor
COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality
More informationTemporal Modeling and Missing Data Estimation for MODIS Vegetation data
Temporal Modeling and Missing Data Estimation for MODIS Vegetation data Rie Honda 1 Introduction The Moderate Resolution Imaging Spectroradiometer (MODIS) is the primary instrument on board NASA s Earth
More informationCONDITIONAL SIMULATION OF TRUNCATED RANDOM FIELDS USING GRADIENT METHODS
CONDITIONAL SIMULATION OF TRUNCATED RANDOM FIELDS USING GRADIENT METHODS Introduction Ning Liu and Dean S. Oliver University of Oklahoma, Norman, Oklahoma, USA; ning@ou.edu The problem of estimating the
More informationStochastic Simulation: Algorithms and Analysis
Soren Asmussen Peter W. Glynn Stochastic Simulation: Algorithms and Analysis et Springer Contents Preface Notation v xii I What This Book Is About 1 1 An Illustrative Example: The Single-Server Queue 1
More informationOptimization of a two-link Robotic Manipulator
Optimization of a two-link Robotic Manipulator Zachary Renwick, Yalım Yıldırım April 22, 2016 Abstract Although robots are used in many processes in research and industry, they are generally not customized
More informationAn Introduction to Markov Chain Monte Carlo
An Introduction to Markov Chain Monte Carlo Markov Chain Monte Carlo (MCMC) refers to a suite of processes for simulating a posterior distribution based on a random (ie. monte carlo) process. In other
More informationLevel-set MCMC Curve Sampling and Geometric Conditional Simulation
Level-set MCMC Curve Sampling and Geometric Conditional Simulation Ayres Fan John W. Fisher III Alan S. Willsky February 16, 2007 Outline 1. Overview 2. Curve evolution 3. Markov chain Monte Carlo 4. Curve
More informationUncertainty Analysis Utilizing Gradient and Hessian Information
Uncertainty Analysis Utilizing Gradient and Hessian Information Markus P. Rumpfkeil, Wataru Yamazaki and Dimitri. Mavriplis Abstract In this paper gradient and Hessian information computed using automatic
More informationA Fast CMS Technique for Computational Efficient System Re-analyses in Structural Dynamics
Paper 23 A Fast CMS Technique for Computational Efficient System Re-analyses in Structural Dynamics D.-C. Papadioti and C. Papadimitriou Department of Mechanical Engineering University of Thessaly, Volos,
More informationACCURACY AND EFFICIENCY OF MONTE CARLO METHOD. Julius Goodman. Bechtel Power Corporation E. Imperial Hwy. Norwalk, CA 90650, U.S.A.
- 430 - ACCURACY AND EFFICIENCY OF MONTE CARLO METHOD Julius Goodman Bechtel Power Corporation 12400 E. Imperial Hwy. Norwalk, CA 90650, U.S.A. ABSTRACT The accuracy of Monte Carlo method of simulating
More informationMRF-based Algorithms for Segmentation of SAR Images
This paper originally appeared in the Proceedings of the 998 International Conference on Image Processing, v. 3, pp. 770-774, IEEE, Chicago, (998) MRF-based Algorithms for Segmentation of SAR Images Robert
More informationSaliency Detection in Aerial Imagery
Saliency Detection in Aerial Imagery using Multi-scale SLIC Segmentation Samir Sahli 1, Daniel A. Lavigne 2 and Yunlong Sheng 1 1- COPL, Image Science group, Laval University, Quebec, Canada 2- Defence
More informationMonte Carlo for Spatial Models
Monte Carlo for Spatial Models Murali Haran Department of Statistics Penn State University Penn State Computational Science Lectures April 2007 Spatial Models Lots of scientific questions involve analyzing
More informationRobotics. Lecture 5: Monte Carlo Localisation. See course website for up to date information.
Robotics Lecture 5: Monte Carlo Localisation See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review:
More informationCecil Jones Academy Mathematics Fundamental Map
Fundamentals Year 7 (Core) Knowledge Unit 1 Unit 2 Solve problems involving prime numbers Use highest common factors to solve problems Use lowest common multiples to solve problems Explore powers and roots
More informationMCMC Methods for data modeling
MCMC Methods for data modeling Kenneth Scerri Department of Automatic Control and Systems Engineering Introduction 1. Symposium on Data Modelling 2. Outline: a. Definition and uses of MCMC b. MCMC algorithms
More informationAdversarial Attacks on Image Recognition*
Adversarial Attacks on Image Recognition* Masha Itkina, Yu Wu, and Bahman Bahmani 3 Abstract This project extends the work done by Papernot et al. in [4] on adversarial attacks in image recognition. We
More informationDesign optimization and design exploration using an open source framework on HPC facilities Presented by: Joel GUERRERO
Workshop HPC Methods for Engineering CINECA (Milan, Italy). June 17th-19th, 2015. Design optimization and design exploration using an open source framework on HPC facilities Presented by: Joel GUERRERO
More informationRandom Search Report An objective look at random search performance for 4 problem sets
Random Search Report An objective look at random search performance for 4 problem sets Dudon Wai Georgia Institute of Technology CS 7641: Machine Learning Atlanta, GA dwai3@gatech.edu Abstract: This report
More informationHumanoid Robotics. Least Squares. Maren Bennewitz
Humanoid Robotics Least Squares Maren Bennewitz Goal of This Lecture Introduction into least squares Use it yourself for odometry calibration, later in the lecture: camera and whole-body self-calibration
More informationCommunication Networks Simulation of Communication Networks
Communication Networks Simulation of Communication Networks Silvia Krug 01.02.2016 Contents 1 Motivation 2 Definition 3 Simulation Environments 4 Simulation 5 Tool Examples Motivation So far: Different
More informationChapter 2 Basic Structure of High-Dimensional Spaces
Chapter 2 Basic Structure of High-Dimensional Spaces Data is naturally represented geometrically by associating each record with a point in the space spanned by the attributes. This idea, although simple,
More information1.1 Temporal and Spatial Reconstruction of Atmospheric Puff Releases using Bayesian Inference
. Temporal and Spatial Reconstruction of Atmospheric Puff Releases using Bayesian Inference Derek Wade and Inanc Senocak High Performance Simulation Laboratory for Thermo-fluids Department of Mechanical
More informationMachine Learning for Pre-emptive Identification of Performance Problems in UNIX Servers Helen Cunningham
Final Report for cs229: Machine Learning for Pre-emptive Identification of Performance Problems in UNIX Servers Helen Cunningham Abstract. The goal of this work is to use machine learning to understand
More informationStage 7 Checklists Have you reached this Standard?
Stage 7 Checklists Have you reached this Standard? Main Criteria for the whole year. J K L Use positive integer powers and associated real roots Apply the four operations with decimal numbers Write a quantity
More informationLandscape Ecology. Lab 2: Indices of Landscape Pattern
Introduction In this lab exercise we explore some metrics commonly used to summarize landscape pattern. You will begin with a presettlement landscape entirely covered in forest. You will then develop this
More informationAccelerometer Gesture Recognition
Accelerometer Gesture Recognition Michael Xie xie@cs.stanford.edu David Pan napdivad@stanford.edu December 12, 2014 Abstract Our goal is to make gesture-based input for smartphones and smartwatches accurate
More informationDesign of Experiments
Seite 1 von 1 Design of Experiments Module Overview In this module, you learn how to create design matrices, screen factors, and perform regression analysis and Monte Carlo simulation using Mathcad. Objectives
More informationOutlier Ensembles. Charu C. Aggarwal IBM T J Watson Research Center Yorktown, NY Keynote, Outlier Detection and Description Workshop, 2013
Charu C. Aggarwal IBM T J Watson Research Center Yorktown, NY 10598 Outlier Ensembles Keynote, Outlier Detection and Description Workshop, 2013 Based on the ACM SIGKDD Explorations Position Paper: Outlier
More informationUsing Machine Learning to Optimize Storage Systems
Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation
More informationA *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,
The structure is a very important aspect in neural network design, it is not only impossible to determine an optimal structure for a given problem, it is even impossible to prove that a given structure
More informationBayesian Estimation for Skew Normal Distributions Using Data Augmentation
The Korean Communications in Statistics Vol. 12 No. 2, 2005 pp. 323-333 Bayesian Estimation for Skew Normal Distributions Using Data Augmentation Hea-Jung Kim 1) Abstract In this paper, we develop a MCMC
More informationCSE 586 Final Programming Project Spring 2011 Due date: Tuesday, May 3
CSE 586 Final Programming Project Spring 2011 Due date: Tuesday, May 3 What I have in mind for our last programming project is to do something with either graphical models or random sampling. A few ideas
More informationSpatial Analysis and Modeling (GIST 4302/5302) Guofeng Cao Department of Geosciences Texas Tech University
Spatial Analysis and Modeling (GIST 4302/5302) Guofeng Cao Department of Geosciences Texas Tech University 1 Outline of This Week Last topic, we learned: Spatial autocorrelation of areal data Spatial regression
More informationNetwork Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Introduction Often, we have only a set of features x = x 1, x 2,, x n, but no associated response y. Therefore we are not interested in prediction nor classification,
More informationModified Metropolis-Hastings algorithm with delayed rejection
Modified Metropolis-Hastings algorithm with delayed reection K.M. Zuev & L.S. Katafygiotis Department of Civil Engineering, Hong Kong University of Science and Technology, Hong Kong, China ABSTRACT: The
More informationComparison of different solvers for two-dimensional steady heat conduction equation ME 412 Project 2
Comparison of different solvers for two-dimensional steady heat conduction equation ME 412 Project 2 Jingwei Zhu March 19, 2014 Instructor: Surya Pratap Vanka 1 Project Description The purpose of this
More informationCFD Project Workflow Guide
CFD Project Workflow Guide Contents Select a problem with known results for proof-of-concept testing... 1 Set up and run a coarse test case... 2 Select and calibrate numerical methods... 3 Minimize & quantify
More informationCHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM
20 CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM 2.1 CLASSIFICATION OF CONVENTIONAL TECHNIQUES Classical optimization methods can be classified into two distinct groups:
More informationLocal Features: Detection, Description & Matching
Local Features: Detection, Description & Matching Lecture 08 Computer Vision Material Citations Dr George Stockman Professor Emeritus, Michigan State University Dr David Lowe Professor, University of British
More informationINTRODUCTION. Model: Deconvolve a 2-D field of random numbers with a simple dip filter, leading to a plane-wave model.
Stanford Exploration Project, Report 105, September 5, 2000, pages 109 123 Short Note Test case for PEF estimation with sparse data II Morgan Brown, Jon Claerbout, and Sergey Fomel 1 INTRODUCTION The two-stage
More informationChapter 16. Microscopic Traffic Simulation Overview Traffic Simulation Models
Chapter 6 Microscopic Traffic Simulation 6. Overview The complexity of traffic stream behaviour and the difficulties in performing experiments with real world traffic make computer simulation an important
More informationEPSRC Centre for Doctoral Training in Industrially Focused Mathematical Modelling
EPSRC Centre for Doctoral Training in Industrially Focused Mathematical Modelling More Accurate Optical Measurements of the Cornea Raquel González Fariña Contents 1. Introduction... 2 Background... 2 2.
More informationRJaCGH, a package for analysis of
RJaCGH, a package for analysis of CGH arrays with Reversible Jump MCMC 1. CGH Arrays: Biological problem: Changes in number of DNA copies are associated to cancer activity. Microarray technology: Oscar
More informationModel Based Impact Location Estimation Using Machine Learning Techniques
Model Based Impact Location Estimation Using Machine Learning Techniques 1. Introduction Impacts on composite structures result in invisible damages that need to be detected and corrected before they lead
More informationLearning Adaptive Parameters with Restricted Genetic Optimization Method
Learning Adaptive Parameters with Restricted Genetic Optimization Method Santiago Garrido and Luis Moreno Universidad Carlos III de Madrid, Leganés 28911, Madrid (Spain) Abstract. Mechanisms for adapting
More informationBox-Cox Transformation for Simple Linear Regression
Chapter 192 Box-Cox Transformation for Simple Linear Regression Introduction This procedure finds the appropriate Box-Cox power transformation (1964) for a dataset containing a pair of variables that are
More informationGroveport Madison Local School District Third Grade Math Content Standards. Planning Sheets
Standard: Patterns, Functions and Algebra A. Analyze and extend patterns, and describe the rule in words. 1. Extend multiplicative and growing patterns, and describe the pattern or rule in words. 2. Analyze
More informationProceedings of the 2015 Winter Simulation Conference L. Yilmaz, W. K. V. Chan, I. Moon, T. M. K. Roeder, C. Macal, and M. D. Rossetti, eds.
Proceedings of the 2015 Winter Simulation Conference L. Yilmaz, W. K. V. Chan, I. Moon, T. M. K. Roeder, C. Macal, and M. D. Rossetti, eds. IMPROVING HIT-AND-RUN WITH SINGLE OBSERVATIONS FOR CONTINUOUS
More informationEE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm
EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm Group 1: Mina A. Makar Stanford University mamakar@stanford.edu Abstract In this report, we investigate the application of the Scale-Invariant
More informationFathom Dynamic Data TM Version 2 Specifications
Data Sources Fathom Dynamic Data TM Version 2 Specifications Use data from one of the many sample documents that come with Fathom. Enter your own data by typing into a case table. Paste data from other
More informationCS 229 Final Project - Using machine learning to enhance a collaborative filtering recommendation system for Yelp
CS 229 Final Project - Using machine learning to enhance a collaborative filtering recommendation system for Yelp Chris Guthrie Abstract In this paper I present my investigation of machine learning as
More informationUninformed Search Methods. Informed Search Methods. Midterm Exam 3/13/18. Thursday, March 15, 7:30 9:30 p.m. room 125 Ag Hall
Midterm Exam Thursday, March 15, 7:30 9:30 p.m. room 125 Ag Hall Covers topics through Decision Trees and Random Forests (does not include constraint satisfaction) Closed book 8.5 x 11 sheet with notes
More informationCS281 Section 9: Graph Models and Practical MCMC
CS281 Section 9: Graph Models and Practical MCMC Scott Linderman November 11, 213 Now that we have a few MCMC inference algorithms in our toolbox, let s try them out on some random graph models. Graphs
More informationHARNESSING CERTAINTY TO SPEED TASK-ALLOCATION ALGORITHMS FOR MULTI-ROBOT SYSTEMS
HARNESSING CERTAINTY TO SPEED TASK-ALLOCATION ALGORITHMS FOR MULTI-ROBOT SYSTEMS An Undergraduate Research Scholars Thesis by DENISE IRVIN Submitted to the Undergraduate Research Scholars program at Texas
More informationSimulation: Solving Dynamic Models ABE 5646 Week 12, Spring 2009
Simulation: Solving Dynamic Models ABE 5646 Week 12, Spring 2009 Week Description Reading Material 12 Mar 23- Mar 27 Uncertainty and Sensitivity Analysis Two forms of crop models Random sampling for stochastic
More informationIssues in MCMC use for Bayesian model fitting. Practical Considerations for WinBUGS Users
Practical Considerations for WinBUGS Users Kate Cowles, Ph.D. Department of Statistics and Actuarial Science University of Iowa 22S:138 Lecture 12 Oct. 3, 2003 Issues in MCMC use for Bayesian model fitting
More informationApplying Supervised Learning
Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains
More informationRecap Randomized Algorithms Comparing SLS Algorithms. Local Search. CPSC 322 CSPs 5. Textbook 4.8. Local Search CPSC 322 CSPs 5, Slide 1
Local Search CPSC 322 CSPs 5 Textbook 4.8 Local Search CPSC 322 CSPs 5, Slide 1 Lecture Overview 1 Recap 2 Randomized Algorithms 3 Comparing SLS Algorithms Local Search CPSC 322 CSPs 5, Slide 2 Stochastic
More informationGas Distribution Modeling Using Sparse Gaussian Process Mixture Models
Gas Distribution Modeling Using Sparse Gaussian Process Mixture Models Cyrill Stachniss, Christian Plagemann, Achim Lilienthal, Wolfram Burgard University of Freiburg, Germany & Örebro University, Sweden
More informationA noninformative Bayesian approach to small area estimation
A noninformative Bayesian approach to small area estimation Glen Meeden School of Statistics University of Minnesota Minneapolis, MN 55455 glen@stat.umn.edu September 2001 Revised May 2002 Research supported
More informationReliability - Based Robust Design Optimization of Centrifugal Pump Impeller for Performance Improvement considering Uncertainties in Design Variable
Reliability - Based Robust Design Optimization of Centrifugal Pump Impeller for Performance Improvement considering Uncertainties in Design Variable M.Balamurugan Assistant Professor, Government College
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods
Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationNeural Network and Deep Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina
Neural Network and Deep Learning Early history of deep learning Deep learning dates back to 1940s: known as cybernetics in the 1940s-60s, connectionism in the 1980s-90s, and under the current name starting
More informationVALERI 2003 : Concepcion site (Mixed Forest) GROUND DATA PROCESSING & PRODUCTION OF THE LEVEL 1 HIGH RESOLUTION MAPS
VALERI 2003 : Concepcion site (Mixed Forest) GROUND DATA PROCESSING & PRODUCTION OF THE LEVEL 1 HIGH RESOLUTION MAPS Marie Weiss 1 Introduction This report describes the production of the high resolution,
More informationNONPARAMETRIC REGRESSION TECHNIQUES
NONPARAMETRIC REGRESSION TECHNIQUES C&PE 940, 28 November 2005 Geoff Bohling Assistant Scientist Kansas Geological Survey geoff@kgs.ku.edu 864-2093 Overheads and other resources available at: http://people.ku.edu/~gbohling/cpe940
More informationStream Function-Vorticity CFD Solver MAE 6263
Stream Function-Vorticity CFD Solver MAE 66 Charles O Neill April, 00 Abstract A finite difference CFD solver was developed for transient, two-dimensional Cartesian viscous flows. Flow parameters are solved
More informationAllstate Insurance Claims Severity: A Machine Learning Approach
Allstate Insurance Claims Severity: A Machine Learning Approach Rajeeva Gaur SUNet ID: rajeevag Jeff Pickelman SUNet ID: pattern Hongyi Wang SUNet ID: hongyiw I. INTRODUCTION The insurance industry has
More informationClustering & Dimensionality Reduction. 273A Intro Machine Learning
Clustering & Dimensionality Reduction 273A Intro Machine Learning What is Unsupervised Learning? In supervised learning we were given attributes & targets (e.g. class labels). In unsupervised learning
More informationMapping of Hierarchical Activation in the Visual Cortex Suman Chakravartula, Denise Jones, Guillaume Leseur CS229 Final Project Report. Autumn 2008.
Mapping of Hierarchical Activation in the Visual Cortex Suman Chakravartula, Denise Jones, Guillaume Leseur CS229 Final Project Report. Autumn 2008. Introduction There is much that is unknown regarding
More information