ISyE 6416: Computational Statistics Spring Lecture 13: Monte Carlo Methods

Similar documents
STAT 725 Notes Monte Carlo Integration

Monte Carlo Methods and Statistical Computing: My Personal E

Chapter 1. Introduction

MULTI-DIMENSIONAL MONTE CARLO INTEGRATION

DEDICATIONS. To my dear parents, and sisters. To my advisor Dr. Mahmoud Alrefaei. To my friends in and out the university.

A Random Number Based Method for Monte Carlo Integration

ACCURACY AND EFFICIENCY OF MONTE CARLO METHOD. Julius Goodman. Bechtel Power Corporation E. Imperial Hwy. Norwalk, CA 90650, U.S.A.

Chapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization. Wolfram Burgard

GAMES Webinar: Rendering Tutorial 2. Monte Carlo Methods. Shuang Zhao

Markov chain Monte Carlo methods

Monte Carlo Integration

In the real world, light sources emit light particles, which travel in space, reflect at objects or scatter in volumetric media (potentially multiple

Homework: Study 6.1 # 1, 5, 7, 13, 25, 19; 3, 17, 27, 53

Optimization and Simulation

VARIANCE REDUCTION TECHNIQUES IN MONTE CARLO SIMULATIONS K. Ming Leung

Lab 5 Monte Carlo integration

Physics 736. Experimental Methods in Nuclear-, Particle-, and Astrophysics. - Statistical Methods -

Simulation. Monte Carlo

Numerical Integration

1. Show that the rectangle of maximum area that has a given perimeter p is a square.

Monte Carlo Integration

Use of Extreme Value Statistics in Modeling Biometric Systems

Laplace Transform of a Lognormal Random Variable

Topic 5 - Joint distributions and the CLT

Monte Carlo Integration of The Rendering Equation. Computer Graphics CMU /15-662, Spring 2017

Monte Carlo Ray Tracing. Computer Graphics CMU /15-662

Bayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19

Physics 736. Experimental Methods in Nuclear-, Particle-, and Astrophysics. Lecture 14

Lecture 14. December 19, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Overview. Monte Carlo Methods. Statistics & Bayesian Inference Lecture 3. Situation At End Of Last Week

Physics 736. Experimental Methods in Nuclear-, Particle-, and Astrophysics. - Statistics and Error Analysis -

Monte Carlo Integration

VARIOUS ESTIMATIONS OF π AS DEMONSTRATIONS OF THE MONTE CARLO METHOD

B. Examples Set up the integral(s) needed to find the area of the region bounded by

Probabilistic Robotics

The Plan: Basic statistics: Random and pseudorandom numbers and their generation: Chapter 16.

Computer Experiments. Designs

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit IV Monte Carlo

Notice that the height of each rectangle is and the width of each rectangle is.

The Cross-Entropy Method

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

Computational Methods for Finding Probabilities of Intervals

Optimization Methods III. The MCMC. Exercises.

Today. Lecture 4: Last time. The EM algorithm. We examine clustering in a little more detail; we went over it a somewhat quickly last time

The Cross-Entropy Method for Mathematical Programming

Application of Importance Sampling in Simulation of Buffer Policies in ATM networks

Chapter 6 Normal Probability Distributions

Integration. Volume Estimation

An Introduction to Markov Chain Monte Carlo

Probability Models.S4 Simulating Random Variables

Biostatistics 615/815 Lecture 16: Importance sampling Single dimensional optimization

Computational Methods. Randomness and Monte Carlo Methods

Nested Sampling: Introduction and Implementation

APPENDIX K MONTE CARLO SIMULATION

Advanced Statistical Computing Week 2: Monte Carlo Study of Statistical Procedures

Motivation. Monte Carlo Path Tracing. Monte Carlo Path Tracing. Monte Carlo Path Tracing. Monte Carlo Path Tracing

Computer Vision Group Prof. Daniel Cremers. 8. Boosting and Bagging

Stochastic Simulation: Algorithms and Analysis

METROPOLIS MONTE CARLO SIMULATION

Probabilistic (Randomized) algorithms

MATH3016: OPTIMIZATION

Bayesian Statistics Group 8th March Slice samplers. (A very brief introduction) The basic idea

Chapter 4: Non-Parametric Techniques

Lecture 8: Jointly distributed random variables

Introduction to Design Optimization: Search Methods

Monte Carlo Simula/on and Copula Func/on. by Gerardo Ferrara

Convexization in Markov Chain Monte Carlo

1-7 Inverse Relations and Functions

Monte Carlo Methods. Lecture slides for Chapter 17 of Deep Learning Ian Goodfellow Last updated

Quantitative Biology II!

10.4 Linear interpolation method Newton s method

Assessing the Quality of the Natural Cubic Spline Approximation

Sampling and Monte-Carlo Integration

B.Stat / B.Math. Entrance Examination 2017

11-2 Probability Distributions

Monte Carlo Integration COS 323

Homework 2. Due: March 2, 2018 at 7:00PM. p = 1 m. (x i ). i=1

Extracting Coactivated Features from Multiple Data Sets

Estimating the Information Rate of Noisy Two-Dimensional Constrained Channels

Instance-based Learning CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2015

Constrained and Unconstrained Optimization

Motivation: Monte Carlo Path Tracing. Sampling and Reconstruction of Visual Appearance. Monte Carlo Path Tracing. Monte Carlo Path Tracing

STA215 Inference about comparing two populations

Introduction to Functions

Simulation. Chapter Copyright 2010 Pearson Education, Inc. Publishing as Prentice Hall

MSA101/MVE Lecture 5

Advanced Graphics. Path Tracing and Photon Mapping Part 2. Path Tracing and Photon Mapping

Unit 8 SUPPLEMENT Normal, T, Chi Square, F, and Sums of Normals

13 Distribution Ray Tracing

Learner Expectations UNIT 1: GRAPICAL AND NUMERIC REPRESENTATIONS OF DATA. Sept. Fathom Lab: Distributions and Best Methods of Display

Monte Carlo Integration COS 323

Chapter 15: Functions of Several Variables

Quasi-Monte Carlo Methods Combating Complexity in Cost Risk Analysis

Tin Ka Ping Secondary School F.3 Mathematics Teaching Syllabus

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Statistics and Computer Application

Monte Carlo Simulation. Ben Kite KU CRMDA 2015 Summer Methodology Institute

CSE446: Linear Regression. Spring 2017

Recursive Estimation

Transcription:

ISyE 6416: Computational Statistics Spring 2017 Lecture 13: Monte Carlo Methods Prof. Yao Xie H. Milton Stewart School of Industrial and Systems Engineering Georgia Institute of Technology

Determine area under a curve determine area under a curve y = h(x), continuous function with 0 h(x) M over the closed interval a x b select points at random from within the rectangular region area under the curve / area of rectangle number of points counted below curve / total number of random points

Monte Carlo integration To compute E[h(X)] = h(x)f(x)dx h(x): integrand, f(x): probability density function Monte Carlo approach to approximate the integration: Take i.i.d. samples X 1,..., X n from pdf f(x). Then take sample average Î 1 = 1 n n h(x i ) i=1 Convergence: law of large numbers, as n 1 n n h(x i ) E[h(X)] i=1

Monte Carlo methods Monte Carlo methods (or Monte Carlo experiments) are a broad class of computational algorithms that rely on repeated random sampling to obtain numerical results. most useful when it is difficult or impossible to use other approaches. Monte Carlo methods are mainly used in three distinct problem classes: optimization, numerical integration, and generating draws from a probability distribution.

Estimating probabilities Evaluating probability can be rewritten using indicator function as P(X A) = I A (x) Widely used in engineering: example, wireless multi-hop networks There are m 3 nodes randomly distributed in the unit square Each node can communicate directly with any other node that is within a distance r of it A two-hop connection arises between nodes 1 and 2 when nodes 1 and 2 are farther than r but are both within distance r of node j for one or more j {3,..., m} Source: A. Owen, Monte Carlo theory, methods and examples.

m = 40 Question: What is the probability of forming a two-way connection? This probability is an integral over 2m = 40 dimensional space Run 10,000 independent replications of this problem Among them, 650 cases have two-hop connection The probability is estimated to be ˆp = 650/10000 = 0.065.

How good is the estimation? CLT-based 99% confidence interval for p ˆp(1 ˆp) leading to ˆp ± z 0.005 }{{} 2.58 n 0.065 0.945 0.065 ± 2.58 = 0.065 ± 0.0064 10000

Estimating tail probability How strong a wind should we construct a building to be resistant to? P(W w) = exp( exp((w 52)/4)) Source: A. Owen, Monte Carlo theory, methods and examples.

Estimating tail probability Suppose the random variable X is binomially distributed with m trials and success probability p. We want to evaluate the right-tail probability α = P{X z} evaluating tail probability is important in hypothesis testing: finding the significance level of the test, we need to evaluate the tail probability of the test statistics under the null distribution tail probability also used for risk management in insurance, and portfolio investment For z much larger than mp, α is very small, and estimating this small probability accurately is not easy

Convergence property If X is square integrable, we have 1 n n h(x i ) N (E[h(X)], 1n ) Var[h(X)] i=1 Estimate the order of the rate of convergence by v/n, v = 1 n 1 n [h(x i ) 1 n i=1 disadvantage: slow convergence rate n 1/2 n h(x j )] 2 j=1 can we reduce v by generating samples smartly?

Important Sampling variance reduction method in Monte Carlo integration main idea: using samples generated from a different distribution rather than the distribution given suppose g(x) is another pdf h(x)f(x)dx = h(x) f(x) g(x) }{{} g(x)dx likelihood ratio change the function to be integrated from h(x) h(x) f(x) g(x) change the pdf from f(x) to g(x)

based on this, now we can sample Y 1, Y 2,..., Y n i.i.d. from g(x) estimate the integration as Î 2 = 1 n n f(y i ) h(y i ) g(y i ) }{{} weighted by likelihood ratio i=1 we can choose g(x) smartly so this estimator Î2 has smaller variance than the one based on direct sampling Î1 choose a distribution to encourage important values

How to choose a pdf to sample from we want to show that var{î2} var{î1}, since both methods are unbiased: E{Î1} = E{Î2}, or equivalently E{Î2 2 } E{Î2 1 }: [h(x) f(x) g(x) ]2 g(x)dx h(x) 2 f(x)dx, ( ) choose g(x) = h(x) f(x) h(v) f(v)dv, we can show (*) is true main idea: choose g(x) to resemble h(x) f(x), the standard error tends to be reduced difficulty: if h(x) f(x) is unknown, we have to approximate it

Proof. [ h(x)f(x) ] 2 g(x)dx g(x) = [h(x)f(x)] 2 /g(x)dx = [h(x)f(x)] 2 /( h(x) f(x))dx [ h(v) f(v)dv] = ( h(x) f(x)dx) 2 Cauchy Schwartz h(x) 2 f(x)dx f(x)dx } {{ } =1 = h 2 (x)f(x)dx