Approximate Bayesian Computation methods and their applications for hierarchical statistical models. University College London, 2015

Size: px
Start display at page:

Download "Approximate Bayesian Computation methods and their applications for hierarchical statistical models. University College London, 2015"

Transcription

1 Approximate Bayesian Computation methods and their applications for hierarchical statistical models University College London, 2015

2 Contents 1. Introduction 2. ABC methods 3. Hierarchical models 4. Application for ovarian cancer detection 5. Conclusion

3 Introduction The likelihood function plays an important role in statistical inference problems For complex models computational costs for evaluating the analytical formula are very high Methods which provide statistical inference bypassing evaluation of the likelihood function gained high popularity

4 ABC methods ABC methods provide ways of evaluating posterior distributions when the likelihood function is analytically or computationally intractable These methods are based on replacing the calculation of the likelihood with a comparison between the observed and simulated data θ Let be a parameter vector to be estimated. Given the prior distribution, the goal is to approximate the posterior distribution, where f(x θ) is the likelihood. π(θ) π(θ x) f(x θ)π(θ)

5 Generic form of ABC methods 1. Sample a candidate parameter vector from some proposal distribution. x* π(θ) θ* 2. Simulate a dataset from the model described by a conditional probability distribution. f(x θ*) 3. Compare the simulated dataset,, with the experimental data,, using a distance function, x 0 x* d ε d(x 0,x*) ε θ* The tolerance ε 0 is the desired level of agreement between x 0 and x*. and tolerance ;; if, accept.

6 Most popular ABC algorithms ABC rejection algorithm ABC MCMC algorithm (Markov Chain Monte Carlo) ABC SMC algorithm (Sequential Monte Carlo)

7 1. Sample from. ABC rejection method θ* π(θ) x* f(x θ*) d(x 0,x*) ε θ* 2. Simulate a dataset from. 3. If, accept, otherwise reject. 4. Return to step 1. Disadvantage: if prior distribution will be very different from the posterior, acceptance rate would be low.

8

9 Markov chain Monte Carlo (WIKIPEDIA) In mathematics, more specifically in statistics, Markov chain Monte Carlo (MCMC) methods are a class of algorithms for sampling from a probability distribution based on constructing a Markov chain that has the desired distribution as its equilibrium distribution. The state of the chain after a number of steps is then used as a sample of the desired distribution. The quality of the sample improves as a function of the number of steps.

10 ABC MCMC 1. Metropolis-Hastings Algorithm 2. Random-walk-Metropolis-Hastings Algorithm 3. Gibbs Sampling Algorithm 4. Metropolis within Gibbs Algorithm

11 q(y x) Metropolis-Hastings Algorithm Let be an arbitrary, friendly distribution (we know how to sample from) called proposal. Choose arbitrarily. Suppose we have generated X 0, X 1,..., X i. To generate X i+1 do the following: (1) Generate a proposal or candidate valuey ~q(y X i ) (2) Evaluate where (3) Set X 0 X i+1 = r r(x i Y) Y with probability r X i with probability 1 r " $ # $ % $! r(x,y)=min f(x)q(y x) f(y)q(x y) #,1 " # $ % # & # '

12 Remarks to Metropolis-Hastings Algorithm A simple way to execute step (3) is to generate U~(0,1) U<r X i+1 =Y X i+1 =X i. If set otherwise. q(y x) N(x,b 2 ) b>0 q q(y x)=q(x y), and! r(x,y)=min f(x) f(y),1 % # # " & # # A common choice for is for some. In this case, proposal density is symmetric, $ '

13 Metropolis-Hastings Algorithm. Example1 Let s simulate a Markov chain whose distribution is Let s take Then f(x)=1 π 1 1+x 2 N(x,b 2 ) r(x, y) = min! " # (The Cauchy distribution) as proposal distribution. f (y),1 f (x) Let s choose, length of chain. $ % & b=1 N=10,000! " # 1+x = min,1 1+y 2 2 $ % &

14 Example 1. Code in R N=10000 b=1 x_values=rep(0,n) x_cauchy=rep(0,n) x_axis=seq(-7,7,by=0.1) x_old=0 x_new=x_old for (i in 1:N) {y=rnorm(1,x_old,b) r=min((1+x_old^2)/(1+y^2),1) p=runif(1) if (p<r) (x_new=y) else (x_new=x_old);; x_values[i]=x_new x_old=x_new} x_cauchy=dcauchy(x_axis) plot(x_axis,x_cauchy,type="p",col="black") points(density(x_values),type="l",col="red",lwd=3)

15

16 Gibbs Sampling Gibbs Sampling is the easiest to use MCMC algorithm in case of dealing with high-dimensional problems as it helps to turn a high-dimensional problem into several one-dimensional problems. One of the examples of high-dimensional problems is hierarchical model

17 Hierarchical model. Example1 Posterior distribution on joint model specified. (θ,σ 2 ) X i ~(θ,σ 2 ), i=1,...,n, θ~n(θ,τ 2 0 ), σ 2 ~IG(a,b), θ 0,τ 2, a, b associated with the

18 Gibbs Sampling algorithm (X,Y) f X, Y (x, y) Suppose that has density. Suppose that it is possible to simulate from the conditional distributions and. Let be f X Y (x y) f Y X (y x) (X 0,Y 0 ) (X 0,Y 0 ),...,(X n,y n ) (X n+1,y n+1 ) starting values. Assume we have drawn Then the Gibbs sampling algorithm for getting : X n+1 ~ f X Y (x Y n ) Y n+1 ~ f Y X (y X n+1 ) repeat

19 Posteriors for the Example1 X i ~(θ,σ 2 ), i=1,...,n, θ~n(θ,τ 2 0 ), σ 2 ~IG(a,b), θ,τ 2 0, a, b f(θ x,σ 2! )~N σ 2 σ 2 +nτ 2θ 0+ nτ 2 σ 2 +nτ x, σ 2 τ 2 # # # # 2 σ 2 +nτ 2 # " " $ f(σ 2 $ " x,θ)~ig n+a, 1 x 2 2 θ $ % $ $ i ' $ '2 $ # & +b $ i $ # % ' ' ' ' ' ' ' ' & $ & & & & & %

20 x=rnorm(1000,10,2) n=length(x) a=3;; b=3 tau2=10 theta0=5 Nsim=5000 Example 1. Code in R xbar=mean(x) sh1=(n/2)+a sigma2=theta=rep(0,nsim) #init arrays sigma2[1]=1/rgamma(1,shape=a,rate=b) #init chains B=sigma2[1]/(sigma2[1]+n*tau2) theta[1]=rnorm(1,m=b*theta0+(1-b)*xbar,sd=sqrt(tau2*b)) for (i in 2:Nsim){ B=sigma2[i-1]/(sigma2[i-1]+n*tau2) theta[i]=rnorm(1,m=b*theta0+(1-b)*xbar,sd=sqrt(tau2*b)) ra1=(1/2)*(sum((x-theta[i])^2))+b sigma2[i]=1/rgamma(1,shape=sh1,rate=ra1) } mean(theta[3000:5000]) mean(sigma2[3000:5000])

21 Conjugate priors In Bayesian probability theory, if posterior distributions are in the same family as the prior distributions, then both prior and posterior are called conjugate distributions and the prior is called conjugate prior. P(θ D)= P(θ)P(D θ) P(θ)P(D θ)dθ

22 Conjugate priors. Example x~n(µ,σ (2) ) x σ (2) Let s consider normal distribution. For normally distributed with fixed variance, the conjugate prior is also normally distributed. For (2) prior µ~n(µ 0,σ 0 ) posterior will be in the form: µ x,σ (2) ~N( ˆµ 0, ˆ σ 0 (2) ), ˆµ 0 = σ (2) x+ σ (2) 0 σ (2) 0 +σ (2) σ (2) +σ µ (2) 0, 0 0 ˆ σ 0 (2) = σ (2) σ 0 (2) σ (2) +σ 0 (2)

23 Conjugate priors. Example Let s consider normal distribution. x~n(µ,σ (2) ) x µ For normally distributed with fixed mean, the conjugate prior is distributed according to inversegamma distribution. For prior σ (2) ~IG(α,β) P(x,µ σ (2) )= 1 σ 2π exp( (x µ)2 ) (σ 2 ) 1/2 exp( 1/2(x µ)2 ) 2σ 2 σ 2 P(σ (2) )=IG(α,β)= β α (σ (2) ) ( α 1) Γ(α) $ exp β & & & & & % σ (2) P(σ (2) x,µ) (σ 2 ) (α+1/2) 1 & & exp β 1/2(x µ)2 & ˆα=α+1/2, ˆβ=β+1(x µ) 2 $ & & % σ (2), ' ) ) ) ) ) ( ' ) ) ) ) ) (

24 ABC SMC A number of sampled parameter values (particles) {θ (1),...,θ (n) } π(θ), sampled from the prior distribution, are propagated through a sequence of intermediate distributions, π(θ d(x 0, x * ) ε i ), i=1,...,t 1, until it represents a sample from the target distribution π(θ d(x 0, x * ) ε T ). The tolerances what mean gradual evolving towards the target posterior. ε 1 >...>ε T 0 For sufficiently large numbers of particles, this approach avoid the problem of getting stuck in areas of low probability (as in ABC MCMC)

25 ABC SMC Algorithm S1. Initialize. Set the population indicator. S2.0 Set the particle indicator. S2.1 If, sample independently from. Else, sample with weights and perturb the particle to obtain, where is a perturbation kernel., return to S2.1. Simulate a candidate dataset. If ε 1,...,ε T t=0 i=1 t=0 θ ** π(θ) θ * (i) from the previous population {θ t 1 } θ ** ~K t (θ θ * ) If π(θ ** )=0 d(x *, x 0 ) ε t w t 1 K t, return to S2.1. x * ~ f(x θ ** )

26 θ t (i) =θ ** ABC SMC Algorithm S2.2 Set and calculate the weight for particle # % 1, if t=0, % (i) w t = π(θ (i) % t ) $, if t>0. % N w(j) t 1Kt (θ (j),θ % (i) t 1 t ) % &% j=1 If, set, go to S2.1. S3 Normalize the weights. t<t i<n i=i+1 t=t+1 If, set, go to S2.0. θ t (i)

27 Ovarian Cancer case study. CA125

28 Risk calculation

29 Change-point hierarchical model for CA125 Controls: Cases: Y ij t ij ~N(θ i,σ 2 ) Y ij t ij,{i i =0}~N(θ i,σ 2 ) Y ij t ij,{i i =1}~N(θ i +γ i (tij τ i) +,σ 2 )

30

31

32 Conditional distributions

33

34 Conclusion 1. ABC methods has great impact on parameters estimation. 1. A lot of applied problems can be reduced to hierarchical model 2. Gibbs Sampling Algorithm is most useful in dealing with hierarchical models

35 Literature 1. Steven J. Skates, Donna K. Pauler, Ian J. Jacobs. Screening Based on the Risk of Cancer Calculation from Bayesian Hierarchical Changepoint and Mixture Models of Longitudinal Markers. Journal of the American Statistical Society, vol. 96 (2001). 2. Wasserman L. All of Statistics. A concise course in Statistical Inference, Springer, Tina Toni, David Welch, Natalja Strelkowa, Andreas Ipsen, Michael P.H. Stumpf. Approximate Bayesian computation scheme for parameter inference and model selection in dynamical systems. Journal of the royal society, 6, (2009). 4. Robert P. Christian, Casella George. Introducing Monte Carlo Methods with R, Springer, 2009.

36 Data processing with caret package in R Data preprocessing Data splitting Data processing Model comparison

37 Data preprocessing preprocess Standardizing Transformation Imputing

38 Data preprocessing. Example data(bloodbrain) # contains array bbbdescr bbbdescr=bbbdescr[,-3] preproc <- preprocess(bbbdescr,method = c("center", "scale")) data <- predict(preproc, bbbdescr) mean(bbbdescr[,1]) mean(data[,1]) var(data[,1]) mean(bbbdescr[,2]) mean(data[,2]) var(data[,2])

39 Data splitting createdatapartition # training/test partition createresample # bootstrap samples createfolds # split the data into k groups createtimeslices # is used for time series data

40 Data splitting. Example data(bloodbrain) # contains array bbbdescr bbbdescr=bbbdescr[,-3] train_part <- createdatapartition(y=bbbdescr[,1], p=0.75, list=false) training <- bbbdescr[train_part,] testing <- bbbdescr[-train_part,] dim(bbbdescr) dim(training) dim(testing)

41 Data processing. Resampling train method= boot # bootstraping boot632 # bootstrapping with adjustment cv # cross validation repeatedcv # repeated cross validation LOOCV # leave one out cross validation

42 Data processing. Example library(mlbench) data(sonar) set.seed(107) intrain <- createdatapartition(y = Sonar$Class, p =.75, list = FALSE) training <- Sonar[ intrain,] testing <- Sonar[-inTrain,] plsfit <- train(class ~., data = training, method = "knn", preproc = c("center", "scale")) plsclasses <- predict(plsfit, newdata = testing) plsclasses

43 Model comparison. Metric options confusionmatrix Continuous outcomes: RMSE # root mean squared error RSquared # R^2 from regression models Categorical outcomes: Accuracy # fraction of correct classes Kappa # measure of concordance

44 Model comparison. Example names(getmodelinfo()) plsfit <- train(class ~., data = training, method = "knn", preproc = c("center", "scale")) plsclasses <- predict(plsfit, newdata = testing) confusionmatrix(data = plsclasses, testing$class) plsfit <- train(class ~., data = training, method = "pls", preproc = c("center", "scale")) plsclasses <- predict(plsfit, newdata = testing) confusionmatrix(data = plsclasses, testing$class) plsfit <- train(class ~., data = training, method = "cforest", preproc = c("center", "scale")) plsclasses <- predict(plsfit, newdata = testing) confusionmatrix(data = plsclasses, testing$class)

45 Literature 1. Max Kuhn. A Short Introduction to the caret Package (2014). 2. Model training and tuning:

46 Questions

A Short Introduction to the caret Package

A Short Introduction to the caret Package A Short Introduction to the caret Package Max Kuhn max.kuhn@pfizer.com October 28, 2016 The caret package (short for classification and regression training) contains functions to streamline the model training

More information

A Short Introduction to the caret Package

A Short Introduction to the caret Package A Short Introduction to the caret Package Max Kuhn max.kuhn@pfizer.com February 12, 2013 The caret package (short for classification and regression training) contains functions to streamline the model

More information

Approximate Bayesian Computation. Alireza Shafaei - April 2016

Approximate Bayesian Computation. Alireza Shafaei - April 2016 Approximate Bayesian Computation Alireza Shafaei - April 2016 The Problem Given a dataset, we are interested in. The Problem Given a dataset, we are interested in. The Problem Given a dataset, we are interested

More information

Optimization Methods III. The MCMC. Exercises.

Optimization Methods III. The MCMC. Exercises. Aula 8. Optimization Methods III. Exercises. 0 Optimization Methods III. The MCMC. Exercises. Anatoli Iambartsev IME-USP Aula 8. Optimization Methods III. Exercises. 1 [RC] A generic Markov chain Monte

More information

Rolling Markov Chain Monte Carlo

Rolling Markov Chain Monte Carlo Rolling Markov Chain Monte Carlo Din-Houn Lau Imperial College London Joint work with Axel Gandy 4 th September 2013 RSS Conference 2013: Newcastle Output predicted final ranks of the each team. Updates

More information

Rolling Markov Chain Monte Carlo

Rolling Markov Chain Monte Carlo Rolling Markov Chain Monte Carlo Din-Houn Lau Imperial College London Joint work with Axel Gandy 4 th July 2013 Predict final ranks of the each team. Updates quick update of predictions. Accuracy control

More information

MCMC Methods for data modeling

MCMC Methods for data modeling MCMC Methods for data modeling Kenneth Scerri Department of Automatic Control and Systems Engineering Introduction 1. Symposium on Data Modelling 2. Outline: a. Definition and uses of MCMC b. MCMC algorithms

More information

Bayesian Statistics Group 8th March Slice samplers. (A very brief introduction) The basic idea

Bayesian Statistics Group 8th March Slice samplers. (A very brief introduction) The basic idea Bayesian Statistics Group 8th March 2000 Slice samplers (A very brief introduction) The basic idea lacements To sample from a distribution, simply sample uniformly from the region under the density function

More information

Markov Chain Monte Carlo (part 1)

Markov Chain Monte Carlo (part 1) Markov Chain Monte Carlo (part 1) Edps 590BAY Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Spring 2018 Depending on the book that you select for

More information

Markov chain Monte Carlo methods

Markov chain Monte Carlo methods Markov chain Monte Carlo methods (supplementary material) see also the applet http://www.lbreyer.com/classic.html February 9 6 Independent Hastings Metropolis Sampler Outline Independent Hastings Metropolis

More information

Statistical Matching using Fractional Imputation

Statistical Matching using Fractional Imputation Statistical Matching using Fractional Imputation Jae-Kwang Kim 1 Iowa State University 1 Joint work with Emily Berg and Taesung Park 1 Introduction 2 Classical Approaches 3 Proposed method 4 Application:

More information

A GENERAL GIBBS SAMPLING ALGORITHM FOR ANALYZING LINEAR MODELS USING THE SAS SYSTEM

A GENERAL GIBBS SAMPLING ALGORITHM FOR ANALYZING LINEAR MODELS USING THE SAS SYSTEM A GENERAL GIBBS SAMPLING ALGORITHM FOR ANALYZING LINEAR MODELS USING THE SAS SYSTEM Jayawant Mandrekar, Daniel J. Sargent, Paul J. Novotny, Jeff A. Sloan Mayo Clinic, Rochester, MN 55905 ABSTRACT A general

More information

BART STAT8810, Fall 2017

BART STAT8810, Fall 2017 BART STAT8810, Fall 2017 M.T. Pratola November 1, 2017 Today BART: Bayesian Additive Regression Trees BART: Bayesian Additive Regression Trees Additive model generalizes the single-tree regression model:

More information

Monte Carlo for Spatial Models

Monte Carlo for Spatial Models Monte Carlo for Spatial Models Murali Haran Department of Statistics Penn State University Penn State Computational Science Lectures April 2007 Spatial Models Lots of scientific questions involve analyzing

More information

Missing Data and Imputation

Missing Data and Imputation Missing Data and Imputation NINA ORWITZ OCTOBER 30 TH, 2017 Outline Types of missing data Simple methods for dealing with missing data Single and multiple imputation R example Missing data is a complex

More information

Approximate Bayesian Computation using Auxiliary Models

Approximate Bayesian Computation using Auxiliary Models Approximate Bayesian Computation using Auxiliary Models Tony Pettitt Co-authors Chris Drovandi, Malcolm Faddy Queensland University of Technology Brisbane MCQMC February 2012 Tony Pettitt () ABC using

More information

1 Methods for Posterior Simulation

1 Methods for Posterior Simulation 1 Methods for Posterior Simulation Let p(θ y) be the posterior. simulation. Koop presents four methods for (posterior) 1. Monte Carlo integration: draw from p(θ y). 2. Gibbs sampler: sequentially drawing

More information

Bayesian Statistics with a Smile: a Resampling-Sampling Perspective

Bayesian Statistics with a Smile: a Resampling-Sampling Perspective Bayesian Statistics with a Smile: a Resampling-Sampling Perspective Hedibert F. Lopes The University of Chicago Booth School of Business 5807 S. Woodlawn Ave., Chicago, IL, 60637. hlopes@chicagobooth.edu

More information

Convexization in Markov Chain Monte Carlo

Convexization in Markov Chain Monte Carlo in Markov Chain Monte Carlo 1 IBM T. J. Watson Yorktown Heights, NY 2 Department of Aerospace Engineering Technion, Israel August 23, 2011 Problem Statement MCMC processes in general are governed by non

More information

Monte Carlo Techniques for Bayesian Statistical Inference A comparative review

Monte Carlo Techniques for Bayesian Statistical Inference A comparative review 1 Monte Carlo Techniques for Bayesian Statistical Inference A comparative review BY Y. FAN, D. WANG School of Mathematics and Statistics, University of New South Wales, Sydney 2052, AUSTRALIA 18th January,

More information

Missing Data Analysis for the Employee Dataset

Missing Data Analysis for the Employee Dataset Missing Data Analysis for the Employee Dataset 67% of the observations have missing values! Modeling Setup Random Variables: Y i =(Y i1,...,y ip ) 0 =(Y i,obs, Y i,miss ) 0 R i =(R i1,...,r ip ) 0 ( 1

More information

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXX 23 An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework Ji Won Yoon arxiv:37.99v [cs.lg] 3 Jul 23 Abstract In order to cluster

More information

Sampling informative/complex a priori probability distributions using Gibbs sampling assisted by sequential simulation

Sampling informative/complex a priori probability distributions using Gibbs sampling assisted by sequential simulation Sampling informative/complex a priori probability distributions using Gibbs sampling assisted by sequential simulation Thomas Mejer Hansen, Klaus Mosegaard, and Knud Skou Cordua 1 1 Center for Energy Resources

More information

Chapter 1. Introduction

Chapter 1. Introduction Chapter 1 Introduction A Monte Carlo method is a compuational method that uses random numbers to compute (estimate) some quantity of interest. Very often the quantity we want to compute is the mean of

More information

Image analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis

Image analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis 7 Computer Vision and Classification 413 / 458 Computer Vision and Classification The k-nearest-neighbor method The k-nearest-neighbor (knn) procedure has been used in data analysis and machine learning

More information

Bayesian Modelling with JAGS and R

Bayesian Modelling with JAGS and R Bayesian Modelling with JAGS and R Martyn Plummer International Agency for Research on Cancer Rencontres R, 3 July 2012 CRAN Task View Bayesian Inference The CRAN Task View Bayesian Inference is maintained

More information

MSA101/MVE Lecture 5

MSA101/MVE Lecture 5 MSA101/MVE187 2017 Lecture 5 Petter Mostad Chalmers University September 12, 2017 1 / 15 Importance sampling MC integration computes h(x)f (x) dx where f (x) is a probability density function, by simulating

More information

STAT 725 Notes Monte Carlo Integration

STAT 725 Notes Monte Carlo Integration STAT 725 Notes Monte Carlo Integration Two major classes of numerical problems arise in statistical inference: optimization and integration. We have already spent some time discussing different optimization

More information

Nested Sampling: Introduction and Implementation

Nested Sampling: Introduction and Implementation UNIVERSITY OF TEXAS AT SAN ANTONIO Nested Sampling: Introduction and Implementation Liang Jing May 2009 1 1 ABSTRACT Nested Sampling is a new technique to calculate the evidence, Z = P(D M) = p(d θ, M)p(θ

More information

Quantitative Biology II!

Quantitative Biology II! Quantitative Biology II! Lecture 3: Markov Chain Monte Carlo! March 9, 2015! 2! Plan for Today!! Introduction to Sampling!! Introduction to MCMC!! Metropolis Algorithm!! Metropolis-Hastings Algorithm!!

More information

INLA: Integrated Nested Laplace Approximations

INLA: Integrated Nested Laplace Approximations INLA: Integrated Nested Laplace Approximations John Paige Statistics Department University of Washington October 10, 2017 1 The problem Markov Chain Monte Carlo (MCMC) takes too long in many settings.

More information

The Bootstrap and Jackknife

The Bootstrap and Jackknife The Bootstrap and Jackknife Summer 2017 Summer Institutes 249 Bootstrap & Jackknife Motivation In scientific research Interest often focuses upon the estimation of some unknown parameter, θ. The parameter

More information

A Short History of Markov Chain Monte Carlo

A Short History of Markov Chain Monte Carlo A Short History of Markov Chain Monte Carlo Christian Robert and George Casella 2010 Introduction Lack of computing machinery, or background on Markov chains, or hesitation to trust in the practicality

More information

Non-parametric Approximate Bayesian Computation for Expensive Simulators

Non-parametric Approximate Bayesian Computation for Expensive Simulators Non-parametric Approximate Bayesian Computation for Expensive Simulators Steven Laan 6036031 Master s thesis 42 EC Master s programme Artificial Intelligence University of Amsterdam Supervisor Ted Meeds

More information

GiRaF: a toolbox for Gibbs Random Fields analysis

GiRaF: a toolbox for Gibbs Random Fields analysis GiRaF: a toolbox for Gibbs Random Fields analysis Julien Stoehr *1, Pierre Pudlo 2, and Nial Friel 1 1 University College Dublin 2 Aix-Marseille Université February 24, 2016 Abstract GiRaF package offers

More information

10-701/15-781, Fall 2006, Final

10-701/15-781, Fall 2006, Final -7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly

More information

Modeling Criminal Careers as Departures From a Unimodal Population Age-Crime Curve: The Case of Marijuana Use

Modeling Criminal Careers as Departures From a Unimodal Population Age-Crime Curve: The Case of Marijuana Use Modeling Criminal Careers as Departures From a Unimodal Population Curve: The Case of Marijuana Use Donatello Telesca, Elena A. Erosheva, Derek A. Kreader, & Ross Matsueda April 15, 2014 extends Telesca

More information

Bayesian Spatiotemporal Modeling with Hierarchical Spatial Priors for fmri

Bayesian Spatiotemporal Modeling with Hierarchical Spatial Priors for fmri Bayesian Spatiotemporal Modeling with Hierarchical Spatial Priors for fmri Galin L. Jones 1 School of Statistics University of Minnesota March 2015 1 Joint with Martin Bezener and John Hughes Experiment

More information

10.4 Linear interpolation method Newton s method

10.4 Linear interpolation method Newton s method 10.4 Linear interpolation method The next best thing one can do is the linear interpolation method, also known as the double false position method. This method works similarly to the bisection method by

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 17 EM CS/CNS/EE 155 Andreas Krause Announcements Project poster session on Thursday Dec 3, 4-6pm in Annenberg 2 nd floor atrium! Easels, poster boards and cookies

More information

CSCI 599 Class Presenta/on. Zach Levine. Markov Chain Monte Carlo (MCMC) HMM Parameter Es/mates

CSCI 599 Class Presenta/on. Zach Levine. Markov Chain Monte Carlo (MCMC) HMM Parameter Es/mates CSCI 599 Class Presenta/on Zach Levine Markov Chain Monte Carlo (MCMC) HMM Parameter Es/mates April 26 th, 2012 Topics Covered in this Presenta2on A (Brief) Review of HMMs HMM Parameter Learning Expecta2on-

More information

Performance of Sequential Imputation Method in Multilevel Applications

Performance of Sequential Imputation Method in Multilevel Applications Section on Survey Research Methods JSM 9 Performance of Sequential Imputation Method in Multilevel Applications Enxu Zhao, Recai M. Yucel New York State Department of Health, 8 N. Pearl St., Albany, NY

More information

Linear Modeling with Bayesian Statistics

Linear Modeling with Bayesian Statistics Linear Modeling with Bayesian Statistics Bayesian Approach I I I I I Estimate probability of a parameter State degree of believe in specific parameter values Evaluate probability of hypothesis given the

More information

10703 Deep Reinforcement Learning and Control

10703 Deep Reinforcement Learning and Control 10703 Deep Reinforcement Learning and Control Russ Salakhutdinov Machine Learning Department rsalakhu@cs.cmu.edu Policy Gradient I Used Materials Disclaimer: Much of the material and slides for this lecture

More information

Clustering Relational Data using the Infinite Relational Model

Clustering Relational Data using the Infinite Relational Model Clustering Relational Data using the Infinite Relational Model Ana Daglis Supervised by: Matthew Ludkin September 4, 2015 Ana Daglis Clustering Data using the Infinite Relational Model September 4, 2015

More information

Stochastic Simulation: Algorithms and Analysis

Stochastic Simulation: Algorithms and Analysis Soren Asmussen Peter W. Glynn Stochastic Simulation: Algorithms and Analysis et Springer Contents Preface Notation v xii I What This Book Is About 1 1 An Illustrative Example: The Single-Server Queue 1

More information

Bayesian Estimation for Skew Normal Distributions Using Data Augmentation

Bayesian Estimation for Skew Normal Distributions Using Data Augmentation The Korean Communications in Statistics Vol. 12 No. 2, 2005 pp. 323-333 Bayesian Estimation for Skew Normal Distributions Using Data Augmentation Hea-Jung Kim 1) Abstract In this paper, we develop a MCMC

More information

Modified Metropolis-Hastings algorithm with delayed rejection

Modified Metropolis-Hastings algorithm with delayed rejection Modified Metropolis-Hastings algorithm with delayed reection K.M. Zuev & L.S. Katafygiotis Department of Civil Engineering, Hong Kong University of Science and Technology, Hong Kong, China ABSTRACT: The

More information

Machine Learning (BSMC-GA 4439) Wenke Liu

Machine Learning (BSMC-GA 4439) Wenke Liu Machine Learning (BSMC-GA 4439) Wenke Liu 01-31-017 Outline Background Defining proximity Clustering methods Determining number of clusters Comparing two solutions Cluster analysis as unsupervised Learning

More information

A noninformative Bayesian approach to small area estimation

A noninformative Bayesian approach to small area estimation A noninformative Bayesian approach to small area estimation Glen Meeden School of Statistics University of Minnesota Minneapolis, MN 55455 glen@stat.umn.edu September 2001 Revised May 2002 Research supported

More information

Overview. Monte Carlo Methods. Statistics & Bayesian Inference Lecture 3. Situation At End Of Last Week

Overview. Monte Carlo Methods. Statistics & Bayesian Inference Lecture 3. Situation At End Of Last Week Statistics & Bayesian Inference Lecture 3 Joe Zuntz Overview Overview & Motivation Metropolis Hastings Monte Carlo Methods Importance sampling Direct sampling Gibbs sampling Monte-Carlo Markov Chains Emcee

More information

Real Coded Genetic Algorithm Particle Filter for Improved Performance

Real Coded Genetic Algorithm Particle Filter for Improved Performance Proceedings of 2012 4th International Conference on Machine Learning and Computing IPCSIT vol. 25 (2012) (2012) IACSIT Press, Singapore Real Coded Genetic Algorithm Particle Filter for Improved Performance

More information

MCMC Diagnostics. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) MCMC Diagnostics MATH / 24

MCMC Diagnostics. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) MCMC Diagnostics MATH / 24 MCMC Diagnostics Yingbo Li Clemson University MATH 9810 Yingbo Li (Clemson) MCMC Diagnostics MATH 9810 1 / 24 Convergence to Posterior Distribution Theory proves that if a Gibbs sampler iterates enough,

More information

Monte Carlo Methods and Statistical Computing: My Personal E

Monte Carlo Methods and Statistical Computing: My Personal E Monte Carlo Methods and Statistical Computing: My Personal Experience Department of Mathematics & Statistics Indian Institute of Technology Kanpur November 29, 2014 Outline Preface 1 Preface 2 3 4 5 6

More information

Journal of Mathematical Psychology

Journal of Mathematical Psychology Journal of Mathematical Psychology 56 (2012) 375 385 Contents lists available at SciVerse ScienceDirect Journal of Mathematical Psychology journal homepage: www.elsevier.com/locate/jmp Approximate Bayesian

More information

Particle Filters for Visual Tracking

Particle Filters for Visual Tracking Particle Filters for Visual Tracking T. Chateau, Pascal Institute, Clermont-Ferrand 1 Content Particle filtering: a probabilistic framework SIR particle filter MCMC particle filter RJMCMC particle filter

More information

Image Segmentation using Gaussian Mixture Models

Image Segmentation using Gaussian Mixture Models Image Segmentation using Gaussian Mixture Models Rahman Farnoosh, Gholamhossein Yari and Behnam Zarpak Department of Applied Mathematics, University of Science and Technology, 16844, Narmak,Tehran, Iran

More information

Hierarchical Bayesian Modeling with Ensemble MCMC. Eric B. Ford (Penn State) Bayesian Computing for Astronomical Data Analysis June 12, 2014

Hierarchical Bayesian Modeling with Ensemble MCMC. Eric B. Ford (Penn State) Bayesian Computing for Astronomical Data Analysis June 12, 2014 Hierarchical Bayesian Modeling with Ensemble MCMC Eric B. Ford (Penn State) Bayesian Computing for Astronomical Data Analysis June 12, 2014 Simple Markov Chain Monte Carlo Initialise chain with θ 0 (initial

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Overview of Part Two Probabilistic Graphical Models Part Two: Inference and Learning Christopher M. Bishop Exact inference and the junction tree MCMC Variational methods and EM Example General variational

More information

CS281 Section 9: Graph Models and Practical MCMC

CS281 Section 9: Graph Models and Practical MCMC CS281 Section 9: Graph Models and Practical MCMC Scott Linderman November 11, 213 Now that we have a few MCMC inference algorithms in our toolbox, let s try them out on some random graph models. Graphs

More information

Exam Issued: May 29, 2017, 13:00 Hand in: May 29, 2017, 16:00

Exam Issued: May 29, 2017, 13:00 Hand in: May 29, 2017, 16:00 P. Hadjidoukas, C. Papadimitriou ETH Zentrum, CTL E 13 CH-8092 Zürich High Performance Computing for Science and Engineering II Exam Issued: May 29, 2017, 13:00 Hand in: May 29, 2017, 16:00 Spring semester

More information

Monte Carlo sampling

Monte Carlo sampling 1 y u theta 0 x 1 Monte Carlo sampling Problem 1 Suppose we want to sample uniformly at random from the triangle defined by the points (0,0), (0,1), (1,0). First Sampling Algorithm: We decide to do this

More information

Package HKprocess. R topics documented: September 6, Type Package. Title Hurst-Kolmogorov process. Version

Package HKprocess. R topics documented: September 6, Type Package. Title Hurst-Kolmogorov process. Version Package HKprocess September 6, 2014 Type Package Title Hurst-Kolmogorov process Version 0.0-1 Date 2014-09-06 Author Maintainer Imports MCMCpack (>= 1.3-3), gtools(>= 3.4.1) Depends

More information

Package mederrrank. R topics documented: July 8, 2015

Package mederrrank. R topics documented: July 8, 2015 Package mederrrank July 8, 2015 Title Bayesian Methods for Identifying the Most Harmful Medication Errors Version 0.0.8 Date 2015-07-06 Two distinct but related statistical approaches to the problem of

More information

arxiv: v3 [stat.co] 27 Apr 2012

arxiv: v3 [stat.co] 27 Apr 2012 A multi-point Metropolis scheme with generic weight functions arxiv:1112.4048v3 stat.co 27 Apr 2012 Abstract Luca Martino, Victor Pascual Del Olmo, Jesse Read Department of Signal Theory and Communications,

More information

A Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data

A Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data A Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data Seyoung Kim, Padhraic Smyth, and Hal Stern Bren School of Information and Computer Sciences University of California,

More information

Topics in Machine Learning-EE 5359 Model Assessment and Selection

Topics in Machine Learning-EE 5359 Model Assessment and Selection Topics in Machine Learning-EE 5359 Model Assessment and Selection Ioannis D. Schizas Electrical Engineering Department University of Texas at Arlington 1 Training and Generalization Training stage: Utilizing

More information

GT "Calcul Ensembliste"

GT Calcul Ensembliste GT "Calcul Ensembliste" Beyond the bounded error framework for non linear state estimation Fahed Abdallah Université de Technologie de Compiègne 9 Décembre 2010 Fahed Abdallah GT "Calcul Ensembliste" 9

More information

ADAPTIVE METROPOLIS-HASTINGS SAMPLING, OR MONTE CARLO KERNEL ESTIMATION

ADAPTIVE METROPOLIS-HASTINGS SAMPLING, OR MONTE CARLO KERNEL ESTIMATION ADAPTIVE METROPOLIS-HASTINGS SAMPLING, OR MONTE CARLO KERNEL ESTIMATION CHRISTOPHER A. SIMS Abstract. A new algorithm for sampling from an arbitrary pdf. 1. Introduction Consider the standard problem of

More information

Accelerated Markov Chain Monte Carlo Sampling in Electrical Capacitance Tomography

Accelerated Markov Chain Monte Carlo Sampling in Electrical Capacitance Tomography Accelerated Markov Chain Monte Carlo Sampling in Electrical Capacitance Tomography Daniel Watzenig, Markus Neumayer and Colin Fox Institute of Electrical Measurement and Measurement Signal Processing,

More information

One-Shot Learning with a Hierarchical Nonparametric Bayesian Model

One-Shot Learning with a Hierarchical Nonparametric Bayesian Model One-Shot Learning with a Hierarchical Nonparametric Bayesian Model R. Salakhutdinov, J. Tenenbaum and A. Torralba MIT Technical Report, 2010 Presented by Esther Salazar Duke University June 10, 2011 E.

More information

Machine Learning (BSMC-GA 4439) Wenke Liu

Machine Learning (BSMC-GA 4439) Wenke Liu Machine Learning (BSMC-GA 4439) Wenke Liu 01-25-2018 Outline Background Defining proximity Clustering methods Determining number of clusters Other approaches Cluster analysis as unsupervised Learning Unsupervised

More information

Markov Random Fields and Gibbs Sampling for Image Denoising

Markov Random Fields and Gibbs Sampling for Image Denoising Markov Random Fields and Gibbs Sampling for Image Denoising Chang Yue Electrical Engineering Stanford University changyue@stanfoed.edu Abstract This project applies Gibbs Sampling based on different Markov

More information

Graphical Models, Bayesian Method, Sampling, and Variational Inference

Graphical Models, Bayesian Method, Sampling, and Variational Inference Graphical Models, Bayesian Method, Sampling, and Variational Inference With Application in Function MRI Analysis and Other Imaging Problems Wei Liu Scientific Computing and Imaging Institute University

More information

BeviMed Guide. Daniel Greene

BeviMed Guide. Daniel Greene BeviMed Guide Daniel Greene 1 Introduction BeviMed [1] is a procedure for evaluating the evidence of association between allele configurations across rare variants, typically within a genomic locus, and

More information

Analysis of Incomplete Multivariate Data

Analysis of Incomplete Multivariate Data Analysis of Incomplete Multivariate Data J. L. Schafer Department of Statistics The Pennsylvania State University USA CHAPMAN & HALL/CRC A CR.C Press Company Boca Raton London New York Washington, D.C.

More information

MULTI-DIMENSIONAL MONTE CARLO INTEGRATION

MULTI-DIMENSIONAL MONTE CARLO INTEGRATION CS580: Computer Graphics KAIST School of Computing Chapter 3 MULTI-DIMENSIONAL MONTE CARLO INTEGRATION 2 1 Monte Carlo Integration This describes a simple technique for the numerical evaluation of integrals

More information

Resampling Methods. Levi Waldron, CUNY School of Public Health. July 13, 2016

Resampling Methods. Levi Waldron, CUNY School of Public Health. July 13, 2016 Resampling Methods Levi Waldron, CUNY School of Public Health July 13, 2016 Outline and introduction Objectives: prediction or inference? Cross-validation Bootstrap Permutation Test Monte Carlo Simulation

More information

Estimating Labels from Label Proportions

Estimating Labels from Label Proportions Estimating Labels from Label Proportions Novi Quadrianto Novi.Quad@gmail.com The Australian National University, Australia NICTA, Statistical Machine Learning Program, Australia Joint work with Alex Smola,

More information

Semiparametric Mixed Effecs with Hierarchical DP Mixture

Semiparametric Mixed Effecs with Hierarchical DP Mixture Semiparametric Mixed Effecs with Hierarchical DP Mixture R topics documented: April 21, 2007 hdpm-package........................................ 1 hdpm............................................ 2 hdpmfitsetup........................................

More information

Bayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19

Bayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19 Bayesian Methods David Rosenberg New York University April 11, 2017 David Rosenberg (New York University) DS-GA 1003 April 11, 2017 1 / 19 Classical Statistics Classical Statistics David Rosenberg (New

More information

Hybrid Quasi-Monte Carlo Method for the Simulation of State Space Models

Hybrid Quasi-Monte Carlo Method for the Simulation of State Space Models The Tenth International Symposium on Operations Research and Its Applications (ISORA 211) Dunhuang, China, August 28 31, 211 Copyright 211 ORSC & APORC, pp. 83 88 Hybrid Quasi-Monte Carlo Method for the

More information

A Deterministic Global Optimization Method for Variational Inference

A Deterministic Global Optimization Method for Variational Inference A Deterministic Global Optimization Method for Variational Inference Hachem Saddiki Mathematics and Statistics University of Massachusetts, Amherst saddiki@math.umass.edu Andrew C. Trapp Operations and

More information

Curve Sampling and Geometric Conditional Simulation

Curve Sampling and Geometric Conditional Simulation Curve Sampling and Geometric Conditional Simulation Ayres Fan Joint work with John Fisher, William Wells, Jonathan Kane, and Alan Willsky S S G Massachusetts Institute of Technology September 19, 2007

More information

Model validation through "Posterior predictive checking" and "Leave-one-out"

Model validation through Posterior predictive checking and Leave-one-out Workshop of the Bayes WG / IBS-DR Mainz, 2006-12-01 G. Nehmiz M. Könen-Bergmann Model validation through "Posterior predictive checking" and "Leave-one-out" Overview The posterior predictive distribution

More information

Project Report: "Bayesian Spam Filter"

Project Report: Bayesian  Spam Filter Humboldt-Universität zu Berlin Lehrstuhl für Maschinelles Lernen Sommersemester 2016 Maschinelles Lernen 1 Project Report: "Bayesian E-Mail Spam Filter" The Bayesians Sabine Bertram, Carolina Gumuljo,

More information

Instability, Sensitivity, and Degeneracy of Discrete Exponential Families

Instability, Sensitivity, and Degeneracy of Discrete Exponential Families Instability, Sensitivity, and Degeneracy of Discrete Exponential Families Michael Schweinberger Pennsylvania State University ONR grant N00014-08-1-1015 Scalable Methods for the Analysis of Network-Based

More information

Bayesian data analysis using R

Bayesian data analysis using R Bayesian data analysis using R BAYESIAN DATA ANALYSIS USING R Jouni Kerman, Samantha Cook, and Andrew Gelman Introduction Bayesian data analysis includes but is not limited to Bayesian inference (Gelman

More information

A Bayesian approach to artificial neural network model selection

A Bayesian approach to artificial neural network model selection A Bayesian approach to artificial neural network model selection Kingston, G. B., H. R. Maier and M. F. Lambert Centre for Applied Modelling in Water Engineering, School of Civil and Environmental Engineering,

More information

Level-set MCMC Curve Sampling and Geometric Conditional Simulation

Level-set MCMC Curve Sampling and Geometric Conditional Simulation Level-set MCMC Curve Sampling and Geometric Conditional Simulation Ayres Fan John W. Fisher III Alan S. Willsky February 16, 2007 Outline 1. Overview 2. Curve evolution 3. Markov chain Monte Carlo 4. Curve

More information

Recap: The E-M algorithm. Biostatistics 615/815 Lecture 22: Gibbs Sampling. Recap - Local minimization methods

Recap: The E-M algorithm. Biostatistics 615/815 Lecture 22: Gibbs Sampling. Recap - Local minimization methods Recap: The E-M algorithm Biostatistics 615/815 Lecture 22: Gibbs Sampling Expectation step (E-step) Given the current estimates of parameters λ (t), calculate the conditional distribution of latent variable

More information

Bayesian Computation: Overview and Methods for Low-Dimensional Models

Bayesian Computation: Overview and Methods for Low-Dimensional Models Bayesian Computation: Overview and Methods for Low-Dimensional Models Tom Loredo Dept. of Astronomy, Cornell University http://www.astro.cornell.edu/staff/loredo/bayes/ IAC Winter School, 3 4 Nov 2014

More information

Statistical Analysis Using Combined Data Sources: Discussion JPSM Distinguished Lecture University of Maryland

Statistical Analysis Using Combined Data Sources: Discussion JPSM Distinguished Lecture University of Maryland Statistical Analysis Using Combined Data Sources: Discussion 2011 JPSM Distinguished Lecture University of Maryland 1 1 University of Michigan School of Public Health April 2011 Complete (Ideal) vs. Observed

More information

Monte Carlo Optimization Introduction

Monte Carlo Optimization Introduction Monte Carlo Methods with R: Monte Carlo Optimization [80] Monte Carlo Optimization Introduction Optimization problems can mostly be seen as one of two kinds: Find the extrema of a function h(θ) over a

More information

Introduction to Optimization

Introduction to Optimization Introduction to Optimization Second Order Optimization Methods Marc Toussaint U Stuttgart Planned Outline Gradient-based optimization (1st order methods) plain grad., steepest descent, conjugate grad.,

More information

A study of classification algorithms using Rapidminer

A study of classification algorithms using Rapidminer Volume 119 No. 12 2018, 15977-15988 ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu ijpam.eu A study of classification algorithms using Rapidminer Dr.J.Arunadevi 1, S.Ramya 2, M.Ramesh Raja

More information

Transitivity and Triads

Transitivity and Triads 1 / 32 Tom A.B. Snijders University of Oxford May 14, 2012 2 / 32 Outline 1 Local Structure Transitivity 2 3 / 32 Local Structure in Social Networks From the standpoint of structural individualism, one

More information

Outline. Bayesian Data Analysis Hierarchical models. Rat tumor data. Errandum: exercise GCSR 3.11

Outline. Bayesian Data Analysis Hierarchical models. Rat tumor data. Errandum: exercise GCSR 3.11 Outline Bayesian Data Analysis Hierarchical models Helle Sørensen May 15, 2009 Today: More about the rat tumor data: model, derivation of posteriors, the actual computations in R. : a hierarchical normal

More information

An Introduction to Markov Chain Monte Carlo

An Introduction to Markov Chain Monte Carlo An Introduction to Markov Chain Monte Carlo Markov Chain Monte Carlo (MCMC) refers to a suite of processes for simulating a posterior distribution based on a random (ie. monte carlo) process. In other

More information

Chapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea

Chapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea Chapter 3 Bootstrap 3.1 Introduction The estimation of parameters in probability distributions is a basic problem in statistics that one tends to encounter already during the very first course on the subject.

More information