Approximate Bayesian Computation methods and their applications for hierarchical statistical models. University College London, 2015
|
|
- Nicholas Ray
- 5 years ago
- Views:
Transcription
1 Approximate Bayesian Computation methods and their applications for hierarchical statistical models University College London, 2015
2 Contents 1. Introduction 2. ABC methods 3. Hierarchical models 4. Application for ovarian cancer detection 5. Conclusion
3 Introduction The likelihood function plays an important role in statistical inference problems For complex models computational costs for evaluating the analytical formula are very high Methods which provide statistical inference bypassing evaluation of the likelihood function gained high popularity
4 ABC methods ABC methods provide ways of evaluating posterior distributions when the likelihood function is analytically or computationally intractable These methods are based on replacing the calculation of the likelihood with a comparison between the observed and simulated data θ Let be a parameter vector to be estimated. Given the prior distribution, the goal is to approximate the posterior distribution, where f(x θ) is the likelihood. π(θ) π(θ x) f(x θ)π(θ)
5 Generic form of ABC methods 1. Sample a candidate parameter vector from some proposal distribution. x* π(θ) θ* 2. Simulate a dataset from the model described by a conditional probability distribution. f(x θ*) 3. Compare the simulated dataset,, with the experimental data,, using a distance function, x 0 x* d ε d(x 0,x*) ε θ* The tolerance ε 0 is the desired level of agreement between x 0 and x*. and tolerance ;; if, accept.
6 Most popular ABC algorithms ABC rejection algorithm ABC MCMC algorithm (Markov Chain Monte Carlo) ABC SMC algorithm (Sequential Monte Carlo)
7 1. Sample from. ABC rejection method θ* π(θ) x* f(x θ*) d(x 0,x*) ε θ* 2. Simulate a dataset from. 3. If, accept, otherwise reject. 4. Return to step 1. Disadvantage: if prior distribution will be very different from the posterior, acceptance rate would be low.
8
9 Markov chain Monte Carlo (WIKIPEDIA) In mathematics, more specifically in statistics, Markov chain Monte Carlo (MCMC) methods are a class of algorithms for sampling from a probability distribution based on constructing a Markov chain that has the desired distribution as its equilibrium distribution. The state of the chain after a number of steps is then used as a sample of the desired distribution. The quality of the sample improves as a function of the number of steps.
10 ABC MCMC 1. Metropolis-Hastings Algorithm 2. Random-walk-Metropolis-Hastings Algorithm 3. Gibbs Sampling Algorithm 4. Metropolis within Gibbs Algorithm
11 q(y x) Metropolis-Hastings Algorithm Let be an arbitrary, friendly distribution (we know how to sample from) called proposal. Choose arbitrarily. Suppose we have generated X 0, X 1,..., X i. To generate X i+1 do the following: (1) Generate a proposal or candidate valuey ~q(y X i ) (2) Evaluate where (3) Set X 0 X i+1 = r r(x i Y) Y with probability r X i with probability 1 r " $ # $ % $! r(x,y)=min f(x)q(y x) f(y)q(x y) #,1 " # $ % # & # '
12 Remarks to Metropolis-Hastings Algorithm A simple way to execute step (3) is to generate U~(0,1) U<r X i+1 =Y X i+1 =X i. If set otherwise. q(y x) N(x,b 2 ) b>0 q q(y x)=q(x y), and! r(x,y)=min f(x) f(y),1 % # # " & # # A common choice for is for some. In this case, proposal density is symmetric, $ '
13 Metropolis-Hastings Algorithm. Example1 Let s simulate a Markov chain whose distribution is Let s take Then f(x)=1 π 1 1+x 2 N(x,b 2 ) r(x, y) = min! " # (The Cauchy distribution) as proposal distribution. f (y),1 f (x) Let s choose, length of chain. $ % & b=1 N=10,000! " # 1+x = min,1 1+y 2 2 $ % &
14 Example 1. Code in R N=10000 b=1 x_values=rep(0,n) x_cauchy=rep(0,n) x_axis=seq(-7,7,by=0.1) x_old=0 x_new=x_old for (i in 1:N) {y=rnorm(1,x_old,b) r=min((1+x_old^2)/(1+y^2),1) p=runif(1) if (p<r) (x_new=y) else (x_new=x_old);; x_values[i]=x_new x_old=x_new} x_cauchy=dcauchy(x_axis) plot(x_axis,x_cauchy,type="p",col="black") points(density(x_values),type="l",col="red",lwd=3)
15
16 Gibbs Sampling Gibbs Sampling is the easiest to use MCMC algorithm in case of dealing with high-dimensional problems as it helps to turn a high-dimensional problem into several one-dimensional problems. One of the examples of high-dimensional problems is hierarchical model
17 Hierarchical model. Example1 Posterior distribution on joint model specified. (θ,σ 2 ) X i ~(θ,σ 2 ), i=1,...,n, θ~n(θ,τ 2 0 ), σ 2 ~IG(a,b), θ 0,τ 2, a, b associated with the
18 Gibbs Sampling algorithm (X,Y) f X, Y (x, y) Suppose that has density. Suppose that it is possible to simulate from the conditional distributions and. Let be f X Y (x y) f Y X (y x) (X 0,Y 0 ) (X 0,Y 0 ),...,(X n,y n ) (X n+1,y n+1 ) starting values. Assume we have drawn Then the Gibbs sampling algorithm for getting : X n+1 ~ f X Y (x Y n ) Y n+1 ~ f Y X (y X n+1 ) repeat
19 Posteriors for the Example1 X i ~(θ,σ 2 ), i=1,...,n, θ~n(θ,τ 2 0 ), σ 2 ~IG(a,b), θ,τ 2 0, a, b f(θ x,σ 2! )~N σ 2 σ 2 +nτ 2θ 0+ nτ 2 σ 2 +nτ x, σ 2 τ 2 # # # # 2 σ 2 +nτ 2 # " " $ f(σ 2 $ " x,θ)~ig n+a, 1 x 2 2 θ $ % $ $ i ' $ '2 $ # & +b $ i $ # % ' ' ' ' ' ' ' ' & $ & & & & & %
20 x=rnorm(1000,10,2) n=length(x) a=3;; b=3 tau2=10 theta0=5 Nsim=5000 Example 1. Code in R xbar=mean(x) sh1=(n/2)+a sigma2=theta=rep(0,nsim) #init arrays sigma2[1]=1/rgamma(1,shape=a,rate=b) #init chains B=sigma2[1]/(sigma2[1]+n*tau2) theta[1]=rnorm(1,m=b*theta0+(1-b)*xbar,sd=sqrt(tau2*b)) for (i in 2:Nsim){ B=sigma2[i-1]/(sigma2[i-1]+n*tau2) theta[i]=rnorm(1,m=b*theta0+(1-b)*xbar,sd=sqrt(tau2*b)) ra1=(1/2)*(sum((x-theta[i])^2))+b sigma2[i]=1/rgamma(1,shape=sh1,rate=ra1) } mean(theta[3000:5000]) mean(sigma2[3000:5000])
21 Conjugate priors In Bayesian probability theory, if posterior distributions are in the same family as the prior distributions, then both prior and posterior are called conjugate distributions and the prior is called conjugate prior. P(θ D)= P(θ)P(D θ) P(θ)P(D θ)dθ
22 Conjugate priors. Example x~n(µ,σ (2) ) x σ (2) Let s consider normal distribution. For normally distributed with fixed variance, the conjugate prior is also normally distributed. For (2) prior µ~n(µ 0,σ 0 ) posterior will be in the form: µ x,σ (2) ~N( ˆµ 0, ˆ σ 0 (2) ), ˆµ 0 = σ (2) x+ σ (2) 0 σ (2) 0 +σ (2) σ (2) +σ µ (2) 0, 0 0 ˆ σ 0 (2) = σ (2) σ 0 (2) σ (2) +σ 0 (2)
23 Conjugate priors. Example Let s consider normal distribution. x~n(µ,σ (2) ) x µ For normally distributed with fixed mean, the conjugate prior is distributed according to inversegamma distribution. For prior σ (2) ~IG(α,β) P(x,µ σ (2) )= 1 σ 2π exp( (x µ)2 ) (σ 2 ) 1/2 exp( 1/2(x µ)2 ) 2σ 2 σ 2 P(σ (2) )=IG(α,β)= β α (σ (2) ) ( α 1) Γ(α) $ exp β & & & & & % σ (2) P(σ (2) x,µ) (σ 2 ) (α+1/2) 1 & & exp β 1/2(x µ)2 & ˆα=α+1/2, ˆβ=β+1(x µ) 2 $ & & % σ (2), ' ) ) ) ) ) ( ' ) ) ) ) ) (
24 ABC SMC A number of sampled parameter values (particles) {θ (1),...,θ (n) } π(θ), sampled from the prior distribution, are propagated through a sequence of intermediate distributions, π(θ d(x 0, x * ) ε i ), i=1,...,t 1, until it represents a sample from the target distribution π(θ d(x 0, x * ) ε T ). The tolerances what mean gradual evolving towards the target posterior. ε 1 >...>ε T 0 For sufficiently large numbers of particles, this approach avoid the problem of getting stuck in areas of low probability (as in ABC MCMC)
25 ABC SMC Algorithm S1. Initialize. Set the population indicator. S2.0 Set the particle indicator. S2.1 If, sample independently from. Else, sample with weights and perturb the particle to obtain, where is a perturbation kernel., return to S2.1. Simulate a candidate dataset. If ε 1,...,ε T t=0 i=1 t=0 θ ** π(θ) θ * (i) from the previous population {θ t 1 } θ ** ~K t (θ θ * ) If π(θ ** )=0 d(x *, x 0 ) ε t w t 1 K t, return to S2.1. x * ~ f(x θ ** )
26 θ t (i) =θ ** ABC SMC Algorithm S2.2 Set and calculate the weight for particle # % 1, if t=0, % (i) w t = π(θ (i) % t ) $, if t>0. % N w(j) t 1Kt (θ (j),θ % (i) t 1 t ) % &% j=1 If, set, go to S2.1. S3 Normalize the weights. t<t i<n i=i+1 t=t+1 If, set, go to S2.0. θ t (i)
27 Ovarian Cancer case study. CA125
28 Risk calculation
29 Change-point hierarchical model for CA125 Controls: Cases: Y ij t ij ~N(θ i,σ 2 ) Y ij t ij,{i i =0}~N(θ i,σ 2 ) Y ij t ij,{i i =1}~N(θ i +γ i (tij τ i) +,σ 2 )
30
31
32 Conditional distributions
33
34 Conclusion 1. ABC methods has great impact on parameters estimation. 1. A lot of applied problems can be reduced to hierarchical model 2. Gibbs Sampling Algorithm is most useful in dealing with hierarchical models
35 Literature 1. Steven J. Skates, Donna K. Pauler, Ian J. Jacobs. Screening Based on the Risk of Cancer Calculation from Bayesian Hierarchical Changepoint and Mixture Models of Longitudinal Markers. Journal of the American Statistical Society, vol. 96 (2001). 2. Wasserman L. All of Statistics. A concise course in Statistical Inference, Springer, Tina Toni, David Welch, Natalja Strelkowa, Andreas Ipsen, Michael P.H. Stumpf. Approximate Bayesian computation scheme for parameter inference and model selection in dynamical systems. Journal of the royal society, 6, (2009). 4. Robert P. Christian, Casella George. Introducing Monte Carlo Methods with R, Springer, 2009.
36 Data processing with caret package in R Data preprocessing Data splitting Data processing Model comparison
37 Data preprocessing preprocess Standardizing Transformation Imputing
38 Data preprocessing. Example data(bloodbrain) # contains array bbbdescr bbbdescr=bbbdescr[,-3] preproc <- preprocess(bbbdescr,method = c("center", "scale")) data <- predict(preproc, bbbdescr) mean(bbbdescr[,1]) mean(data[,1]) var(data[,1]) mean(bbbdescr[,2]) mean(data[,2]) var(data[,2])
39 Data splitting createdatapartition # training/test partition createresample # bootstrap samples createfolds # split the data into k groups createtimeslices # is used for time series data
40 Data splitting. Example data(bloodbrain) # contains array bbbdescr bbbdescr=bbbdescr[,-3] train_part <- createdatapartition(y=bbbdescr[,1], p=0.75, list=false) training <- bbbdescr[train_part,] testing <- bbbdescr[-train_part,] dim(bbbdescr) dim(training) dim(testing)
41 Data processing. Resampling train method= boot # bootstraping boot632 # bootstrapping with adjustment cv # cross validation repeatedcv # repeated cross validation LOOCV # leave one out cross validation
42 Data processing. Example library(mlbench) data(sonar) set.seed(107) intrain <- createdatapartition(y = Sonar$Class, p =.75, list = FALSE) training <- Sonar[ intrain,] testing <- Sonar[-inTrain,] plsfit <- train(class ~., data = training, method = "knn", preproc = c("center", "scale")) plsclasses <- predict(plsfit, newdata = testing) plsclasses
43 Model comparison. Metric options confusionmatrix Continuous outcomes: RMSE # root mean squared error RSquared # R^2 from regression models Categorical outcomes: Accuracy # fraction of correct classes Kappa # measure of concordance
44 Model comparison. Example names(getmodelinfo()) plsfit <- train(class ~., data = training, method = "knn", preproc = c("center", "scale")) plsclasses <- predict(plsfit, newdata = testing) confusionmatrix(data = plsclasses, testing$class) plsfit <- train(class ~., data = training, method = "pls", preproc = c("center", "scale")) plsclasses <- predict(plsfit, newdata = testing) confusionmatrix(data = plsclasses, testing$class) plsfit <- train(class ~., data = training, method = "cforest", preproc = c("center", "scale")) plsclasses <- predict(plsfit, newdata = testing) confusionmatrix(data = plsclasses, testing$class)
45 Literature 1. Max Kuhn. A Short Introduction to the caret Package (2014). 2. Model training and tuning:
46 Questions
A Short Introduction to the caret Package
A Short Introduction to the caret Package Max Kuhn max.kuhn@pfizer.com October 28, 2016 The caret package (short for classification and regression training) contains functions to streamline the model training
More informationA Short Introduction to the caret Package
A Short Introduction to the caret Package Max Kuhn max.kuhn@pfizer.com February 12, 2013 The caret package (short for classification and regression training) contains functions to streamline the model
More informationApproximate Bayesian Computation. Alireza Shafaei - April 2016
Approximate Bayesian Computation Alireza Shafaei - April 2016 The Problem Given a dataset, we are interested in. The Problem Given a dataset, we are interested in. The Problem Given a dataset, we are interested
More informationOptimization Methods III. The MCMC. Exercises.
Aula 8. Optimization Methods III. Exercises. 0 Optimization Methods III. The MCMC. Exercises. Anatoli Iambartsev IME-USP Aula 8. Optimization Methods III. Exercises. 1 [RC] A generic Markov chain Monte
More informationRolling Markov Chain Monte Carlo
Rolling Markov Chain Monte Carlo Din-Houn Lau Imperial College London Joint work with Axel Gandy 4 th September 2013 RSS Conference 2013: Newcastle Output predicted final ranks of the each team. Updates
More informationRolling Markov Chain Monte Carlo
Rolling Markov Chain Monte Carlo Din-Houn Lau Imperial College London Joint work with Axel Gandy 4 th July 2013 Predict final ranks of the each team. Updates quick update of predictions. Accuracy control
More informationMCMC Methods for data modeling
MCMC Methods for data modeling Kenneth Scerri Department of Automatic Control and Systems Engineering Introduction 1. Symposium on Data Modelling 2. Outline: a. Definition and uses of MCMC b. MCMC algorithms
More informationBayesian Statistics Group 8th March Slice samplers. (A very brief introduction) The basic idea
Bayesian Statistics Group 8th March 2000 Slice samplers (A very brief introduction) The basic idea lacements To sample from a distribution, simply sample uniformly from the region under the density function
More informationMarkov Chain Monte Carlo (part 1)
Markov Chain Monte Carlo (part 1) Edps 590BAY Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Spring 2018 Depending on the book that you select for
More informationMarkov chain Monte Carlo methods
Markov chain Monte Carlo methods (supplementary material) see also the applet http://www.lbreyer.com/classic.html February 9 6 Independent Hastings Metropolis Sampler Outline Independent Hastings Metropolis
More informationStatistical Matching using Fractional Imputation
Statistical Matching using Fractional Imputation Jae-Kwang Kim 1 Iowa State University 1 Joint work with Emily Berg and Taesung Park 1 Introduction 2 Classical Approaches 3 Proposed method 4 Application:
More informationA GENERAL GIBBS SAMPLING ALGORITHM FOR ANALYZING LINEAR MODELS USING THE SAS SYSTEM
A GENERAL GIBBS SAMPLING ALGORITHM FOR ANALYZING LINEAR MODELS USING THE SAS SYSTEM Jayawant Mandrekar, Daniel J. Sargent, Paul J. Novotny, Jeff A. Sloan Mayo Clinic, Rochester, MN 55905 ABSTRACT A general
More informationBART STAT8810, Fall 2017
BART STAT8810, Fall 2017 M.T. Pratola November 1, 2017 Today BART: Bayesian Additive Regression Trees BART: Bayesian Additive Regression Trees Additive model generalizes the single-tree regression model:
More informationMonte Carlo for Spatial Models
Monte Carlo for Spatial Models Murali Haran Department of Statistics Penn State University Penn State Computational Science Lectures April 2007 Spatial Models Lots of scientific questions involve analyzing
More informationMissing Data and Imputation
Missing Data and Imputation NINA ORWITZ OCTOBER 30 TH, 2017 Outline Types of missing data Simple methods for dealing with missing data Single and multiple imputation R example Missing data is a complex
More informationApproximate Bayesian Computation using Auxiliary Models
Approximate Bayesian Computation using Auxiliary Models Tony Pettitt Co-authors Chris Drovandi, Malcolm Faddy Queensland University of Technology Brisbane MCQMC February 2012 Tony Pettitt () ABC using
More information1 Methods for Posterior Simulation
1 Methods for Posterior Simulation Let p(θ y) be the posterior. simulation. Koop presents four methods for (posterior) 1. Monte Carlo integration: draw from p(θ y). 2. Gibbs sampler: sequentially drawing
More informationBayesian Statistics with a Smile: a Resampling-Sampling Perspective
Bayesian Statistics with a Smile: a Resampling-Sampling Perspective Hedibert F. Lopes The University of Chicago Booth School of Business 5807 S. Woodlawn Ave., Chicago, IL, 60637. hlopes@chicagobooth.edu
More informationConvexization in Markov Chain Monte Carlo
in Markov Chain Monte Carlo 1 IBM T. J. Watson Yorktown Heights, NY 2 Department of Aerospace Engineering Technion, Israel August 23, 2011 Problem Statement MCMC processes in general are governed by non
More informationMonte Carlo Techniques for Bayesian Statistical Inference A comparative review
1 Monte Carlo Techniques for Bayesian Statistical Inference A comparative review BY Y. FAN, D. WANG School of Mathematics and Statistics, University of New South Wales, Sydney 2052, AUSTRALIA 18th January,
More informationMissing Data Analysis for the Employee Dataset
Missing Data Analysis for the Employee Dataset 67% of the observations have missing values! Modeling Setup Random Variables: Y i =(Y i1,...,y ip ) 0 =(Y i,obs, Y i,miss ) 0 R i =(R i1,...,r ip ) 0 ( 1
More informationAn Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework
IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXX 23 An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework Ji Won Yoon arxiv:37.99v [cs.lg] 3 Jul 23 Abstract In order to cluster
More informationSampling informative/complex a priori probability distributions using Gibbs sampling assisted by sequential simulation
Sampling informative/complex a priori probability distributions using Gibbs sampling assisted by sequential simulation Thomas Mejer Hansen, Klaus Mosegaard, and Knud Skou Cordua 1 1 Center for Energy Resources
More informationChapter 1. Introduction
Chapter 1 Introduction A Monte Carlo method is a compuational method that uses random numbers to compute (estimate) some quantity of interest. Very often the quantity we want to compute is the mean of
More informationImage analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis
7 Computer Vision and Classification 413 / 458 Computer Vision and Classification The k-nearest-neighbor method The k-nearest-neighbor (knn) procedure has been used in data analysis and machine learning
More informationBayesian Modelling with JAGS and R
Bayesian Modelling with JAGS and R Martyn Plummer International Agency for Research on Cancer Rencontres R, 3 July 2012 CRAN Task View Bayesian Inference The CRAN Task View Bayesian Inference is maintained
More informationMSA101/MVE Lecture 5
MSA101/MVE187 2017 Lecture 5 Petter Mostad Chalmers University September 12, 2017 1 / 15 Importance sampling MC integration computes h(x)f (x) dx where f (x) is a probability density function, by simulating
More informationSTAT 725 Notes Monte Carlo Integration
STAT 725 Notes Monte Carlo Integration Two major classes of numerical problems arise in statistical inference: optimization and integration. We have already spent some time discussing different optimization
More informationNested Sampling: Introduction and Implementation
UNIVERSITY OF TEXAS AT SAN ANTONIO Nested Sampling: Introduction and Implementation Liang Jing May 2009 1 1 ABSTRACT Nested Sampling is a new technique to calculate the evidence, Z = P(D M) = p(d θ, M)p(θ
More informationQuantitative Biology II!
Quantitative Biology II! Lecture 3: Markov Chain Monte Carlo! March 9, 2015! 2! Plan for Today!! Introduction to Sampling!! Introduction to MCMC!! Metropolis Algorithm!! Metropolis-Hastings Algorithm!!
More informationINLA: Integrated Nested Laplace Approximations
INLA: Integrated Nested Laplace Approximations John Paige Statistics Department University of Washington October 10, 2017 1 The problem Markov Chain Monte Carlo (MCMC) takes too long in many settings.
More informationThe Bootstrap and Jackknife
The Bootstrap and Jackknife Summer 2017 Summer Institutes 249 Bootstrap & Jackknife Motivation In scientific research Interest often focuses upon the estimation of some unknown parameter, θ. The parameter
More informationA Short History of Markov Chain Monte Carlo
A Short History of Markov Chain Monte Carlo Christian Robert and George Casella 2010 Introduction Lack of computing machinery, or background on Markov chains, or hesitation to trust in the practicality
More informationNon-parametric Approximate Bayesian Computation for Expensive Simulators
Non-parametric Approximate Bayesian Computation for Expensive Simulators Steven Laan 6036031 Master s thesis 42 EC Master s programme Artificial Intelligence University of Amsterdam Supervisor Ted Meeds
More informationGiRaF: a toolbox for Gibbs Random Fields analysis
GiRaF: a toolbox for Gibbs Random Fields analysis Julien Stoehr *1, Pierre Pudlo 2, and Nial Friel 1 1 University College Dublin 2 Aix-Marseille Université February 24, 2016 Abstract GiRaF package offers
More information10-701/15-781, Fall 2006, Final
-7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly
More informationModeling Criminal Careers as Departures From a Unimodal Population Age-Crime Curve: The Case of Marijuana Use
Modeling Criminal Careers as Departures From a Unimodal Population Curve: The Case of Marijuana Use Donatello Telesca, Elena A. Erosheva, Derek A. Kreader, & Ross Matsueda April 15, 2014 extends Telesca
More informationBayesian Spatiotemporal Modeling with Hierarchical Spatial Priors for fmri
Bayesian Spatiotemporal Modeling with Hierarchical Spatial Priors for fmri Galin L. Jones 1 School of Statistics University of Minnesota March 2015 1 Joint with Martin Bezener and John Hughes Experiment
More information10.4 Linear interpolation method Newton s method
10.4 Linear interpolation method The next best thing one can do is the linear interpolation method, also known as the double false position method. This method works similarly to the bisection method by
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 17 EM CS/CNS/EE 155 Andreas Krause Announcements Project poster session on Thursday Dec 3, 4-6pm in Annenberg 2 nd floor atrium! Easels, poster boards and cookies
More informationCSCI 599 Class Presenta/on. Zach Levine. Markov Chain Monte Carlo (MCMC) HMM Parameter Es/mates
CSCI 599 Class Presenta/on Zach Levine Markov Chain Monte Carlo (MCMC) HMM Parameter Es/mates April 26 th, 2012 Topics Covered in this Presenta2on A (Brief) Review of HMMs HMM Parameter Learning Expecta2on-
More informationPerformance of Sequential Imputation Method in Multilevel Applications
Section on Survey Research Methods JSM 9 Performance of Sequential Imputation Method in Multilevel Applications Enxu Zhao, Recai M. Yucel New York State Department of Health, 8 N. Pearl St., Albany, NY
More informationLinear Modeling with Bayesian Statistics
Linear Modeling with Bayesian Statistics Bayesian Approach I I I I I Estimate probability of a parameter State degree of believe in specific parameter values Evaluate probability of hypothesis given the
More information10703 Deep Reinforcement Learning and Control
10703 Deep Reinforcement Learning and Control Russ Salakhutdinov Machine Learning Department rsalakhu@cs.cmu.edu Policy Gradient I Used Materials Disclaimer: Much of the material and slides for this lecture
More informationClustering Relational Data using the Infinite Relational Model
Clustering Relational Data using the Infinite Relational Model Ana Daglis Supervised by: Matthew Ludkin September 4, 2015 Ana Daglis Clustering Data using the Infinite Relational Model September 4, 2015
More informationStochastic Simulation: Algorithms and Analysis
Soren Asmussen Peter W. Glynn Stochastic Simulation: Algorithms and Analysis et Springer Contents Preface Notation v xii I What This Book Is About 1 1 An Illustrative Example: The Single-Server Queue 1
More informationBayesian Estimation for Skew Normal Distributions Using Data Augmentation
The Korean Communications in Statistics Vol. 12 No. 2, 2005 pp. 323-333 Bayesian Estimation for Skew Normal Distributions Using Data Augmentation Hea-Jung Kim 1) Abstract In this paper, we develop a MCMC
More informationModified Metropolis-Hastings algorithm with delayed rejection
Modified Metropolis-Hastings algorithm with delayed reection K.M. Zuev & L.S. Katafygiotis Department of Civil Engineering, Hong Kong University of Science and Technology, Hong Kong, China ABSTRACT: The
More informationMachine Learning (BSMC-GA 4439) Wenke Liu
Machine Learning (BSMC-GA 4439) Wenke Liu 01-31-017 Outline Background Defining proximity Clustering methods Determining number of clusters Comparing two solutions Cluster analysis as unsupervised Learning
More informationA noninformative Bayesian approach to small area estimation
A noninformative Bayesian approach to small area estimation Glen Meeden School of Statistics University of Minnesota Minneapolis, MN 55455 glen@stat.umn.edu September 2001 Revised May 2002 Research supported
More informationOverview. Monte Carlo Methods. Statistics & Bayesian Inference Lecture 3. Situation At End Of Last Week
Statistics & Bayesian Inference Lecture 3 Joe Zuntz Overview Overview & Motivation Metropolis Hastings Monte Carlo Methods Importance sampling Direct sampling Gibbs sampling Monte-Carlo Markov Chains Emcee
More informationReal Coded Genetic Algorithm Particle Filter for Improved Performance
Proceedings of 2012 4th International Conference on Machine Learning and Computing IPCSIT vol. 25 (2012) (2012) IACSIT Press, Singapore Real Coded Genetic Algorithm Particle Filter for Improved Performance
More informationMCMC Diagnostics. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) MCMC Diagnostics MATH / 24
MCMC Diagnostics Yingbo Li Clemson University MATH 9810 Yingbo Li (Clemson) MCMC Diagnostics MATH 9810 1 / 24 Convergence to Posterior Distribution Theory proves that if a Gibbs sampler iterates enough,
More informationMonte Carlo Methods and Statistical Computing: My Personal E
Monte Carlo Methods and Statistical Computing: My Personal Experience Department of Mathematics & Statistics Indian Institute of Technology Kanpur November 29, 2014 Outline Preface 1 Preface 2 3 4 5 6
More informationJournal of Mathematical Psychology
Journal of Mathematical Psychology 56 (2012) 375 385 Contents lists available at SciVerse ScienceDirect Journal of Mathematical Psychology journal homepage: www.elsevier.com/locate/jmp Approximate Bayesian
More informationParticle Filters for Visual Tracking
Particle Filters for Visual Tracking T. Chateau, Pascal Institute, Clermont-Ferrand 1 Content Particle filtering: a probabilistic framework SIR particle filter MCMC particle filter RJMCMC particle filter
More informationImage Segmentation using Gaussian Mixture Models
Image Segmentation using Gaussian Mixture Models Rahman Farnoosh, Gholamhossein Yari and Behnam Zarpak Department of Applied Mathematics, University of Science and Technology, 16844, Narmak,Tehran, Iran
More informationHierarchical Bayesian Modeling with Ensemble MCMC. Eric B. Ford (Penn State) Bayesian Computing for Astronomical Data Analysis June 12, 2014
Hierarchical Bayesian Modeling with Ensemble MCMC Eric B. Ford (Penn State) Bayesian Computing for Astronomical Data Analysis June 12, 2014 Simple Markov Chain Monte Carlo Initialise chain with θ 0 (initial
More informationProbabilistic Graphical Models
Overview of Part Two Probabilistic Graphical Models Part Two: Inference and Learning Christopher M. Bishop Exact inference and the junction tree MCMC Variational methods and EM Example General variational
More informationCS281 Section 9: Graph Models and Practical MCMC
CS281 Section 9: Graph Models and Practical MCMC Scott Linderman November 11, 213 Now that we have a few MCMC inference algorithms in our toolbox, let s try them out on some random graph models. Graphs
More informationExam Issued: May 29, 2017, 13:00 Hand in: May 29, 2017, 16:00
P. Hadjidoukas, C. Papadimitriou ETH Zentrum, CTL E 13 CH-8092 Zürich High Performance Computing for Science and Engineering II Exam Issued: May 29, 2017, 13:00 Hand in: May 29, 2017, 16:00 Spring semester
More informationMonte Carlo sampling
1 y u theta 0 x 1 Monte Carlo sampling Problem 1 Suppose we want to sample uniformly at random from the triangle defined by the points (0,0), (0,1), (1,0). First Sampling Algorithm: We decide to do this
More informationPackage HKprocess. R topics documented: September 6, Type Package. Title Hurst-Kolmogorov process. Version
Package HKprocess September 6, 2014 Type Package Title Hurst-Kolmogorov process Version 0.0-1 Date 2014-09-06 Author Maintainer Imports MCMCpack (>= 1.3-3), gtools(>= 3.4.1) Depends
More informationPackage mederrrank. R topics documented: July 8, 2015
Package mederrrank July 8, 2015 Title Bayesian Methods for Identifying the Most Harmful Medication Errors Version 0.0.8 Date 2015-07-06 Two distinct but related statistical approaches to the problem of
More informationarxiv: v3 [stat.co] 27 Apr 2012
A multi-point Metropolis scheme with generic weight functions arxiv:1112.4048v3 stat.co 27 Apr 2012 Abstract Luca Martino, Victor Pascual Del Olmo, Jesse Read Department of Signal Theory and Communications,
More informationA Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data
A Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data Seyoung Kim, Padhraic Smyth, and Hal Stern Bren School of Information and Computer Sciences University of California,
More informationTopics in Machine Learning-EE 5359 Model Assessment and Selection
Topics in Machine Learning-EE 5359 Model Assessment and Selection Ioannis D. Schizas Electrical Engineering Department University of Texas at Arlington 1 Training and Generalization Training stage: Utilizing
More informationGT "Calcul Ensembliste"
GT "Calcul Ensembliste" Beyond the bounded error framework for non linear state estimation Fahed Abdallah Université de Technologie de Compiègne 9 Décembre 2010 Fahed Abdallah GT "Calcul Ensembliste" 9
More informationADAPTIVE METROPOLIS-HASTINGS SAMPLING, OR MONTE CARLO KERNEL ESTIMATION
ADAPTIVE METROPOLIS-HASTINGS SAMPLING, OR MONTE CARLO KERNEL ESTIMATION CHRISTOPHER A. SIMS Abstract. A new algorithm for sampling from an arbitrary pdf. 1. Introduction Consider the standard problem of
More informationAccelerated Markov Chain Monte Carlo Sampling in Electrical Capacitance Tomography
Accelerated Markov Chain Monte Carlo Sampling in Electrical Capacitance Tomography Daniel Watzenig, Markus Neumayer and Colin Fox Institute of Electrical Measurement and Measurement Signal Processing,
More informationOne-Shot Learning with a Hierarchical Nonparametric Bayesian Model
One-Shot Learning with a Hierarchical Nonparametric Bayesian Model R. Salakhutdinov, J. Tenenbaum and A. Torralba MIT Technical Report, 2010 Presented by Esther Salazar Duke University June 10, 2011 E.
More informationMachine Learning (BSMC-GA 4439) Wenke Liu
Machine Learning (BSMC-GA 4439) Wenke Liu 01-25-2018 Outline Background Defining proximity Clustering methods Determining number of clusters Other approaches Cluster analysis as unsupervised Learning Unsupervised
More informationMarkov Random Fields and Gibbs Sampling for Image Denoising
Markov Random Fields and Gibbs Sampling for Image Denoising Chang Yue Electrical Engineering Stanford University changyue@stanfoed.edu Abstract This project applies Gibbs Sampling based on different Markov
More informationGraphical Models, Bayesian Method, Sampling, and Variational Inference
Graphical Models, Bayesian Method, Sampling, and Variational Inference With Application in Function MRI Analysis and Other Imaging Problems Wei Liu Scientific Computing and Imaging Institute University
More informationBeviMed Guide. Daniel Greene
BeviMed Guide Daniel Greene 1 Introduction BeviMed [1] is a procedure for evaluating the evidence of association between allele configurations across rare variants, typically within a genomic locus, and
More informationAnalysis of Incomplete Multivariate Data
Analysis of Incomplete Multivariate Data J. L. Schafer Department of Statistics The Pennsylvania State University USA CHAPMAN & HALL/CRC A CR.C Press Company Boca Raton London New York Washington, D.C.
More informationMULTI-DIMENSIONAL MONTE CARLO INTEGRATION
CS580: Computer Graphics KAIST School of Computing Chapter 3 MULTI-DIMENSIONAL MONTE CARLO INTEGRATION 2 1 Monte Carlo Integration This describes a simple technique for the numerical evaluation of integrals
More informationResampling Methods. Levi Waldron, CUNY School of Public Health. July 13, 2016
Resampling Methods Levi Waldron, CUNY School of Public Health July 13, 2016 Outline and introduction Objectives: prediction or inference? Cross-validation Bootstrap Permutation Test Monte Carlo Simulation
More informationEstimating Labels from Label Proportions
Estimating Labels from Label Proportions Novi Quadrianto Novi.Quad@gmail.com The Australian National University, Australia NICTA, Statistical Machine Learning Program, Australia Joint work with Alex Smola,
More informationSemiparametric Mixed Effecs with Hierarchical DP Mixture
Semiparametric Mixed Effecs with Hierarchical DP Mixture R topics documented: April 21, 2007 hdpm-package........................................ 1 hdpm............................................ 2 hdpmfitsetup........................................
More informationBayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19
Bayesian Methods David Rosenberg New York University April 11, 2017 David Rosenberg (New York University) DS-GA 1003 April 11, 2017 1 / 19 Classical Statistics Classical Statistics David Rosenberg (New
More informationHybrid Quasi-Monte Carlo Method for the Simulation of State Space Models
The Tenth International Symposium on Operations Research and Its Applications (ISORA 211) Dunhuang, China, August 28 31, 211 Copyright 211 ORSC & APORC, pp. 83 88 Hybrid Quasi-Monte Carlo Method for the
More informationA Deterministic Global Optimization Method for Variational Inference
A Deterministic Global Optimization Method for Variational Inference Hachem Saddiki Mathematics and Statistics University of Massachusetts, Amherst saddiki@math.umass.edu Andrew C. Trapp Operations and
More informationCurve Sampling and Geometric Conditional Simulation
Curve Sampling and Geometric Conditional Simulation Ayres Fan Joint work with John Fisher, William Wells, Jonathan Kane, and Alan Willsky S S G Massachusetts Institute of Technology September 19, 2007
More informationModel validation through "Posterior predictive checking" and "Leave-one-out"
Workshop of the Bayes WG / IBS-DR Mainz, 2006-12-01 G. Nehmiz M. Könen-Bergmann Model validation through "Posterior predictive checking" and "Leave-one-out" Overview The posterior predictive distribution
More informationProject Report: "Bayesian Spam Filter"
Humboldt-Universität zu Berlin Lehrstuhl für Maschinelles Lernen Sommersemester 2016 Maschinelles Lernen 1 Project Report: "Bayesian E-Mail Spam Filter" The Bayesians Sabine Bertram, Carolina Gumuljo,
More informationInstability, Sensitivity, and Degeneracy of Discrete Exponential Families
Instability, Sensitivity, and Degeneracy of Discrete Exponential Families Michael Schweinberger Pennsylvania State University ONR grant N00014-08-1-1015 Scalable Methods for the Analysis of Network-Based
More informationBayesian data analysis using R
Bayesian data analysis using R BAYESIAN DATA ANALYSIS USING R Jouni Kerman, Samantha Cook, and Andrew Gelman Introduction Bayesian data analysis includes but is not limited to Bayesian inference (Gelman
More informationA Bayesian approach to artificial neural network model selection
A Bayesian approach to artificial neural network model selection Kingston, G. B., H. R. Maier and M. F. Lambert Centre for Applied Modelling in Water Engineering, School of Civil and Environmental Engineering,
More informationLevel-set MCMC Curve Sampling and Geometric Conditional Simulation
Level-set MCMC Curve Sampling and Geometric Conditional Simulation Ayres Fan John W. Fisher III Alan S. Willsky February 16, 2007 Outline 1. Overview 2. Curve evolution 3. Markov chain Monte Carlo 4. Curve
More informationRecap: The E-M algorithm. Biostatistics 615/815 Lecture 22: Gibbs Sampling. Recap - Local minimization methods
Recap: The E-M algorithm Biostatistics 615/815 Lecture 22: Gibbs Sampling Expectation step (E-step) Given the current estimates of parameters λ (t), calculate the conditional distribution of latent variable
More informationBayesian Computation: Overview and Methods for Low-Dimensional Models
Bayesian Computation: Overview and Methods for Low-Dimensional Models Tom Loredo Dept. of Astronomy, Cornell University http://www.astro.cornell.edu/staff/loredo/bayes/ IAC Winter School, 3 4 Nov 2014
More informationStatistical Analysis Using Combined Data Sources: Discussion JPSM Distinguished Lecture University of Maryland
Statistical Analysis Using Combined Data Sources: Discussion 2011 JPSM Distinguished Lecture University of Maryland 1 1 University of Michigan School of Public Health April 2011 Complete (Ideal) vs. Observed
More informationMonte Carlo Optimization Introduction
Monte Carlo Methods with R: Monte Carlo Optimization [80] Monte Carlo Optimization Introduction Optimization problems can mostly be seen as one of two kinds: Find the extrema of a function h(θ) over a
More informationIntroduction to Optimization
Introduction to Optimization Second Order Optimization Methods Marc Toussaint U Stuttgart Planned Outline Gradient-based optimization (1st order methods) plain grad., steepest descent, conjugate grad.,
More informationA study of classification algorithms using Rapidminer
Volume 119 No. 12 2018, 15977-15988 ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu ijpam.eu A study of classification algorithms using Rapidminer Dr.J.Arunadevi 1, S.Ramya 2, M.Ramesh Raja
More informationTransitivity and Triads
1 / 32 Tom A.B. Snijders University of Oxford May 14, 2012 2 / 32 Outline 1 Local Structure Transitivity 2 3 / 32 Local Structure in Social Networks From the standpoint of structural individualism, one
More informationOutline. Bayesian Data Analysis Hierarchical models. Rat tumor data. Errandum: exercise GCSR 3.11
Outline Bayesian Data Analysis Hierarchical models Helle Sørensen May 15, 2009 Today: More about the rat tumor data: model, derivation of posteriors, the actual computations in R. : a hierarchical normal
More informationAn Introduction to Markov Chain Monte Carlo
An Introduction to Markov Chain Monte Carlo Markov Chain Monte Carlo (MCMC) refers to a suite of processes for simulating a posterior distribution based on a random (ie. monte carlo) process. In other
More informationChapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea
Chapter 3 Bootstrap 3.1 Introduction The estimation of parameters in probability distributions is a basic problem in statistics that one tends to encounter already during the very first course on the subject.
More information