Outline. Bayesian Data Analysis Hierarchical models. Rat tumor data. Errandum: exercise GCSR 3.11
|
|
- Clement Lucas
- 5 years ago
- Views:
Transcription
1 Outline Bayesian Data Analysis Hierarchical models Helle Sørensen May 15, 2009 Today: More about the rat tumor data: model, derivation of posteriors, the actual computations in R. : a hierarchical normal model, derivation of posteriors. The actual comutations will be an exercise next week. Future: Chapter 10 11: MCMC simulation. Tuesday May 26: Irene Mantzouni. Be there! Bayes (Lecture 6) Hierarchical models 1 / 22 Bayes (Lecture 6) Hierarchical models 2 / 22 Errandum: exercise GCSR 3.11 Rat tumor data Remember the problems with in exercise 3.11 (bioassay data)? Posteriors Posteriors for α: Normal prior: 0.96 (-0.71, 2.82) Uniform prior: 1.32 (-0.57,3.80) As we would expect from the figure. More information in normal prior. Results not very robust (small dataset)! GCSR, page and Interested in θ the probability of tumor in a population of female lab rats of type F344 that receice a zero dose of a drug. Data: 71 experiments with female type F344 rats In j th study: n j rats of which y j developed a tumor NB.: In Section 5.1, data only consists of experiment 71 only, the rest are used as prior information. I find the section confusing and non-illuminating Bayes (Lecture 6) Hierarchical models 3 / 22 Bayes (Lecture 6) Hierarchical models 4 / 22
2 Rat tumor data: the hierarchical model Rat tumor data Three levels: The data: conditionally on θ 1,...,θ the data y 1,...,y are independent with y j p(y j θ j ). The random effects: conditionally of φ the θ s are iid. with θ j p(θ j φ). Recall: θ j s are not observed. The (hyper)prior distribution p(φ) Here: φ = (α,β) and y j θ j Bin(n j,θ j ); θ j Beta(α,β); p(α,β) = (α + β) 5/2. Note: for now, the conditional dist n of y j given θ j depends on θ j only (no covariates). The strategy will be the following. 0. Set up the model: p(y θ) and p(θ α,β). Note: the hyperprior p(α,β) is not yet specified. Enters as p(α,β) in oint posterior of (α,β,θ) 2. Posterior of θ for fixed α and β 3. Posterior of (α,β) Finally comes the actual prior and the actual computations: 4. Specify a hyperprior p(α,β) 5. Contours of posterior p(α,β y) 6. Posterior simulation of (α,β) 7. Posterior simulation of θ Bayes (Lecture 6) Hierarchical models 5 / 22 Bayes (Lecture 6) Hierarchical models 6 / Model 1 3. Posteriors 1. oint posterior of (α,β,θ): Recall the three levels: Given θ the y j s are independent and y j θ j Bin(n j,θ j ). Given (α,β) the θ j s are independent and θ j α,β Beta(α,β). Hyperprior (α,β) p(α,β) 1 p(α,β,θ y) p(α,β) B(α,β) n θ α+y j 1 j (1 θ j ) β +n j y j Posterior dist n of θ for fixed values of (α,β): independent and 3. Posterior of (α,β): p(α,β y) = θ j α,β,y Beta(α + y j,β + n j y j ) p(α,β,θ y)dθ 1 = p(α,β) B(α,β) n Beta(α + y j,β + n j y j ) Bayes (Lecture 6) Hierarchical models 7 / 22 Bayes (Lecture 6) Hierarchical models 8 / 22
3 4. Hyperprior distribution 4. Hyperprior distribution Recall: θ j α,β Beta(α,β) so E(θ j α,β) = α α+β. What is a reasonable prior distibution of (α,β)? Three different useful parameterizations: (α,β) ( ) (ξ 1,ξ 2 ) = α α+β,(α + β) 1/2 ( ) ) (η 1,η 2 ) = log( α β,log(α + β) Second try: flat in (ξ 1,ξ 2 ), ie. p(ξ 1,ξ 2 ) 1. What does this correspond to in the other parameterizations? By the transformation theorem multiply with the correct acobian we find that p(ξ 1,ξ 2 ) 1 corresponds to p(α,β) (α + β) 5/2 p(η 1,η 2 ) αβ(α + β) 5/2 First try: flat in (η 1,η 2 ), ie. p(η 1,η 2 ) 1. Problem: this yields an improper posterior. Bayes (Lecture 6) Hierarchical models 9 / 22 Bayes (Lecture 6) Hierarchical models 10 / Contour of posterior for (α,β) 6. Posterior simulation of (α,β) Want contour plot of the posterior of p(η 1,η 2 y). Strategy: Recall what is p(α,β y). Multiply with αβ due to the η-parameterization Need the density as a function of η rather than (α,β). Compute log p(η 1,η 2 y) and then p(η 1,η 2 y) on η-grid Since we now have (unnormalized) posterior density p(η 1,η 2 y) we can apply the grid simulation strategy : Simulate η 1 from its marginal posterior p(η 1 y) on grid-values Simulate η 2 from its marginal conditional p(η 2 y,eta 1 ) given η 1 on grid-values Add random jitter to get a continuous distribution Transform the simulated values of (η 1,η 2 ) to (α,β). Bayes (Lecture 6) Hierarchical models 11 / 22 Bayes (Lecture 6) Hierarchical models 12 / 22
4 7. Posterior simulation of θ Recall: conditional on (α,β,y) the θ s are independent with p(θ j α,β,y) Beta(α + y j,β + n j y j ). Hence, for each (α,β)-draw, draw θ 1,...,θ 71 according the relevant Beta-distribution. Plot of posterior means E(θ j y) or medians against y j /n j (page 131) shows that extreme values of y j /n j are shrunk towards the pooled mean Compromise between the complete pooling and the separate analyses. GCSR Section 5.5, page Model and analysis method described in Section schools Treatment is preparation for SAT-V test. Two groups of students at each school: treated and controls For each student: result of test Data: estimated treatment effect from each school (y 1,...,y 8 ) and corresponding SE s (σ 1,...,σ 8 ) Think of y j as average over treated students minus average over control students (although estimated from a regression model). The σ j are considered known (although estimated from student data) Interested in the treatment effect (θ) Bayes (Lecture 6) Hierarchical models 13 / 22 Bayes (Lecture 6) Hierarchical models 14 / 22 : complete pooling and separate analyses : hierarchical model Complete pooling: same θ for all schools, y j θ N(θ,σ 2 j ) MLE would be weighted average of y 1,...,y 8 Bayesian analysis would combine this average and a prior for θ Completely separate analysis: different θ s in all schools, y j θ N(θ j,σ 2 j ) Data from each shool estimated separately MLE s would be the observations themselves, ˆθ j = y j Bayesian analysis would combine data y j with a prior for θ j In the hierarchical model the schools has different treatment effects but they are drawn from the same distribtion. Three levels: Conditional on θ: y 1,...,y 8 are independent and θ y j θ N(θ j,σj 2 ). In particular, the distribution of y j depends on θ only via θ j. Conditional on hyperprior (µ,τ): θ 1,...,θ 8 are independent and θ j N(µ,τ 2 ). Hyperpriors: (µ,τ) p(µ,τ) = p(τ)p(µ τ). Will assume that p(µ τ) 1 and later also that p(τ) 1. Drawing of model structure similar to that on page 119. NB. Likelihood analysis of the mixed model without the hyperprior level would yield a negative estimate of τ 2. Unpleasant... Bayes (Lecture 6) Hierarchical models 15 / 22 Bayes (Lecture 6) Hierarchical models 16 / 22
5 : analysis 0. Model sepcification: that s what we just did. 1. Posterior of random effects and hyperparameters: p(θ, µ,τ y) p(µ,τ) N(θ j µ,τ 2 ) 2. Posterior of random effects given hyperparameters: As a function of θ for fixed (µ,τ): [ p(θ µ,τ,y) = e 1 2τ 2 (θ j µ) 2 e 1 2σ 2 j N(y j θ j,σ 2 j ) (y j θ j ) 2] Look at each term separately and recognize the simple model with normal prior and normal prior. Hence p(θ j µ,τ,y) = N(ˆθ j,v 2 j ) where 3. Posterior of hyperparameter, p(µ,τ 2 y) p(y µ,τ)p(µ,τ). In this particular case we can find p(y µ,τ) explicitly: conditional on (µ,τ) the y j s are independent and Why is that so? Hence, p(y j µ,τ) = N(µ,σ 2 j + τ 2 ) p(µ,τ y) = p(µ,τ) N(µ,σ 2 j + τ 2 ) ˆθ j = y j/σj 2 + µ/τ 2 1/σj 2 + 1/τ ; 1 = 1 2 V j j σ τ 2 Bayes (Lecture 6) Hierarchical models 17 / 22 Bayes (Lecture 6) Hierarchical models 18 / 22 3.a Posterior of µ given τ, p(µ τ,y): Assume that p(µ τ) 1 so that p(µ,τ) = p(τ). As a function if µ for fixed τ: p(µ τ,y) ( ) 1 exp 2σj 2 + τ (y 2 j µ) 2 Recognize for each term the posterior for the normal model with known variance σ 2 j + τ 2 and uniform prior. Hence p(µ τ,y) = N(ˆµ,V µ ) where ˆµ = 1 y σj 2 j +τ2 1 ; σj 2+τ2 1 = V µ 1 σj 2 + τ 2 3.b Posterior of τ: p(µ,τ y) = p(µ τ,y)p(τ y) so p(τ y) = p(µ,τ y) p(µ τ,y) = p(ˆµ,τ y) p(ˆµ τ,y) = p(τ) N(µ,σj 2 + τ 2 ) 1/ V µ Note that ˆµ and V µ are functions of τ. 4. Hyperprior distribution. Already specified p(µ τ) 1. Moreover, let p(τ) 1. NB. p(logτ) 1 leads to improper posterior try it out yourself! A bit confusion in the book regarding standard deviation (τ) and variance (τ 2 ). Bayes (Lecture 6) Hierarchical models 19 / 22 Bayes (Lecture 6) Hierarchical models 20 / 22
6 Computations and conclusions 5. Contours of p(µ,τ y). 6. Posterior simulation of (µ, τ). Two steps: Simulate τ from p(τ y) by a grid approximation Simulate µ from p(µ τ,y) for the simluted τ-values 7. Posterior simulation of θ: Simulate each θ j separately from p(θ j µ,τ,y) for the simulated values of (µ, τ). You will do the actual computations in an exercise next week, so just a few comments at this point. Figure 5.6 in the book: which analysis does τ = 0 correspond to? Which analysis does a large τ correspond to? Table 5.2 and 5.3: compare y j to posterior means of θ j. What has happened? What would you do to see if there is a treatment effect? Bayes (Lecture 6) Hierarchical models 21 / 22 Bayes (Lecture 6) Hierarchical models 22 / 22
Bayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19
Bayesian Methods David Rosenberg New York University April 11, 2017 David Rosenberg (New York University) DS-GA 1003 April 11, 2017 1 / 19 Classical Statistics Classical Statistics David Rosenberg (New
More informationMCMC Methods for data modeling
MCMC Methods for data modeling Kenneth Scerri Department of Automatic Control and Systems Engineering Introduction 1. Symposium on Data Modelling 2. Outline: a. Definition and uses of MCMC b. MCMC algorithms
More informationMarkov Chain Monte Carlo (part 1)
Markov Chain Monte Carlo (part 1) Edps 590BAY Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Spring 2018 Depending on the book that you select for
More informationEstimating Labels from Label Proportions
Estimating Labels from Label Proportions Novi Quadrianto Novi.Quad@gmail.com The Australian National University, Australia NICTA, Statistical Machine Learning Program, Australia Joint work with Alex Smola,
More informationBayesian model selection and diagnostics
Bayesian model selection and diagnostics A typical Bayesian analysis compares a handful of models. Example 1: Consider the spline model for the motorcycle data, how many basis functions? Example 2: Consider
More informationGeneral Factorial Models
In Chapter 8 in Oehlert STAT:5201 Week 9 - Lecture 1 1 / 31 It is possible to have many factors in a factorial experiment. We saw some three-way factorials earlier in the DDD book (HW 1 with 3 factors:
More informationCanopy Light: Synthesizing multiple data sources
Canopy Light: Synthesizing multiple data sources Tree growth depends upon light (previous example, lab 7) Hard to measure how much light an ADULT tree receives Multiple sources of proxy data Exposed Canopy
More information1 Methods for Posterior Simulation
1 Methods for Posterior Simulation Let p(θ y) be the posterior. simulation. Koop presents four methods for (posterior) 1. Monte Carlo integration: draw from p(θ y). 2. Gibbs sampler: sequentially drawing
More informationISyE8843A, Brani Vidakovic Handout 14
ISyE8843A, Brani Vidakovic Handout 4 BUGS BUGS is freely available software for constructing Bayesian statistical models and evaluating them using MCMC methodology. BUGS and WINBUGS are distributed freely
More informationBART STAT8810, Fall 2017
BART STAT8810, Fall 2017 M.T. Pratola November 1, 2017 Today BART: Bayesian Additive Regression Trees BART: Bayesian Additive Regression Trees Additive model generalizes the single-tree regression model:
More informationA. Incorrect! This would be the negative of the range. B. Correct! The range is the maximum data value minus the minimum data value.
AP Statistics - Problem Drill 05: Measures of Variation No. 1 of 10 1. The range is calculated as. (A) The minimum data value minus the maximum data value. (B) The maximum data value minus the minimum
More informationINFO Wednesday, September 14, 2016
INFO 1301 Wednesday, September 14, 2016 Topics to be covered Last class we spoke about the middle (central tendency) of a data set. Today we talk about variability in a data set. Variability of Data Variance
More informationGeneral Factorial Models
In Chapter 8 in Oehlert STAT:5201 Week 9 - Lecture 2 1 / 34 It is possible to have many factors in a factorial experiment. In DDD we saw an example of a 3-factor study with ball size, height, and surface
More informationMonte Carlo for Spatial Models
Monte Carlo for Spatial Models Murali Haran Department of Statistics Penn State University Penn State Computational Science Lectures April 2007 Spatial Models Lots of scientific questions involve analyzing
More informationAn Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework
IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXX 23 An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework Ji Won Yoon arxiv:37.99v [cs.lg] 3 Jul 23 Abstract In order to cluster
More informationCalibration and emulation of TIE-GCM
Calibration and emulation of TIE-GCM Serge Guillas School of Mathematics Georgia Institute of Technology Jonathan Rougier University of Bristol Big Thanks to Crystal Linkletter (SFU-SAMSI summer school)
More informationFMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu
FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)
More informationLinear Modeling with Bayesian Statistics
Linear Modeling with Bayesian Statistics Bayesian Approach I I I I I Estimate probability of a parameter State degree of believe in specific parameter values Evaluate probability of hypothesis given the
More informationA GENERAL GIBBS SAMPLING ALGORITHM FOR ANALYZING LINEAR MODELS USING THE SAS SYSTEM
A GENERAL GIBBS SAMPLING ALGORITHM FOR ANALYZING LINEAR MODELS USING THE SAS SYSTEM Jayawant Mandrekar, Daniel J. Sargent, Paul J. Novotny, Jeff A. Sloan Mayo Clinic, Rochester, MN 55905 ABSTRACT A general
More informationScalable Bayes Clustering for Outlier Detection Under Informative Sampling
Scalable Bayes Clustering for Outlier Detection Under Informative Sampling Based on JMLR paper of T. D. Savitsky Terrance D. Savitsky Office of Survey Methods Research FCSM - 2018 March 7-9, 2018 1 / 21
More informationClustering Lecture 5: Mixture Model
Clustering Lecture 5: Mixture Model Jing Gao SUNY Buffalo 1 Outline Basics Motivation, definition, evaluation Methods Partitional Hierarchical Density-based Mixture model Spectral methods Advanced topics
More informationMini-project 2 CMPSCI 689 Spring 2015 Due: Tuesday, April 07, in class
Mini-project 2 CMPSCI 689 Spring 2015 Due: Tuesday, April 07, in class Guidelines Submission. Submit a hardcopy of the report containing all the figures and printouts of code in class. For readability
More informationLaboratorio di Problemi Inversi Esercitazione 4: metodi Bayesiani e importance sampling
Laboratorio di Problemi Inversi Esercitazione 4: metodi Bayesiani e importance sampling Luca Calatroni Dipartimento di Matematica, Universitá degli studi di Genova May 19, 2016. Luca Calatroni (DIMA, Unige)
More informationOne-Shot Learning with a Hierarchical Nonparametric Bayesian Model
One-Shot Learning with a Hierarchical Nonparametric Bayesian Model R. Salakhutdinov, J. Tenenbaum and A. Torralba MIT Technical Report, 2010 Presented by Esther Salazar Duke University June 10, 2011 E.
More informationOnline Supplement to Bayesian Simultaneous Edit and Imputation for Multivariate Categorical Data
Online Supplement to Bayesian Simultaneous Edit and Imputation for Multivariate Categorical Data Daniel Manrique-Vallier and Jerome P. Reiter August 3, 2016 This supplement includes the algorithm for processing
More informationBayesian Model Averaging over Directed Acyclic Graphs With Implications for Prediction in Structural Equation Modeling
ing over Directed Acyclic Graphs With Implications for Prediction in ing David Kaplan Department of Educational Psychology Case April 13th, 2015 University of Nebraska-Lincoln 1 / 41 ing Case This work
More informationCSE 446 Bias-Variance & Naïve Bayes
CSE 446 Bias-Variance & Naïve Bayes Administrative Homework 1 due next week on Friday Good to finish early Homework 2 is out on Monday Check the course calendar Start early (midterm is right before Homework
More informationThe Multi Stage Gibbs Sampling: Data Augmentation Dutch Example
The Multi Stage Gibbs Sampling: Data Augmentation Dutch Example Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601 Module 8 1 Example: Data augmentation / Auxiliary variables A commonly-used
More informationWeek 4: Simple Linear Regression II
Week 4: Simple Linear Regression II Marcelo Coca Perraillon University of Colorado Anschutz Medical Campus Health Services Research Methods I HSMP 7607 2017 c 2017 PERRAILLON ARR 1 Outline Algebraic properties
More informationImage analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis
7 Computer Vision and Classification 413 / 458 Computer Vision and Classification The k-nearest-neighbor method The k-nearest-neighbor (knn) procedure has been used in data analysis and machine learning
More informationStatistics I 2011/2012 Notes about the third Computer Class: Simulation of samples and goodness of fit; Central Limit Theorem; Confidence intervals.
Statistics I 2011/2012 Notes about the third Computer Class: Simulation of samples and goodness of fit; Central Limit Theorem; Confidence intervals. In this Computer Class we are going to use Statgraphics
More informationINLA: Integrated Nested Laplace Approximations
INLA: Integrated Nested Laplace Approximations John Paige Statistics Department University of Washington October 10, 2017 1 The problem Markov Chain Monte Carlo (MCMC) takes too long in many settings.
More informationmritc: A Package for MRI Tissue Classification
mritc: A Package for MRI Tissue Classification Dai Feng 1 Luke Tierney 2 1 Merck Research Labratories 2 University of Iowa July 2010 Feng & Tierney (Merck & U of Iowa) MRI Tissue Classification July 2010
More informationOptimal designs for comparing curves
Optimal designs for comparing curves Holger Dette, Ruhr-Universität Bochum Maria Konstantinou, Ruhr-Universität Bochum Kirsten Schorning, Ruhr-Universität Bochum FP7 HEALTH 2013-602552 Outline 1 Motivation
More informationCSSS 510: Lab 2. Introduction to Maximum Likelihood Estimation
CSSS 510: Lab 2 Introduction to Maximum Likelihood Estimation 2018-10-12 0. Agenda 1. Housekeeping: simcf, tile 2. Questions about Homework 1 or lecture 3. Simulating heteroskedastic normal data 4. Fitting
More informationMissing Data Analysis for the Employee Dataset
Missing Data Analysis for the Employee Dataset 67% of the observations have missing values! Modeling Setup Random Variables: Y i =(Y i1,...,y ip ) 0 =(Y i,obs, Y i,miss ) 0 R i =(R i1,...,r ip ) 0 ( 1
More informationClustering. Mihaela van der Schaar. January 27, Department of Engineering Science University of Oxford
Department of Engineering Science University of Oxford January 27, 2017 Many datasets consist of multiple heterogeneous subsets. Cluster analysis: Given an unlabelled data, want algorithms that automatically
More informationGAMs semi-parametric GLMs. Simon Wood Mathematical Sciences, University of Bath, U.K.
GAMs semi-parametric GLMs Simon Wood Mathematical Sciences, University of Bath, U.K. Generalized linear models, GLM 1. A GLM models a univariate response, y i as g{e(y i )} = X i β where y i Exponential
More informationMachine Learning A WS15/16 1sst KU Version: January 11, b) [1 P] For the probability distribution P (A, B, C, D) with the factorization
Machine Learning A 708.064 WS15/16 1sst KU Version: January 11, 2016 Exercises Problems marked with * are optional. 1 Conditional Independence I [3 P] a) [1 P] For the probability distribution P (A, B,
More informationMissing Data and Imputation
Missing Data and Imputation NINA ORWITZ OCTOBER 30 TH, 2017 Outline Types of missing data Simple methods for dealing with missing data Single and multiple imputation R example Missing data is a complex
More information(Not That) Advanced Hierarchical Models
(Not That) Advanced Hierarchical Models Ben Goodrich StanCon: January 10, 2018 Ben Goodrich Advanced Hierarchical Models StanCon 1 / 13 Obligatory Disclosure Ben is an employee of Columbia University,
More informationNested Sampling: Introduction and Implementation
UNIVERSITY OF TEXAS AT SAN ANTONIO Nested Sampling: Introduction and Implementation Liang Jing May 2009 1 1 ABSTRACT Nested Sampling is a new technique to calculate the evidence, Z = P(D M) = p(d θ, M)p(θ
More informationClassification. 1 o Semestre 2007/2008
Classification Departamento de Engenharia Informática Instituto Superior Técnico 1 o Semestre 2007/2008 Slides baseados nos slides oficiais do livro Mining the Web c Soumen Chakrabarti. Outline 1 2 3 Single-Class
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 13: The bootstrap (v3) Ramesh Johari ramesh.johari@stanford.edu 1 / 30 Resampling 2 / 30 Sampling distribution of a statistic For this lecture: There is a population model
More informationChapters 5-6: Statistical Inference Methods
Chapters 5-6: Statistical Inference Methods Chapter 5: Estimation (of population parameters) Ex. Based on GSS data, we re 95% confident that the population mean of the variable LONELY (no. of days in past
More informationUniversity of Cambridge Engineering Part IIB Paper 4F10: Statistical Pattern Processing Handout 11: Non-Parametric Techniques.
. Non-Parameteric Techniques University of Cambridge Engineering Part IIB Paper 4F: Statistical Pattern Processing Handout : Non-Parametric Techniques Mark Gales mjfg@eng.cam.ac.uk Michaelmas 23 Introduction
More informationA. Using the data provided above, calculate the sampling variance and standard error for S for each week s data.
WILD 502 Lab 1 Estimating Survival when Animal Fates are Known Today s lab will give you hands-on experience with estimating survival rates using logistic regression to estimate the parameters in a variety
More informationChapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea
Chapter 3 Bootstrap 3.1 Introduction The estimation of parameters in probability distributions is a basic problem in statistics that one tends to encounter already during the very first course on the subject.
More informationA Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data
A Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data Seyoung Kim, Padhraic Smyth, and Hal Stern Bren School of Information and Computer Sciences University of California,
More informationUniversity of Cambridge Engineering Part IIB Paper 4F10: Statistical Pattern Processing Handout 11: Non-Parametric Techniques
University of Cambridge Engineering Part IIB Paper 4F10: Statistical Pattern Processing Handout 11: Non-Parametric Techniques Mark Gales mjfg@eng.cam.ac.uk Michaelmas 2015 11. Non-Parameteric Techniques
More informationExtreme Value Theory in (Hourly) Precipitation
Extreme Value Theory in (Hourly) Precipitation Uli Schneider Geophysical Statistics Project, NCAR GSP Miniseries at CSU November 17, 2003 Outline Project overview Extreme value theory 101 Applying extreme
More informationNew Results of Fully Bayesian
of Fully Bayesian UCI February 7, 2012 of Fully Bayesian Calibration Samples Principle Component Analysis Model Building Three source parameter sampling schemes Simulation Quasar data sets New data sets
More informationOverview. Monte Carlo Methods. Statistics & Bayesian Inference Lecture 3. Situation At End Of Last Week
Statistics & Bayesian Inference Lecture 3 Joe Zuntz Overview Overview & Motivation Metropolis Hastings Monte Carlo Methods Importance sampling Direct sampling Gibbs sampling Monte-Carlo Markov Chains Emcee
More informationEstimation of Item Response Models
Estimation of Item Response Models Lecture #5 ICPSR Item Response Theory Workshop Lecture #5: 1of 39 The Big Picture of Estimation ESTIMATOR = Maximum Likelihood; Mplus Any questions? answers Lecture #5:
More informationMissing Data Analysis for the Employee Dataset
Missing Data Analysis for the Employee Dataset 67% of the observations have missing values! Modeling Setup For our analysis goals we would like to do: Y X N (X, 2 I) and then interpret the coefficients
More informationEdge-exchangeable graphs and sparsity
Edge-exchangeable graphs and sparsity Tamara Broderick Department of EECS Massachusetts Institute of Technology tbroderick@csail.mit.edu Diana Cai Department of Statistics University of Chicago dcai@uchicago.edu
More informationBayesian Computation: Overview and Methods for Low-Dimensional Models
Bayesian Computation: Overview and Methods for Low-Dimensional Models Tom Loredo Dept. of Astronomy, Cornell University http://www.astro.cornell.edu/staff/loredo/bayes/ IAC Winter School, 3 4 Nov 2014
More informationUse of Extreme Value Statistics in Modeling Biometric Systems
Use of Extreme Value Statistics in Modeling Biometric Systems Similarity Scores Two types of matching: Genuine sample Imposter sample Matching scores Enrolled sample 0.95 0.32 Probability Density Decision
More informationBayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis
Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis Xavier Le Faucheur a, Brani Vidakovic b and Allen Tannenbaum a a School of Electrical and Computer Engineering, b Department of Biomedical
More informationToday. Lecture 4: Last time. The EM algorithm. We examine clustering in a little more detail; we went over it a somewhat quickly last time
Today Lecture 4: We examine clustering in a little more detail; we went over it a somewhat quickly last time The CAD data will return and give us an opportunity to work with curves (!) We then examine
More informationComputational Cognitive Science
Computational Cognitive Science Lecture 5: Maximum Likelihood Estimation; Parameter Uncertainty Chris Lucas (Slides adapted from Frank Keller s) School of Informatics University of Edinburgh clucas2@inf.ed.ac.uk
More informationCS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas
CS839: Probabilistic Graphical Models Lecture 10: Learning with Partially Observed Data Theo Rekatsinas 1 Partially Observed GMs Speech recognition 2 Partially Observed GMs Evolution 3 Partially Observed
More informationSupplementary Figure 1. Decoding results broken down for different ROIs
Supplementary Figure 1 Decoding results broken down for different ROIs Decoding results for areas V1, V2, V3, and V1 V3 combined. (a) Decoded and presented orientations are strongly correlated in areas
More informationSTA 570 Spring Lecture 5 Tuesday, Feb 1
STA 570 Spring 2011 Lecture 5 Tuesday, Feb 1 Descriptive Statistics Summarizing Univariate Data o Standard Deviation, Empirical Rule, IQR o Boxplots Summarizing Bivariate Data o Contingency Tables o Row
More informationRegularization and model selection
CS229 Lecture notes Andrew Ng Part VI Regularization and model selection Suppose we are trying select among several different models for a learning problem. For instance, we might be using a polynomial
More informationR package mcll for Monte Carlo local likelihood estimation
R package mcll for Monte Carlo local likelihood estimation Minjeong Jeon University of California, Berkeley Sophia Rabe-Hesketh University of California, Berkeley February 4, 2013 Cari Kaufman University
More informationInf2B assignment 2. Natural images classification. Hiroshi Shimodaira and Pol Moreno. Submission due: 4pm, Wednesday 30 March 2016.
Inf2B assignment 2 (Ver. 1.2) Natural images classification Submission due: 4pm, Wednesday 30 March 2016 Hiroshi Shimodaira and Pol Moreno This assignment is out of 100 marks and forms 12.5% of your final
More informationHomework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures:
Homework Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression 3.0-3.2 Pod-cast lecture on-line Next lectures: I posted a rough plan. It is flexible though so please come with suggestions Bayes
More information6.867 Machine Learning
6.867 Machine Learning Problem set 3 Due Tuesday, October 22, in class What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove.
More informationMachine Learning A W 1sst KU. b) [1 P] For the probability distribution P (A, B, C, D) with the factorization
Machine Learning A 708.064 13W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence a) [1 P] For the probability distribution P (A, B, C, D) with the factorization P (A, B,
More informationBeviMed Guide. Daniel Greene
BeviMed Guide Daniel Greene 1 Introduction BeviMed [1] is a procedure for evaluating the evidence of association between allele configurations across rare variants, typically within a genomic locus, and
More informationMachine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves
Machine Learning A 708.064 11W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence I [2 P] a) [1 P] Give an example for a probability distribution P (A, B, C) that disproves
More informationGlobal modelling of air pollution using multiple data sources
Global modelling of air pollution using multiple data sources Matthew Thomas M.L.Thomas@bath.ac.uk Supervised by Dr. Gavin Shaddick In collaboration with IHME and WHO June 14, 2016 1/ 1 MOTIVATION Air
More informationProgramming Exercise 1: Linear Regression
Programming Exercise 1: Linear Regression Machine Learning Introduction In this exercise, you will implement linear regression and get to see it work on data. Before starting on this programming exercise,
More informationECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov
ECE521: Week 11, Lecture 20 27 March 2017: HMM learning/inference With thanks to Russ Salakhutdinov Examples of other perspectives Murphy 17.4 End of Russell & Norvig 15.2 (Artificial Intelligence: A Modern
More informationGraphical Models. Dmitrij Lagutin, T Machine Learning: Basic Principles
Graphical Models Dmitrij Lagutin, dlagutin@cc.hut.fi T-61.6020 - Machine Learning: Basic Principles 12.3.2007 Contents Introduction to graphical models Bayesian networks Conditional independence Markov
More informationDeep Generative Models Variational Autoencoders
Deep Generative Models Variational Autoencoders Sudeshna Sarkar 5 April 2017 Generative Nets Generative models that represent probability distributions over multiple variables in some way. Directed Generative
More informationAn imputation approach for analyzing mixed-mode surveys
An imputation approach for analyzing mixed-mode surveys Jae-kwang Kim 1 Iowa State University June 4, 2013 1 Joint work with S. Park and S. Kim Ouline Introduction Proposed Methodology Application to Private
More informationLecture 12. August 23, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.
Lecture 12 Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University August 23, 2007 1 2 3 4 5 1 2 Introduce the bootstrap 3 the bootstrap algorithm 4 Example
More informationVariability in Annual Temperature Profiles
Variability in Annual Temperature Profiles A Multivariate Spatial Analysis of Regional Climate Model Output Tamara Greasby, Stephan Sain Institute for Mathematics Applied to Geosciences, National Center
More informationSemiparametric Mixed Effecs with Hierarchical DP Mixture
Semiparametric Mixed Effecs with Hierarchical DP Mixture R topics documented: April 21, 2007 hdpm-package........................................ 1 hdpm............................................ 2 hdpmfitsetup........................................
More informationAdvanced Machine Learning Practical 3: Classification (SVM, RVM & AdaBoost)
Advanced Machine Learning Practical : Classification (SVM, RVM & AdaBoost) Professor: Aude Billard Assistants: Nadia Figueroa, Ilaria Lauzana and Brice Platerrier E-mails: aude.billard@epfl.ch, nadia.figueroafernandez@epfl.ch
More informationGenerative and discriminative classification techniques
Generative and discriminative classification techniques Machine Learning and Category Representation 2014-2015 Jakob Verbeek, November 28, 2014 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.14.15
More information10. MLSP intro. (Clustering: K-means, EM, GMM, etc.)
10. MLSP intro. (Clustering: K-means, EM, GMM, etc.) Rahil Mahdian 01.04.2016 LSV Lab, Saarland University, Germany What is clustering? Clustering is the classification of objects into different groups,
More information1. The Normal Distribution, continued
Math 1125-Introductory Statistics Lecture 16 10/9/06 1. The Normal Distribution, continued Recall that the standard normal distribution is symmetric about z = 0, so the area to the right of zero is 0.5000.
More informationLab 7: Bayesian analysis of a dice toss problem using C++ instead of python
Lab 7: Bayesian analysis of a dice toss problem using C++ instead of python Due date: Monday March 27, 11:59pm Short version of the assignment Take your python file from lab 6 and convert it into lab7
More information10-701/15-781, Fall 2006, Final
-7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly
More informationSTAT 725 Notes Monte Carlo Integration
STAT 725 Notes Monte Carlo Integration Two major classes of numerical problems arise in statistical inference: optimization and integration. We have already spent some time discussing different optimization
More informationComputer vision: models, learning and inference. Chapter 10 Graphical Models
Computer vision: models, learning and inference Chapter 10 Graphical Models Independence Two variables x 1 and x 2 are independent if their joint probability distribution factorizes as Pr(x 1, x 2 )=Pr(x
More informationA quick introduction to First Bayes
A quick introduction to First Bayes Lawrence Joseph October 1, 2003 1 Introduction This document very briefly reviews the main features of the First Bayes statistical teaching package. For full details,
More informationBayesian Inference for Sample Surveys
Bayesian Inference for Sample Surveys Trivellore Raghunathan (Raghu) Director, Survey Research Center Professor of Biostatistics University of Michigan Distinctive features of survey inference 1. Primary
More informationINLA: an introduction
INLA: an introduction Håvard Rue 1 Norwegian University of Science and Technology Trondheim, Norway May 2009 1 Joint work with S.Martino (Trondheim) and N.Chopin (Paris) Latent Gaussian models Background
More informationWarped Mixture Models
Warped Mixture Models Tomoharu Iwata, David Duvenaud, Zoubin Ghahramani Cambridge University Computational and Biological Learning Lab March 11, 2013 OUTLINE Motivation Gaussian Process Latent Variable
More informationLab #9: ANOVA and TUKEY tests
Lab #9: ANOVA and TUKEY tests Objectives: 1. Column manipulation in SAS 2. Analysis of variance 3. Tukey test 4. Least Significant Difference test 5. Analysis of variance with PROC GLM 6. Levene test for
More informationIssues in MCMC use for Bayesian model fitting. Practical Considerations for WinBUGS Users
Practical Considerations for WinBUGS Users Kate Cowles, Ph.D. Department of Statistics and Actuarial Science University of Iowa 22S:138 Lecture 12 Oct. 3, 2003 Issues in MCMC use for Bayesian model fitting
More informationNetwork Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Sources Hastie, Tibshirani, Friedman: The Elements of Statistical Learning James, Witten, Hastie, Tibshirani: An Introduction to Statistical Learning Andrew Ng:
More informationPackage EBglmnet. January 30, 2016
Type Package Package EBglmnet January 30, 2016 Title Empirical Bayesian Lasso and Elastic Net Methods for Generalized Linear Models Version 4.1 Date 2016-01-15 Author Anhui Huang, Dianting Liu Maintainer
More informationProblem 1 (20 pt) Answer the following questions, and provide an explanation for each question.
Problem 1 Answer the following questions, and provide an explanation for each question. (5 pt) Can linear regression work when all X values are the same? When all Y values are the same? (5 pt) Can linear
More informationGraphical Models & HMMs
Graphical Models & HMMs Henrik I. Christensen Robotics & Intelligent Machines @ GT Georgia Institute of Technology, Atlanta, GA 30332-0280 hic@cc.gatech.edu Henrik I. Christensen (RIM@GT) Graphical Models
More information08 An Introduction to Dense Continuous Robotic Mapping
NAVARCH/EECS 568, ROB 530 - Winter 2018 08 An Introduction to Dense Continuous Robotic Mapping Maani Ghaffari March 14, 2018 Previously: Occupancy Grid Maps Pose SLAM graph and its associated dense occupancy
More information