INLA: an introduction
|
|
- Bathsheba Burns
- 5 years ago
- Views:
Transcription
1 INLA: an introduction Håvard Rue 1 Norwegian University of Science and Technology Trondheim, Norway May Joint work with S.Martino (Trondheim) and N.Chopin (Paris)
2 Latent Gaussian models Background Latent Gaussian models Stage 1 Observed data y = (y i ), y i x, θ π(y i x i, θ) Stage 2 Latent Gaussian field x θ N (µ, Q(θ) 1 ), Ax = 0 Stage 3 Priors for the hyperparameters θ π(θ) Unify many of the most used models in statistics
3 Latent Gaussian models Background Latent Gaussian models Stage 1 Observed data y = (y i ), y i x, θ π(y i x i, θ) Stage 2 Latent Gaussian field x θ N (µ, Q(θ) 1 ), Ax = 0 Stage 3 Priors for the hyperparameters θ π(θ) Unify many of the most used models in statistics
4 Latent Gaussian models Background Latent Gaussian models Stage 1 Observed data y = (y i ), y i x, θ π(y i x i, θ) Stage 2 Latent Gaussian field x θ N (µ, Q(θ) 1 ), Ax = 0 Stage 3 Priors for the hyperparameters θ π(θ) Unify many of the most used models in statistics
5 Latent Gaussian models Background Latent Gaussian models Stage 1 Observed data y = (y i ), y i x, θ π(y i x i, θ) Stage 2 Latent Gaussian field x θ N (µ, Q(θ) 1 ), Ax = 0 Stage 3 Priors for the hyperparameters θ π(θ) Unify many of the most used models in statistics
6 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ
7 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ
8 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ
9 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ
10 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ
11 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ
12 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ
13 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ
14 Latent Gaussian models Background Structured additive regression models η i = k β k z ki + j w ji f j (z ji ) + ɛ i Observations y π(y x, θ) = i π(y i η i, θ) from an (f.ex) exponential family with mean µ i = g 1 (η i ). Latent Gaussian model if x = ({β k }, {f ji }, {η i }) θ N (µ, Q(θ) 1 )
15 Latent Gaussian models Background Structured additive regression models η i = k β k z ki + j w ji f j (z ji ) + ɛ i Observations y π(y x, θ) = i π(y i η i, θ) from an (f.ex) exponential family with mean µ i = g 1 (η i ). Latent Gaussian model if x = ({β k }, {f ji }, {η i }) θ N (µ, Q(θ) 1 )
16 Latent Gaussian models Background Structured additive regression models η i = k β k z ki + j w ji f j (z ji ) + ɛ i Observations y π(y x, θ) = i π(y i η i, θ) from an (f.ex) exponential family with mean µ i = g 1 (η i ). Latent Gaussian model if x = ({β k }, {f ji }, {η i }) θ N (µ, Q(θ) 1 )
17 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
18 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
19 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
20 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
21 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
22 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
23 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
24 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
25 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
26 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
27 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models
28 Latent Gaussian models Examples Example: Disease mapping (BYM-model) Data y i Poisson(E i exp(η i )) Log-relative risk η i = u i + v i + β T z i Structured component u Unstructured component v Covariates z i Log-precisions log κ u and log κ v
29 Latent Gaussian models Examples Characteristic features Large dimension of the latent Gaussian field: A lot of conditional independence in the latent Gaussian field Few hyperparameters θ: dim(θ) between 1 and 5 Non-Gaussian data
30 Latent Gaussian models Examples Characteristic features Large dimension of the latent Gaussian field: A lot of conditional independence in the latent Gaussian field Few hyperparameters θ: dim(θ) between 1 and 5 Non-Gaussian data
31 Latent Gaussian models Examples Characteristic features Large dimension of the latent Gaussian field: A lot of conditional independence in the latent Gaussian field Few hyperparameters θ: dim(θ) between 1 and 5 Non-Gaussian data
32 Latent Gaussian models Examples Characteristic features Large dimension of the latent Gaussian field: A lot of conditional independence in the latent Gaussian field Few hyperparameters θ: dim(θ) between 1 and 5 Non-Gaussian data
33 Task Main task Compute the posterior marginals for the latent field π(x i y), i = 1,..., n Compute the posterior marginals for the hyperparameters π(θ j y), j = 1,..., dim(θ) Today s standard approach, is to make use of MCMC Main difficulties CPU-time Additive MC-errors
34 Task Main task Compute the posterior marginals for the latent field π(x i y), i = 1,..., n Compute the posterior marginals for the hyperparameters π(θ j y), j = 1,..., dim(θ) Today s standard approach, is to make use of MCMC Main difficulties CPU-time Additive MC-errors
35 Task Main task Compute the posterior marginals for the latent field π(x i y), i = 1,..., n Compute the posterior marginals for the hyperparameters π(θ j y), j = 1,..., dim(θ) Today s standard approach, is to make use of MCMC Main difficulties CPU-time Additive MC-errors
36 Task Main task Compute the posterior marginals for the latent field π(x i y), i = 1,..., n Compute the posterior marginals for the hyperparameters π(θ j y), j = 1,..., dim(θ) Today s standard approach, is to make use of MCMC Main difficulties CPU-time Additive MC-errors
37 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations
38 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations
39 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations
40 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations
41 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations
42 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations
43 Our approach Summary of results Main results HUGE improvement in both speed and accuracy compared to MCMC alternatives Relative error Practically exact results 2 Extensions: Marginal likelihood, DIC, Cross-validation,... INLA enable us to treat Bayesian latent Gaussian models properly and bring these models from the research communities to the end-users 2 Can construct counter-examples
44 Our approach Summary of results Main results HUGE improvement in both speed and accuracy compared to MCMC alternatives Relative error Practically exact results 2 Extensions: Marginal likelihood, DIC, Cross-validation,... INLA enable us to treat Bayesian latent Gaussian models properly and bring these models from the research communities to the end-users 2 Can construct counter-examples
45 Our approach Summary of results Main results HUGE improvement in both speed and accuracy compared to MCMC alternatives Relative error Practically exact results 2 Extensions: Marginal likelihood, DIC, Cross-validation,... INLA enable us to treat Bayesian latent Gaussian models properly and bring these models from the research communities to the end-users 2 Can construct counter-examples
46 Our approach Summary of results Main results HUGE improvement in both speed and accuracy compared to MCMC alternatives Relative error Practically exact results 2 Extensions: Marginal likelihood, DIC, Cross-validation,... INLA enable us to treat Bayesian latent Gaussian models properly and bring these models from the research communities to the end-users 2 Can construct counter-examples
47 Our approach Summary of results Main results HUGE improvement in both speed and accuracy compared to MCMC alternatives Relative error Practically exact results 2 Extensions: Marginal likelihood, DIC, Cross-validation,... INLA enable us to treat Bayesian latent Gaussian models properly and bring these models from the research communities to the end-users 2 Can construct counter-examples
48 Main ideas Main ideas (I) π(z) = π(x, z) π(x z) leading to π(z) = π(x, z) π(x z) mode(z) When π(x z) is the Gaussian-approximation, this is the Laplace-approximation Want π(x z) to be almost Gaussian.
49 Main ideas Main ideas (I) π(z) = π(x, z) π(x z) leading to π(z) = π(x, z) π(x z) mode(z) When π(x z) is the Gaussian-approximation, this is the Laplace-approximation Want π(x z) to be almost Gaussian.
50 Main ideas Main ideas (I) π(z) = π(x, z) π(x z) leading to π(z) = π(x, z) π(x z) mode(z) When π(x z) is the Gaussian-approximation, this is the Laplace-approximation Want π(x z) to be almost Gaussian.
51 Main ideas Main ideas (II) Posterior π(x, θ y) π(θ) π(x θ) i I π(y i x i, θ) Do the integration wrt θ numerically
52 Main ideas Main ideas (II) Posterior π(x, θ y) π(θ) π(x θ) i I π(y i x i, θ) Do the integration wrt θ numerically π(x i y) = π(θ j y) = π(θ y) π(x i θ, y) dθ π(θ y) dθ j
53 Main ideas Main ideas (II) Posterior π(x, θ y) π(θ) π(x θ) i I π(y i x i, θ) Do the integration wrt θ numerically π(x i y) = π(θ j y) = π(θ y) π(x i θ, y) dθ π(θ y) dθ j
54 Main ideas Remarks 1. Expect π(θ y) to be accurate, since x θ is a priori Gaussian Likelihood models are well-behaved so is almost Gaussian. π(x θ, y) 2. There are no distributional assumptions on θ y 3. Similar remarks are valid to π(x i θ, y)
55 Main ideas Remarks 1. Expect π(θ y) to be accurate, since x θ is a priori Gaussian Likelihood models are well-behaved so is almost Gaussian. π(x θ, y) 2. There are no distributional assumptions on θ y 3. Similar remarks are valid to π(x i θ, y)
56 Main ideas Remarks 1. Expect π(θ y) to be accurate, since x θ is a priori Gaussian Likelihood models are well-behaved so is almost Gaussian. π(x θ, y) 2. There are no distributional assumptions on θ y 3. Similar remarks are valid to π(x i θ, y)
57 Main ideas Remarks 1. Expect π(θ y) to be accurate, since x θ is a priori Gaussian Likelihood models are well-behaved so is almost Gaussian. π(x θ, y) 2. There are no distributional assumptions on θ y 3. Similar remarks are valid to π(x i θ, y)
58 Main ideas Remarks 1. Expect π(θ y) to be accurate, since x θ is a priori Gaussian Likelihood models are well-behaved so is almost Gaussian. π(x θ, y) 2. There are no distributional assumptions on θ y 3. Similar remarks are valid to π(x i θ, y)
59 Computational tools Background Computational issues Our approach in its raw form is not computational feasible Main issue is the large dimension, n, of the latent field Various strategies/tricks are required for obtaining a practical good solution
60 Computational tools Background Computational issues Our approach in its raw form is not computational feasible Main issue is the large dimension, n, of the latent field Various strategies/tricks are required for obtaining a practical good solution
61 Computational tools Background Computational issues Our approach in its raw form is not computational feasible Main issue is the large dimension, n, of the latent field Various strategies/tricks are required for obtaining a practical good solution
62 Computational tools Gaussian Markov random fields Gaussian Markov random fields Make use of the conditional independence properties in the latent field x i x j x ij Q ij = 0 where Q is the precision matrix (inverse covariance) Use numerical methods for sparse matrices!
63 Computational tools Gaussian Markov random fields Gaussian Markov random fields Make use of the conditional independence properties in the latent field x i x j x ij Q ij = 0 where Q is the precision matrix (inverse covariance) Use numerical methods for sparse matrices!
64 Computational tools Rank one updates Rank one updates Have computed the Cholesky factorisation Prec(x) = Q = LL T Want to compute the conditional mean and variances for x i x i No need to factorise Prec(x i x i ); can compute the correction for conditioning on x i.
65 Computational tools Rank one updates Rank one updates Have computed the Cholesky factorisation Prec(x) = Q = LL T Want to compute the conditional mean and variances for x i x i No need to factorise Prec(x i x i ); can compute the correction for conditioning on x i.
66 Computational tools Rank one updates Rank one updates Have computed the Cholesky factorisation Prec(x) = Q = LL T Want to compute the conditional mean and variances for x i x i No need to factorise Prec(x i x i ); can compute the correction for conditioning on x i.
67 Computational tools Simplified Laplace Simplified Laplace Approximation Expand the Laplace approximation of π(x i θ, y): log π(x i θ, y) = 1 2 x 2 i + b i x i d i x 3 i + Remarks Correct the Gaussian approximation for error in shift and skewness through b i and d i Fit a skew-normal density Computational fast 2φ(x)Φ(ax) Sufficient accurate for most applications
68 Computational tools Simplified Laplace Simplified Laplace Approximation Expand the Laplace approximation of π(x i θ, y): log π(x i θ, y) = 1 2 x 2 i + b i x i d i x 3 i + Remarks Correct the Gaussian approximation for error in shift and skewness through b i and d i Fit a skew-normal density Computational fast 2φ(x)Φ(ax) Sufficient accurate for most applications
69 Computational tools Simplified Laplace Simplified Laplace Approximation Expand the Laplace approximation of π(x i θ, y): log π(x i θ, y) = 1 2 x 2 i + b i x i d i x 3 i + Remarks Correct the Gaussian approximation for error in shift and skewness through b i and d i Fit a skew-normal density Computational fast 2φ(x)Φ(ax) Sufficient accurate for most applications
70 Computational tools Simplified Laplace Simplified Laplace Approximation Expand the Laplace approximation of π(x i θ, y): log π(x i θ, y) = 1 2 x 2 i + b i x i d i x 3 i + Remarks Correct the Gaussian approximation for error in shift and skewness through b i and d i Fit a skew-normal density Computational fast 2φ(x)Φ(ax) Sufficient accurate for most applications
71 Computational tools Simplified Laplace Simplified Laplace Approximation Expand the Laplace approximation of π(x i θ, y): log π(x i θ, y) = 1 2 x 2 i + b i x i d i x 3 i + Remarks Correct the Gaussian approximation for error in shift and skewness through b i and d i Fit a skew-normal density Computational fast 2φ(x)Φ(ax) Sufficient accurate for most applications
72 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) I Step I Explore π(θ y) Locate the mode Use the Hessian to construct new variables Grid-search
73 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) I Step I Explore π(θ y) Locate the mode Use the Hessian to construct new variables Grid-search
74 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) I Step I Explore π(θ y) Locate the mode Use the Hessian to construct new variables Grid-search
75 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) II Step II For each θ j For each i, compute the (simplified) Laplace approximation for x i
76 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) III Step III Sum out θ j For each i, sum out θ π(x i y) j π(x i y, θ j ) π(θ j y) Build a log-spline corrected Gaussian N (x i ; µ i, σ 2 i ) exp(spline) to represent π(x i y).
77 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) III Step III Sum out θ j For each i, sum out θ π(x i y) j π(x i y, θ j ) π(θ j y) Build a log-spline corrected Gaussian N (x i ; µ i, σ 2 i ) exp(spline) to represent π(x i y).
78 The Integrated nested Laplace-approximation (INLA) Assessing the error How can we assess the errors in the approximations? Important, but asymptotic arguments are difficult: dim(y) = O(n) and dim(x) = O(n)
79 The Integrated nested Laplace-approximation (INLA) Assessing the error Errors in the approximations of π(x i y) Compare a sequence of improved approximations 1. Gaussian approximation 2. Simplified Laplace 3. Laplace Compute the full Laplace-approximation for π(x i y, θ j ) only if the Gaussian and the Simplified Laplace approximation disagree.
80 The Integrated nested Laplace-approximation (INLA) Assessing the error Errors in the approximations of π(x i y) Compare a sequence of improved approximations 1. Gaussian approximation 2. Simplified Laplace 3. Laplace Compute the full Laplace-approximation for π(x i y, θ j ) only if the Gaussian and the Simplified Laplace approximation disagree.
81 The Integrated nested Laplace-approximation (INLA) Assessing the error Overall check: Equivalent number of replicates Tool 3: Estimate the effective number of parameters From the Deviance Information Criteria: p D (θ) n trace ( Q prior (θ) Q post. (θ) 1) Compare with the number of observations: #observations/p D (θ) high ratio is good Theoretical justification
82 The Integrated nested Laplace-approximation (INLA) Assessing the error Overall check: Equivalent number of replicates Tool 3: Estimate the effective number of parameters From the Deviance Information Criteria: p D (θ) n trace ( Q prior (θ) Q post. (θ) 1) Compare with the number of observations: #observations/p D (θ) high ratio is good Theoretical justification
83 The Integrated nested Laplace-approximation (INLA) Assessing the error Overall check: Equivalent number of replicates Tool 3: Estimate the effective number of parameters From the Deviance Information Criteria: p D (θ) n trace ( Q prior (θ) Q post. (θ) 1) Compare with the number of observations: #observations/p D (θ) high ratio is good Theoretical justification
84 Examples Examples 10:30-11:30: Andrea Riebler: Performance of INLA analysing bivariate meta-regression and age-period-cohort models. 12:30-13:30: Birgit Schrödle: Spatio-temporal disease mapping using INLA. 13:45-14:45: Virgilio Gómez-Rubio: Approximate Bayesian Inference for Small Area Estimation 15:00-16:00: Lea Fortunato: About the Rapid Inquiry Facility, and spatial analyses with WinBUGS and INLA. 09:00-10:00: Rupali Akerkar: Approximate Bayesian Inference for Survival models. 10:15-11:30: Ingelin Steinsland & Anna Marie Holand: Animal Model and INLA.
85 Extensions Marginal likelihood Marginal likelihood Marginal likelihood is the normalising constant for π(θ y)
86 Extensions DIC Deviance Information Criteria D(x; θ) = 2 i log(y i x i, θ) DIC = 2 Mean (D(x; θ)) D(Mean(x); θ )
87 Cross-validation Cross-validation Based on we can compute π(x i y i, θ) π(x i y, θ) π(y i x i, θ) π(y i y i ) Similar with π(θ y i ) Keep the integration points {θ j } fixed. Detect surprising observations: Prob(y new i y i y i )
88 Cross-validation Cross-validation Based on we can compute π(x i y i, θ) π(x i y, θ) π(y i x i, θ) π(y i y i ) Similar with π(θ y i ) Keep the integration points {θ j } fixed. Detect surprising observations: Prob(y new i y i y i )
89 Cross-validation Cross-validation Based on we can compute π(x i y i, θ) π(x i y, θ) π(y i x i, θ) π(y i y i ) Similar with π(θ y i ) Keep the integration points {θ j } fixed. Detect surprising observations: Prob(y new i y i y i )
90 Cross-validation Cross-validation Based on we can compute π(x i y i, θ) π(x i y, θ) π(y i x i, θ) π(y i y i ) Similar with π(θ y i ) Keep the integration points {θ j } fixed. Detect surprising observations: Prob(y new i y i y i )
91 Recent developments Recent developments in (R-)INLA Improving the code: speedups and improved parallel performance Improving the R-interface Extending the building-blocks: prior-models and likelihood-models Lot of things still todo... Documentation Worked out examples Webpage +++
92 Recent developments Improving the code The two largest changes the last year (summer 2008) Change the way inla works in a multi-core environment. Especially important for more than dual-core (xmas 2008) Implement a more efficient rank-one update formula (Thanks Birgit!). Especially important for models with many constraints, but gave a huge speedup also for models with a single constraint.
93 Recent developments Improving the code The two largest changes the last year (summer 2008) Change the way inla works in a multi-core environment. Especially important for more than dual-core (xmas 2008) Implement a more efficient rank-one update formula (Thanks Birgit!). Especially important for models with many constraints, but gave a huge speedup also for models with a single constraint.
94 Recent developments inla & multi-core Gradient and Hessian computations are done in parallel π(θ y) θ i and 2 θ i θ j π(θ y) All computations of π(x i θ j, y) is now done in parallel wrt j, and not i as before.
95 Recent developments Multi-core example inla-example 7: CANCER-INCIDENCE, n = 7138, dim(θ) = 3 Number of cores Time used s s 4 9.1s 6 8.6s 8 8.3s Linear algebra (Cholesky/Solve/Inverse): 50% Administration: 50%
96 Recent developments Recent developments in (R-)INLA Improved R-interface inla-binary is now bundled with the R-package. Make use of model.matrix() in R: formula ~... + x*z will expand as formula ~... Allow for formula ~... + x + z + x:z + offset(a) + offset(b) defining a+b to be fixed offset in formula Larger change for survival-models (more on this tomorrow...)
97 Recent developments Recent developments in (R-)INLA Improved R-interface inla-binary is now bundled with the R-package. Make use of model.matrix() in R: formula ~... + x*z will expand as formula ~... Allow for formula ~... + x + z + x:z + offset(a) + offset(b) defining a+b to be fixed offset in formula Larger change for survival-models (more on this tomorrow...)
98 Recent developments Recent developments in (R-)INLA Improved R-interface inla-binary is now bundled with the R-package. Make use of model.matrix() in R: formula ~... + x*z will expand as formula ~... Allow for formula ~... + x + z + x:z + offset(a) + offset(b) defining a+b to be fixed offset in formula Larger change for survival-models (more on this tomorrow...)
99 Recent developments Recent developments in (R-)INLA Improved R-interface inla-binary is now bundled with the R-package. Make use of model.matrix() in R: formula ~... + x*z will expand as formula ~... Allow for formula ~... + x + z + x:z + offset(a) + offset(b) defining a+b to be fixed offset in formula Larger change for survival-models (more on this tomorrow...)
100 Recent developments Building blocks [13] "3diidwishartpart0" "3diidwishartpart1" "3diidwishartpart2" > names(inla.models()$models) [1] "iid" "rw1" "rw2" [4] "crw2" "seasonal" "besag" [7] "ar1" "generic" "2diid" [10] "2diidwishart" "2diidwishartpart0" "2diidwishartpart1" [16] "z" "rw2d" "matern2d"
101 Recent developments The z -model The z-models is η =... + Zz +... where Z is a n k matrix and z N k (0, τi).
102 Recent developments Likelihood models > names(inla.models()$lmodels) [1] "poisson" "binomial" [3] "exponential" "piecewise.constant" [5] "piecewise.linear" "gaussian" [7] "laplace" "weibull" [9] "weibullcure" "stochvol_t" [11] "zeroinflated_poisson_0" "zeroinflated_poisson_1" [13] "zeroinflated_binomial_0" "zeroinflated_binomial_1" [15] "T" "stochvol.nig"
103 Recent developments Zero-inflated models The (type 0) likelihood is defined as Prob(y...) = p 1 [y=0] + (1 p) Poission(y y > 0) The (type 1) likelihood is defined as Prob(y...) = p 1 [y=0] + (1 p) Poission(y) [Birgit: Mixture over p is still on the list... sorry.]
104 Recent developments The TODO-list Better support for spatial (GMRF) models. More on this tomorrow (Finn.L). Alternative spline-models, B-splines? (Thomas.K?) Simultaneous credibility intervals Documentation Worked out examples Webpage +++
105 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source
106 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source
107 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source
108 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source
109 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source
110 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source
111 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source
112 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source
113 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source
114 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source
115 Summary Summary (II) It is our view that the prospects of this work are more important than this work itself: Make latent Gaussian models, more applicable, useful and appealing for the end user Allow us to use latent Gaussian models as baseline models, even in cases where more complex models are more appropriate
116 Summary Summary (II) It is our view that the prospects of this work are more important than this work itself: Make latent Gaussian models, more applicable, useful and appealing for the end user Allow us to use latent Gaussian models as baseline models, even in cases where more complex models are more appropriate
117 Summary Summary (II) It is our view that the prospects of this work are more important than this work itself: Make latent Gaussian models, more applicable, useful and appealing for the end user Allow us to use latent Gaussian models as baseline models, even in cases where more complex models are more appropriate
INLA: Integrated Nested Laplace Approximations
INLA: Integrated Nested Laplace Approximations John Paige Statistics Department University of Washington October 10, 2017 1 The problem Markov Chain Monte Carlo (MCMC) takes too long in many settings.
More informationMonte Carlo for Spatial Models
Monte Carlo for Spatial Models Murali Haran Department of Statistics Penn State University Penn State Computational Science Lectures April 2007 Spatial Models Lots of scientific questions involve analyzing
More informationLudwig Fahrmeir Gerhard Tute. Statistical odelling Based on Generalized Linear Model. íecond Edition. . Springer
Ludwig Fahrmeir Gerhard Tute Statistical odelling Based on Generalized Linear Model íecond Edition. Springer Preface to the Second Edition Preface to the First Edition List of Examples List of Figures
More informationarxiv: v1 [stat.co] 13 Feb 2017
Spatial Models with the Integrated Nested Laplace Approximation within Markov Chain Monte Carlo Virgilio Gómez-Rubio a,, Francisco Palmí-Perales a arxiv:1702.03891v1 [stat.co] 13 Feb 2017 a Department
More informationIntroduction to the R-INLA R package. Elias T. Krainski
Introduction to the R-INLA R package Elias T. Krainski 2 Outline Overview Tokyo example GAG urine data Close look at GAG urine example R-INLA details 3 What is R-INLA? The algorithm name: Integrated Nested
More informationGlobal modelling of air pollution using multiple data sources
Global modelling of air pollution using multiple data sources Matthew Thomas SAMBa, University of Bath Email: M.L.Thomas@bath.ac.uk November 11, 015 1/ 3 OUTLINE Motivation Data Sources Existing Approaches
More informationGAMs semi-parametric GLMs. Simon Wood Mathematical Sciences, University of Bath, U.K.
GAMs semi-parametric GLMs Simon Wood Mathematical Sciences, University of Bath, U.K. Generalized linear models, GLM 1. A GLM models a univariate response, y i as g{e(y i )} = X i β where y i Exponential
More informationA Fast and Exact Simulation Algorithm for General Gaussian Markov Random Fields
A Fast and Exact Simulation Algorithm for General Gaussian Markov Random Fields HÅVARD RUE DEPARTMENT OF MATHEMATICAL SCIENCES NTNU, NORWAY FIRST VERSION: FEBRUARY 23, 1999 REVISED: APRIL 23, 1999 SUMMARY
More informationTime Series Analysis by State Space Methods
Time Series Analysis by State Space Methods Second Edition J. Durbin London School of Economics and Political Science and University College London S. J. Koopman Vrije Universiteit Amsterdam OXFORD UNIVERSITY
More informationGeneralized Additive Models
:p Texts in Statistical Science Generalized Additive Models An Introduction with R Simon N. Wood Contents Preface XV 1 Linear Models 1 1.1 A simple linear model 2 Simple least squares estimation 3 1.1.1
More informationGlobal modelling of air pollution using multiple data sources
Global modelling of air pollution using multiple data sources Matthew Thomas M.L.Thomas@bath.ac.uk Supervised by Dr. Gavin Shaddick In collaboration with IHME and WHO June 14, 2016 1/ 1 MOTIVATION Air
More informationEstimation of Item Response Models
Estimation of Item Response Models Lecture #5 ICPSR Item Response Theory Workshop Lecture #5: 1of 39 The Big Picture of Estimation ESTIMATOR = Maximum Likelihood; Mplus Any questions? answers Lecture #5:
More informationDATA ANALYSIS USING HIERARCHICAL GENERALIZED LINEAR MODELS WITH R
DATA ANALYSIS USING HIERARCHICAL GENERALIZED LINEAR MODELS WITH R Lee, Rönnegård & Noh LRN@du.se Lee, Rönnegård & Noh HGLM book 1 / 25 Overview 1 Background to the book 2 A motivating example from my own
More informationDATA ANALYSIS USING HIERARCHICAL GENERALIZED LINEAR MODELS WITH R
DATA ANALYSIS USING HIERARCHICAL GENERALIZED LINEAR MODELS WITH R Lee, Rönnegård & Noh LRN@du.se Lee, Rönnegård & Noh HGLM book 1 / 24 Overview 1 Background to the book 2 Crack growth example 3 Contents
More informationA spatio-temporal model for extreme precipitation simulated by a climate model.
A spatio-temporal model for extreme precipitation simulated by a climate model. Jonathan Jalbert Joint work with Anne-Catherine Favre, Claude Bélisle and Jean-François Angers STATMOS Workshop: Climate
More informationPackage INLABMA. February 4, 2017
Version 0.1-8 Date 2017-02-03 Encoding latin1 Title Bayesian Model Averaging with INLA Package INLABMA February 4, 2017 Author, Roger Bivand Maintainer Depends R(>= 2.15.0), parallel,
More information08 An Introduction to Dense Continuous Robotic Mapping
NAVARCH/EECS 568, ROB 530 - Winter 2018 08 An Introduction to Dense Continuous Robotic Mapping Maani Ghaffari March 14, 2018 Previously: Occupancy Grid Maps Pose SLAM graph and its associated dense occupancy
More informationBayesian nonparametric regression and density estimation using integrated nested Laplace approximations
Cleveland Clinic Lerner Research Institute From the SelectedWorks of Xiaofeng Wang 2013 Bayesian nonparametric regression and density estimation using integrated nested Laplace approximations Xiaofeng
More informationTheoretical Concepts of Machine Learning
Theoretical Concepts of Machine Learning Part 2 Institute of Bioinformatics Johannes Kepler University, Linz, Austria Outline 1 Introduction 2 Generalization Error 3 Maximum Likelihood 4 Noise Models 5
More informationNORGES TEKNISK-NATURVITENSKAPELIGE UNIVERSITET. Approximating Hidden Gaussian Markov Random Fields PREPRINT STATISTICS NO. 1/2003
NORGES TEKNISK-NATURVITENSKAPELIGE UNIVERSITET Approximating Hidden Gaussian Markov Random Fields by Håvard Rue, Ingelin Steinsland & Sveinung Erland PREPRINT STATISTICS NO. 1/2003 NORWEGIAN UNIVERSITY
More informationVariability in Annual Temperature Profiles
Variability in Annual Temperature Profiles A Multivariate Spatial Analysis of Regional Climate Model Output Tamara Greasby, Stephan Sain Institute for Mathematics Applied to Geosciences, National Center
More informationBig Data in Environmental Research Pre-World Congress Meeting of New Researchers in Statistics and Probability Problem sheet
Big Data in Environmental Research Pre-World Congress Meeting of New Researchers in Statistics and Probability Problem sheet Website: http://www.stat.ubc.ca/~gavin/stepibooknewstyle/course_toronto.html
More informationPackage acebayes. R topics documented: November 21, Type Package
Type Package Package acebayes November 21, 2018 Title Optimal Bayesian Experimental Design using the ACE Algorithm Version 1.5.2 Date 2018-11-21 Author Antony M. Overstall, David C. Woods & Maria Adamou
More informationNested Sampling: Introduction and Implementation
UNIVERSITY OF TEXAS AT SAN ANTONIO Nested Sampling: Introduction and Implementation Liang Jing May 2009 1 1 ABSTRACT Nested Sampling is a new technique to calculate the evidence, Z = P(D M) = p(d θ, M)p(θ
More informationQuantitative Biology II!
Quantitative Biology II! Lecture 3: Markov Chain Monte Carlo! March 9, 2015! 2! Plan for Today!! Introduction to Sampling!! Introduction to MCMC!! Metropolis Algorithm!! Metropolis-Hastings Algorithm!!
More informationFMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu
FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)
More informationBayesian Analysis of Differential Gene Expression
Bayesian Analysis of Differential Gene Expression Biostat Journal Club Chuan Zhou chuan.zhou@vanderbilt.edu Department of Biostatistics Vanderbilt University Bayesian Modeling p. 1/1 Lewin et al., 2006
More informationSTAT 705 Introduction to generalized additive models
STAT 705 Introduction to generalized additive models Timothy Hanson Department of Statistics, University of South Carolina Stat 705: Data Analysis II 1 / 22 Generalized additive models Consider a linear
More informationImage analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis
7 Computer Vision and Classification 413 / 458 Computer Vision and Classification The k-nearest-neighbor method The k-nearest-neighbor (knn) procedure has been used in data analysis and machine learning
More informationLocal spatial-predictor selection
University of Wollongong Research Online Centre for Statistical & Survey Methodology Working Paper Series Faculty of Engineering and Information Sciences 2013 Local spatial-predictor selection Jonathan
More informationCHAPTER 1 INTRODUCTION
Introduction CHAPTER 1 INTRODUCTION Mplus is a statistical modeling program that provides researchers with a flexible tool to analyze their data. Mplus offers researchers a wide choice of models, estimators,
More informationMarkov Chain Monte Carlo (part 1)
Markov Chain Monte Carlo (part 1) Edps 590BAY Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Spring 2018 Depending on the book that you select for
More informationMachine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme, Nicolas Schilling
Machine Learning B. Unsupervised Learning B.1 Cluster Analysis Lars Schmidt-Thieme, Nicolas Schilling Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University of Hildesheim,
More informationCalibration and emulation of TIE-GCM
Calibration and emulation of TIE-GCM Serge Guillas School of Mathematics Georgia Institute of Technology Jonathan Rougier University of Bristol Big Thanks to Crystal Linkletter (SFU-SAMSI summer school)
More informationCSC412: Stochastic Variational Inference. David Duvenaud
CSC412: Stochastic Variational Inference David Duvenaud Admin A3 will be released this week and will be shorter Motivation for REINFORCE Class projects Class Project ideas Develop a generative model for
More informationMixture Models and the EM Algorithm
Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine c 2017 1 Finite Mixture Models Say we have a data set D = {x 1,..., x N } where x i is
More informationBootstrapping Methods
Bootstrapping Methods example of a Monte Carlo method these are one Monte Carlo statistical method some Bayesian statistical methods are Monte Carlo we can also simulate models using Monte Carlo methods
More informationBayesian model selection and diagnostics
Bayesian model selection and diagnostics A typical Bayesian analysis compares a handful of models. Example 1: Consider the spline model for the motorcycle data, how many basis functions? Example 2: Consider
More informationApplying multiple imputation on waterbird census data Comparing two imputation methods
Applying multiple imputation on waterbird census data Comparing two imputation methods ISEC, Montpellier, 1 july 2014 Thierry Onkelinx, Koen Devos & Paul Quataert Thierry.Onkelinx@inbo.be Contents 1 Introduction
More informationGenerative and discriminative classification techniques
Generative and discriminative classification techniques Machine Learning and Category Representation 2014-2015 Jakob Verbeek, November 28, 2014 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.14.15
More informationDeep Generative Models Variational Autoencoders
Deep Generative Models Variational Autoencoders Sudeshna Sarkar 5 April 2017 Generative Nets Generative models that represent probability distributions over multiple variables in some way. Directed Generative
More informationCS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas
CS839: Probabilistic Graphical Models Lecture 10: Learning with Partially Observed Data Theo Rekatsinas 1 Partially Observed GMs Speech recognition 2 Partially Observed GMs Evolution 3 Partially Observed
More informationAn Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework
IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXX 23 An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework Ji Won Yoon arxiv:37.99v [cs.lg] 3 Jul 23 Abstract In order to cluster
More informationMonte Carlo Methods and Statistical Computing: My Personal E
Monte Carlo Methods and Statistical Computing: My Personal Experience Department of Mathematics & Statistics Indian Institute of Technology Kanpur November 29, 2014 Outline Preface 1 Preface 2 3 4 5 6
More informationMachine Learning A WS15/16 1sst KU Version: January 11, b) [1 P] For the probability distribution P (A, B, C, D) with the factorization
Machine Learning A 708.064 WS15/16 1sst KU Version: January 11, 2016 Exercises Problems marked with * are optional. 1 Conditional Independence I [3 P] a) [1 P] For the probability distribution P (A, B,
More informationModeling Criminal Careers as Departures From a Unimodal Population Age-Crime Curve: The Case of Marijuana Use
Modeling Criminal Careers as Departures From a Unimodal Population Curve: The Case of Marijuana Use Donatello Telesca, Elena A. Erosheva, Derek A. Kreader, & Ross Matsueda April 15, 2014 extends Telesca
More informationMachine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves
Machine Learning A 708.064 11W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence I [2 P] a) [1 P] Give an example for a probability distribution P (A, B, C) that disproves
More information1 Methods for Posterior Simulation
1 Methods for Posterior Simulation Let p(θ y) be the posterior. simulation. Koop presents four methods for (posterior) 1. Monte Carlo integration: draw from p(θ y). 2. Gibbs sampler: sequentially drawing
More informationHow to Price a House
How to Price a House An Interpretable Bayesian Approach Dustin Lennon dustin@inferentialist.com Inferentialist Consulting Seattle, WA April 9, 2014 Introduction Project to tie up loose ends / came out
More informationBayesian Modelling with JAGS and R
Bayesian Modelling with JAGS and R Martyn Plummer International Agency for Research on Cancer Rencontres R, 3 July 2012 CRAN Task View Bayesian Inference The CRAN Task View Bayesian Inference is maintained
More informationRobot Mapping. Least Squares Approach to SLAM. Cyrill Stachniss
Robot Mapping Least Squares Approach to SLAM Cyrill Stachniss 1 Three Main SLAM Paradigms Kalman filter Particle filter Graphbased least squares approach to SLAM 2 Least Squares in General Approach for
More informationGraphbased. Kalman filter. Particle filter. Three Main SLAM Paradigms. Robot Mapping. Least Squares Approach to SLAM. Least Squares in General
Robot Mapping Three Main SLAM Paradigms Least Squares Approach to SLAM Kalman filter Particle filter Graphbased Cyrill Stachniss least squares approach to SLAM 1 2 Least Squares in General! Approach for
More informationBART STAT8810, Fall 2017
BART STAT8810, Fall 2017 M.T. Pratola November 1, 2017 Today BART: Bayesian Additive Regression Trees BART: Bayesian Additive Regression Trees Additive model generalizes the single-tree regression model:
More informationWhat is machine learning?
Machine learning, pattern recognition and statistical data modelling Lecture 12. The last lecture Coryn Bailer-Jones 1 What is machine learning? Data description and interpretation finding simpler relationship
More informationPackage glmmml. R topics documented: March 25, Encoding UTF-8 Version Date Title Generalized Linear Models with Clustering
Encoding UTF-8 Version 1.0.3 Date 2018-03-25 Title Generalized Linear Models with Clustering Package glmmml March 25, 2018 Binomial and Poisson regression for clustered data, fixed and random effects with
More informationMachine Learning A W 1sst KU. b) [1 P] For the probability distribution P (A, B, C, D) with the factorization
Machine Learning A 708.064 13W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence a) [1 P] For the probability distribution P (A, B, C, D) with the factorization P (A, B,
More informationPattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition
Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant
More informationCS Introduction to Data Mining Instructor: Abdullah Mueen
CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen LECTURE 8: ADVANCED CLUSTERING (FUZZY AND CO -CLUSTERING) Review: Basic Cluster Analysis Methods (Chap. 10) Cluster Analysis: Basic Concepts
More informationExpectation-Maximization Methods in Population Analysis. Robert J. Bauer, Ph.D. ICON plc.
Expectation-Maximization Methods in Population Analysis Robert J. Bauer, Ph.D. ICON plc. 1 Objective The objective of this tutorial is to briefly describe the statistical basis of Expectation-Maximization
More informationMCMC Diagnostics. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) MCMC Diagnostics MATH / 24
MCMC Diagnostics Yingbo Li Clemson University MATH 9810 Yingbo Li (Clemson) MCMC Diagnostics MATH 9810 1 / 24 Convergence to Posterior Distribution Theory proves that if a Gibbs sampler iterates enough,
More informationMissing Data Analysis for the Employee Dataset
Missing Data Analysis for the Employee Dataset 67% of the observations have missing values! Modeling Setup Random Variables: Y i =(Y i1,...,y ip ) 0 =(Y i,obs, Y i,miss ) 0 R i =(R i1,...,r ip ) 0 ( 1
More informationUnified Methods for Censored Longitudinal Data and Causality
Mark J. van der Laan James M. Robins Unified Methods for Censored Longitudinal Data and Causality Springer Preface v Notation 1 1 Introduction 8 1.1 Motivation, Bibliographic History, and an Overview of
More informationLecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010
Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010 Tim Hanson, Ph.D. University of South Carolina T. Hanson (USC) Stat 704: Data Analysis I, Fall 2010 1 / 26 Additive predictors
More informationStatistics (STAT) Statistics (STAT) 1. Prerequisites: grade in C- or higher in STAT 1200 or STAT 1300 or STAT 1400
Statistics (STAT) 1 Statistics (STAT) STAT 1200: Introductory Statistical Reasoning Statistical concepts for critically evaluation quantitative information. Descriptive statistics, probability, estimation,
More informationBayesian Computation: Overview and Methods for Low-Dimensional Models
Bayesian Computation: Overview and Methods for Low-Dimensional Models Tom Loredo Dept. of Astronomy, Cornell University http://www.astro.cornell.edu/staff/loredo/bayes/ IAC Winter School, 3 4 Nov 2014
More informationInference and Representation
Inference and Representation Rachel Hodos New York University Lecture 5, October 6, 2015 Rachel Hodos Lecture 5: Inference and Representation Today: Learning with hidden variables Outline: Unsupervised
More informationK-Means and Gaussian Mixture Models
K-Means and Gaussian Mixture Models David Rosenberg New York University June 15, 2015 David Rosenberg (New York University) DS-GA 1003 June 15, 2015 1 / 43 K-Means Clustering Example: Old Faithful Geyser
More informationAcknowledgments. Acronyms
Acknowledgments Preface Acronyms xi xiii xv 1 Basic Tools 1 1.1 Goals of inference 1 1.1.1 Population or process? 1 1.1.2 Probability samples 2 1.1.3 Sampling weights 3 1.1.4 Design effects. 5 1.2 An introduction
More informationPackage ICsurv. February 19, 2015
Package ICsurv February 19, 2015 Type Package Title A package for semiparametric regression analysis of interval-censored data Version 1.0 Date 2014-6-9 Author Christopher S. McMahan and Lianming Wang
More informationMultiple imputation using chained equations: Issues and guidance for practice
Multiple imputation using chained equations: Issues and guidance for practice Ian R. White, Patrick Royston and Angela M. Wood http://onlinelibrary.wiley.com/doi/10.1002/sim.4067/full By Gabrielle Simoneau
More informationPreface to the Second Edition. Preface to the First Edition. 1 Introduction 1
Preface to the Second Edition Preface to the First Edition vii xi 1 Introduction 1 2 Overview of Supervised Learning 9 2.1 Introduction... 9 2.2 Variable Types and Terminology... 9 2.3 Two Simple Approaches
More informationAn introduction to SPSS
An introduction to SPSS To open the SPSS software using U of Iowa Virtual Desktop... Go to https://virtualdesktop.uiowa.edu and choose SPSS 24. Contents NOTE: Save data files in a drive that is accessible
More informationStructured Learning. Jun Zhu
Structured Learning Jun Zhu Supervised learning Given a set of I.I.D. training samples Learn a prediction function b r a c e Supervised learning (cont d) Many different choices Logistic Regression Maximum
More informationBayesian Estimation for Skew Normal Distributions Using Data Augmentation
The Korean Communications in Statistics Vol. 12 No. 2, 2005 pp. 323-333 Bayesian Estimation for Skew Normal Distributions Using Data Augmentation Hea-Jung Kim 1) Abstract In this paper, we develop a MCMC
More informationHierarchical Modelling for Large Spatial Datasets
Hierarchical Modelling for Large Spatial Datasets Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics,
More informationThe glmmml Package. August 20, 2006
The glmmml Package August 20, 2006 Version 0.65-1 Date 2006/08/20 Title Generalized linear models with clustering A Maximum Likelihood and bootstrap approach to mixed models. License GPL version 2 or newer.
More informationSGN (4 cr) Chapter 11
SGN-41006 (4 cr) Chapter 11 Clustering Jussi Tohka & Jari Niemi Department of Signal Processing Tampere University of Technology February 25, 2014 J. Tohka & J. Niemi (TUT-SGN) SGN-41006 (4 cr) Chapter
More informationBayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19
Bayesian Methods David Rosenberg New York University April 11, 2017 David Rosenberg (New York University) DS-GA 1003 April 11, 2017 1 / 19 Classical Statistics Classical Statistics David Rosenberg (New
More informationGAMs, GAMMs and other penalized GLMs using mgcv in R. Simon Wood Mathematical Sciences, University of Bath, U.K.
GAMs, GAMMs and other penalied GLMs using mgcv in R Simon Wood Mathematical Sciences, University of Bath, U.K. Simple eample Consider a very simple dataset relating the timber volume of cherry trees to
More informationECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov
ECE521: Week 11, Lecture 20 27 March 2017: HMM learning/inference With thanks to Russ Salakhutdinov Examples of other perspectives Murphy 17.4 End of Russell & Norvig 15.2 (Artificial Intelligence: A Modern
More informationSummary: A Tutorial on Learning With Bayesian Networks
Summary: A Tutorial on Learning With Bayesian Networks Markus Kalisch May 5, 2006 We primarily summarize [4]. When we think that it is appropriate, we comment on additional facts and more recent developments.
More informationPackage mcemglm. November 29, 2015
Type Package Package mcemglm November 29, 2015 Title Maximum Likelihood Estimation for Generalized Linear Mixed Models Version 1.1 Date 2015-11-28 Author Felipe Acosta Archila Maintainer Maximum likelihood
More informationScalable Multidimensional Hierarchical Bayesian Modeling on Spark
Scalable Multidimensional Hierarchical Bayesian Modeling on Spark Robert Ormandi, Hongxia Yang and Quan Lu Yahoo! Sunnyvale, CA 2015 Click-Through-Rate (CTR) Prediction Estimating the probability of click
More informationDiscussion on Bayesian Model Selection and Parameter Estimation in Extragalactic Astronomy by Martin Weinberg
Discussion on Bayesian Model Selection and Parameter Estimation in Extragalactic Astronomy by Martin Weinberg Phil Gregory Physics and Astronomy Univ. of British Columbia Introduction Martin Weinberg reported
More informationA Bayesian General Linear Modeling Approach to Cortical Surface fmri Data Analysis
A Bayesian General Linear Modeling Approach to Cortical Surface fmri Data Analysis Amanda F. Mejia a, Yu Ryan Yue b, David Bolin c, Finn Lindgren d and Martin A. Lindquist e arxiv:1706.00959v1 [stat.ap]
More information3D Human Motion Analysis and Manifolds
D E P A R T M E N T O F C O M P U T E R S C I E N C E U N I V E R S I T Y O F C O P E N H A G E N 3D Human Motion Analysis and Manifolds Kim Steenstrup Pedersen DIKU Image group and E-Science center Motivation
More informationLinear Modeling with Bayesian Statistics
Linear Modeling with Bayesian Statistics Bayesian Approach I I I I I Estimate probability of a parameter State degree of believe in specific parameter values Evaluate probability of hypothesis given the
More informationlow bias high variance high bias low variance error test set training set high low Model Complexity Typical Behaviour Lecture 11:
Lecture 11: Overfitting and Capacity Control high bias low variance Typical Behaviour low bias high variance Sam Roweis error test set training set November 23, 4 low Model Complexity high Generalization,
More informationIntroduction to Mobile Robotics
Introduction to Mobile Robotics Gaussian Processes Wolfram Burgard Cyrill Stachniss Giorgio Grisetti Maren Bennewitz Christian Plagemann SS08, University of Freiburg, Department for Computer Science Announcement
More informationSplines and penalized regression
Splines and penalized regression November 23 Introduction We are discussing ways to estimate the regression function f, where E(y x) = f(x) One approach is of course to assume that f has a certain shape,
More informationWildfire Chances and Probabilistic Risk Assessment
Wildfire Chances and Probabilistic Risk Assessment D R Brillinger, H K Preisler and H M Naderi Statistics Department University of California Berkeley, CA, 97-386 Pacific Southwest Research Station USDA
More informationIssues in MCMC use for Bayesian model fitting. Practical Considerations for WinBUGS Users
Practical Considerations for WinBUGS Users Kate Cowles, Ph.D. Department of Statistics and Actuarial Science University of Iowa 22S:138 Lecture 12 Oct. 3, 2003 Issues in MCMC use for Bayesian model fitting
More informationMachine Learning / Jan 27, 2010
Revisiting Logistic Regression & Naïve Bayes Aarti Singh Machine Learning 10-701/15-781 Jan 27, 2010 Generative and Discriminative Classifiers Training classifiers involves learning a mapping f: X -> Y,
More informationSpatial Statistics using R-INLA and Gaussian Markov random fields
Spatial Statistics using R-INLA and Gaussian Markov random fields David Bolin and Johan Lindström 1 Introduction In this lab we will look at an example of how to use the SPDE models in the INLA package.
More informationInfectious Disease Models. Angie Dobbs. A Thesis. Presented to. The University of Guelph. In partial fulfilment of requirements.
Issues of Computational Efficiency and Model Approximation for Individual-Level Infectious Disease Models by Angie Dobbs A Thesis Presented to The University of Guelph In partial fulfilment of requirements
More informationMissing Data Analysis for the Employee Dataset
Missing Data Analysis for the Employee Dataset 67% of the observations have missing values! Modeling Setup For our analysis goals we would like to do: Y X N (X, 2 I) and then interpret the coefficients
More informationRegression. Dr. G. Bharadwaja Kumar VIT Chennai
Regression Dr. G. Bharadwaja Kumar VIT Chennai Introduction Statistical models normally specify how one set of variables, called dependent variables, functionally depend on another set of variables, called
More informationSpatial Interpolation & Geostatistics
(Z i Z j ) 2 / 2 Spatial Interpolation & Geostatistics Lag Lag Mean Distance between pairs of points 1 Tobler s Law All places are related, but nearby places are related more than distant places Corollary:
More informationECE521 Lecture 18 Graphical Models Hidden Markov Models
ECE521 Lecture 18 Graphical Models Hidden Markov Models Outline Graphical models Conditional independence Conditional independence after marginalization Sequence models hidden Markov models 2 Graphical
More informationComputer vision: models, learning and inference. Chapter 10 Graphical Models
Computer vision: models, learning and inference Chapter 10 Graphical Models Independence Two variables x 1 and x 2 are independent if their joint probability distribution factorizes as Pr(x 1, x 2 )=Pr(x
More information