INLA: an introduction

Size: px
Start display at page:

Download "INLA: an introduction"

Transcription

1 INLA: an introduction Håvard Rue 1 Norwegian University of Science and Technology Trondheim, Norway May Joint work with S.Martino (Trondheim) and N.Chopin (Paris)

2 Latent Gaussian models Background Latent Gaussian models Stage 1 Observed data y = (y i ), y i x, θ π(y i x i, θ) Stage 2 Latent Gaussian field x θ N (µ, Q(θ) 1 ), Ax = 0 Stage 3 Priors for the hyperparameters θ π(θ) Unify many of the most used models in statistics

3 Latent Gaussian models Background Latent Gaussian models Stage 1 Observed data y = (y i ), y i x, θ π(y i x i, θ) Stage 2 Latent Gaussian field x θ N (µ, Q(θ) 1 ), Ax = 0 Stage 3 Priors for the hyperparameters θ π(θ) Unify many of the most used models in statistics

4 Latent Gaussian models Background Latent Gaussian models Stage 1 Observed data y = (y i ), y i x, θ π(y i x i, θ) Stage 2 Latent Gaussian field x θ N (µ, Q(θ) 1 ), Ax = 0 Stage 3 Priors for the hyperparameters θ π(θ) Unify many of the most used models in statistics

5 Latent Gaussian models Background Latent Gaussian models Stage 1 Observed data y = (y i ), y i x, θ π(y i x i, θ) Stage 2 Latent Gaussian field x θ N (µ, Q(θ) 1 ), Ax = 0 Stage 3 Priors for the hyperparameters θ π(θ) Unify many of the most used models in statistics

6 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ

7 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ

8 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ

9 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ

10 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ

11 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ

12 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ

13 Latent Gaussian models Background Structured additive regression models Linear predictor η i = k β k z ki + j w ji f j (z ji ) + ɛ i Linear effects of covariates {z ki } Effects of f j ( ) Fixed weights {w ji } Commonly: f j (z ji ) = f j,zji Account for smooth response Temporal or spatially indexed covariates Unstructured terms ( random effects ) Depend on some parameters θ

14 Latent Gaussian models Background Structured additive regression models η i = k β k z ki + j w ji f j (z ji ) + ɛ i Observations y π(y x, θ) = i π(y i η i, θ) from an (f.ex) exponential family with mean µ i = g 1 (η i ). Latent Gaussian model if x = ({β k }, {f ji }, {η i }) θ N (µ, Q(θ) 1 )

15 Latent Gaussian models Background Structured additive regression models η i = k β k z ki + j w ji f j (z ji ) + ɛ i Observations y π(y x, θ) = i π(y i η i, θ) from an (f.ex) exponential family with mean µ i = g 1 (η i ). Latent Gaussian model if x = ({β k }, {f ji }, {η i }) θ N (µ, Q(θ) 1 )

16 Latent Gaussian models Background Structured additive regression models η i = k β k z ki + j w ji f j (z ji ) + ɛ i Observations y π(y x, θ) = i π(y i η i, θ) from an (f.ex) exponential family with mean µ i = g 1 (η i ). Latent Gaussian model if x = ({β k }, {f ji }, {η i }) θ N (µ, Q(θ) 1 )

17 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

18 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

19 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

20 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

21 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

22 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

23 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

24 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

25 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

26 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

27 Latent Gaussian models Examples Examples Dynamic linear models Stochastic volatility Generalised linear (mixed) models Generalised additive (mixed) models Spline smoothing Semiparametric regression Space-varying (semiparametric) regression models Disease mapping Log-Gaussian Cox-processes Model-based geostatistics (*) Spatio-temporal models

28 Latent Gaussian models Examples Example: Disease mapping (BYM-model) Data y i Poisson(E i exp(η i )) Log-relative risk η i = u i + v i + β T z i Structured component u Unstructured component v Covariates z i Log-precisions log κ u and log κ v

29 Latent Gaussian models Examples Characteristic features Large dimension of the latent Gaussian field: A lot of conditional independence in the latent Gaussian field Few hyperparameters θ: dim(θ) between 1 and 5 Non-Gaussian data

30 Latent Gaussian models Examples Characteristic features Large dimension of the latent Gaussian field: A lot of conditional independence in the latent Gaussian field Few hyperparameters θ: dim(θ) between 1 and 5 Non-Gaussian data

31 Latent Gaussian models Examples Characteristic features Large dimension of the latent Gaussian field: A lot of conditional independence in the latent Gaussian field Few hyperparameters θ: dim(θ) between 1 and 5 Non-Gaussian data

32 Latent Gaussian models Examples Characteristic features Large dimension of the latent Gaussian field: A lot of conditional independence in the latent Gaussian field Few hyperparameters θ: dim(θ) between 1 and 5 Non-Gaussian data

33 Task Main task Compute the posterior marginals for the latent field π(x i y), i = 1,..., n Compute the posterior marginals for the hyperparameters π(θ j y), j = 1,..., dim(θ) Today s standard approach, is to make use of MCMC Main difficulties CPU-time Additive MC-errors

34 Task Main task Compute the posterior marginals for the latent field π(x i y), i = 1,..., n Compute the posterior marginals for the hyperparameters π(θ j y), j = 1,..., dim(θ) Today s standard approach, is to make use of MCMC Main difficulties CPU-time Additive MC-errors

35 Task Main task Compute the posterior marginals for the latent field π(x i y), i = 1,..., n Compute the posterior marginals for the hyperparameters π(θ j y), j = 1,..., dim(θ) Today s standard approach, is to make use of MCMC Main difficulties CPU-time Additive MC-errors

36 Task Main task Compute the posterior marginals for the latent field π(x i y), i = 1,..., n Compute the posterior marginals for the hyperparameters π(θ j y), j = 1,..., dim(θ) Today s standard approach, is to make use of MCMC Main difficulties CPU-time Additive MC-errors

37 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations

38 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations

39 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations

40 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations

41 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations

42 Our approach INLA Integrated nested Laplace approximations (INLA) Utilise of the latent Gaussian field Laplace approximations Utilise the conditional independence properties of the latent Gaussian field Numerical algorithms for sparse matrices Utilise small dim(θ) Integrated Nested Laplace approximations

43 Our approach Summary of results Main results HUGE improvement in both speed and accuracy compared to MCMC alternatives Relative error Practically exact results 2 Extensions: Marginal likelihood, DIC, Cross-validation,... INLA enable us to treat Bayesian latent Gaussian models properly and bring these models from the research communities to the end-users 2 Can construct counter-examples

44 Our approach Summary of results Main results HUGE improvement in both speed and accuracy compared to MCMC alternatives Relative error Practically exact results 2 Extensions: Marginal likelihood, DIC, Cross-validation,... INLA enable us to treat Bayesian latent Gaussian models properly and bring these models from the research communities to the end-users 2 Can construct counter-examples

45 Our approach Summary of results Main results HUGE improvement in both speed and accuracy compared to MCMC alternatives Relative error Practically exact results 2 Extensions: Marginal likelihood, DIC, Cross-validation,... INLA enable us to treat Bayesian latent Gaussian models properly and bring these models from the research communities to the end-users 2 Can construct counter-examples

46 Our approach Summary of results Main results HUGE improvement in both speed and accuracy compared to MCMC alternatives Relative error Practically exact results 2 Extensions: Marginal likelihood, DIC, Cross-validation,... INLA enable us to treat Bayesian latent Gaussian models properly and bring these models from the research communities to the end-users 2 Can construct counter-examples

47 Our approach Summary of results Main results HUGE improvement in both speed and accuracy compared to MCMC alternatives Relative error Practically exact results 2 Extensions: Marginal likelihood, DIC, Cross-validation,... INLA enable us to treat Bayesian latent Gaussian models properly and bring these models from the research communities to the end-users 2 Can construct counter-examples

48 Main ideas Main ideas (I) π(z) = π(x, z) π(x z) leading to π(z) = π(x, z) π(x z) mode(z) When π(x z) is the Gaussian-approximation, this is the Laplace-approximation Want π(x z) to be almost Gaussian.

49 Main ideas Main ideas (I) π(z) = π(x, z) π(x z) leading to π(z) = π(x, z) π(x z) mode(z) When π(x z) is the Gaussian-approximation, this is the Laplace-approximation Want π(x z) to be almost Gaussian.

50 Main ideas Main ideas (I) π(z) = π(x, z) π(x z) leading to π(z) = π(x, z) π(x z) mode(z) When π(x z) is the Gaussian-approximation, this is the Laplace-approximation Want π(x z) to be almost Gaussian.

51 Main ideas Main ideas (II) Posterior π(x, θ y) π(θ) π(x θ) i I π(y i x i, θ) Do the integration wrt θ numerically

52 Main ideas Main ideas (II) Posterior π(x, θ y) π(θ) π(x θ) i I π(y i x i, θ) Do the integration wrt θ numerically π(x i y) = π(θ j y) = π(θ y) π(x i θ, y) dθ π(θ y) dθ j

53 Main ideas Main ideas (II) Posterior π(x, θ y) π(θ) π(x θ) i I π(y i x i, θ) Do the integration wrt θ numerically π(x i y) = π(θ j y) = π(θ y) π(x i θ, y) dθ π(θ y) dθ j

54 Main ideas Remarks 1. Expect π(θ y) to be accurate, since x θ is a priori Gaussian Likelihood models are well-behaved so is almost Gaussian. π(x θ, y) 2. There are no distributional assumptions on θ y 3. Similar remarks are valid to π(x i θ, y)

55 Main ideas Remarks 1. Expect π(θ y) to be accurate, since x θ is a priori Gaussian Likelihood models are well-behaved so is almost Gaussian. π(x θ, y) 2. There are no distributional assumptions on θ y 3. Similar remarks are valid to π(x i θ, y)

56 Main ideas Remarks 1. Expect π(θ y) to be accurate, since x θ is a priori Gaussian Likelihood models are well-behaved so is almost Gaussian. π(x θ, y) 2. There are no distributional assumptions on θ y 3. Similar remarks are valid to π(x i θ, y)

57 Main ideas Remarks 1. Expect π(θ y) to be accurate, since x θ is a priori Gaussian Likelihood models are well-behaved so is almost Gaussian. π(x θ, y) 2. There are no distributional assumptions on θ y 3. Similar remarks are valid to π(x i θ, y)

58 Main ideas Remarks 1. Expect π(θ y) to be accurate, since x θ is a priori Gaussian Likelihood models are well-behaved so is almost Gaussian. π(x θ, y) 2. There are no distributional assumptions on θ y 3. Similar remarks are valid to π(x i θ, y)

59 Computational tools Background Computational issues Our approach in its raw form is not computational feasible Main issue is the large dimension, n, of the latent field Various strategies/tricks are required for obtaining a practical good solution

60 Computational tools Background Computational issues Our approach in its raw form is not computational feasible Main issue is the large dimension, n, of the latent field Various strategies/tricks are required for obtaining a practical good solution

61 Computational tools Background Computational issues Our approach in its raw form is not computational feasible Main issue is the large dimension, n, of the latent field Various strategies/tricks are required for obtaining a practical good solution

62 Computational tools Gaussian Markov random fields Gaussian Markov random fields Make use of the conditional independence properties in the latent field x i x j x ij Q ij = 0 where Q is the precision matrix (inverse covariance) Use numerical methods for sparse matrices!

63 Computational tools Gaussian Markov random fields Gaussian Markov random fields Make use of the conditional independence properties in the latent field x i x j x ij Q ij = 0 where Q is the precision matrix (inverse covariance) Use numerical methods for sparse matrices!

64 Computational tools Rank one updates Rank one updates Have computed the Cholesky factorisation Prec(x) = Q = LL T Want to compute the conditional mean and variances for x i x i No need to factorise Prec(x i x i ); can compute the correction for conditioning on x i.

65 Computational tools Rank one updates Rank one updates Have computed the Cholesky factorisation Prec(x) = Q = LL T Want to compute the conditional mean and variances for x i x i No need to factorise Prec(x i x i ); can compute the correction for conditioning on x i.

66 Computational tools Rank one updates Rank one updates Have computed the Cholesky factorisation Prec(x) = Q = LL T Want to compute the conditional mean and variances for x i x i No need to factorise Prec(x i x i ); can compute the correction for conditioning on x i.

67 Computational tools Simplified Laplace Simplified Laplace Approximation Expand the Laplace approximation of π(x i θ, y): log π(x i θ, y) = 1 2 x 2 i + b i x i d i x 3 i + Remarks Correct the Gaussian approximation for error in shift and skewness through b i and d i Fit a skew-normal density Computational fast 2φ(x)Φ(ax) Sufficient accurate for most applications

68 Computational tools Simplified Laplace Simplified Laplace Approximation Expand the Laplace approximation of π(x i θ, y): log π(x i θ, y) = 1 2 x 2 i + b i x i d i x 3 i + Remarks Correct the Gaussian approximation for error in shift and skewness through b i and d i Fit a skew-normal density Computational fast 2φ(x)Φ(ax) Sufficient accurate for most applications

69 Computational tools Simplified Laplace Simplified Laplace Approximation Expand the Laplace approximation of π(x i θ, y): log π(x i θ, y) = 1 2 x 2 i + b i x i d i x 3 i + Remarks Correct the Gaussian approximation for error in shift and skewness through b i and d i Fit a skew-normal density Computational fast 2φ(x)Φ(ax) Sufficient accurate for most applications

70 Computational tools Simplified Laplace Simplified Laplace Approximation Expand the Laplace approximation of π(x i θ, y): log π(x i θ, y) = 1 2 x 2 i + b i x i d i x 3 i + Remarks Correct the Gaussian approximation for error in shift and skewness through b i and d i Fit a skew-normal density Computational fast 2φ(x)Φ(ax) Sufficient accurate for most applications

71 Computational tools Simplified Laplace Simplified Laplace Approximation Expand the Laplace approximation of π(x i θ, y): log π(x i θ, y) = 1 2 x 2 i + b i x i d i x 3 i + Remarks Correct the Gaussian approximation for error in shift and skewness through b i and d i Fit a skew-normal density Computational fast 2φ(x)Φ(ax) Sufficient accurate for most applications

72 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) I Step I Explore π(θ y) Locate the mode Use the Hessian to construct new variables Grid-search

73 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) I Step I Explore π(θ y) Locate the mode Use the Hessian to construct new variables Grid-search

74 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) I Step I Explore π(θ y) Locate the mode Use the Hessian to construct new variables Grid-search

75 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) II Step II For each θ j For each i, compute the (simplified) Laplace approximation for x i

76 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) III Step III Sum out θ j For each i, sum out θ π(x i y) j π(x i y, θ j ) π(θ j y) Build a log-spline corrected Gaussian N (x i ; µ i, σ 2 i ) exp(spline) to represent π(x i y).

77 The Integrated nested Laplace-approximation (INLA) Summary The integrated nested Laplace approximation (INLA) III Step III Sum out θ j For each i, sum out θ π(x i y) j π(x i y, θ j ) π(θ j y) Build a log-spline corrected Gaussian N (x i ; µ i, σ 2 i ) exp(spline) to represent π(x i y).

78 The Integrated nested Laplace-approximation (INLA) Assessing the error How can we assess the errors in the approximations? Important, but asymptotic arguments are difficult: dim(y) = O(n) and dim(x) = O(n)

79 The Integrated nested Laplace-approximation (INLA) Assessing the error Errors in the approximations of π(x i y) Compare a sequence of improved approximations 1. Gaussian approximation 2. Simplified Laplace 3. Laplace Compute the full Laplace-approximation for π(x i y, θ j ) only if the Gaussian and the Simplified Laplace approximation disagree.

80 The Integrated nested Laplace-approximation (INLA) Assessing the error Errors in the approximations of π(x i y) Compare a sequence of improved approximations 1. Gaussian approximation 2. Simplified Laplace 3. Laplace Compute the full Laplace-approximation for π(x i y, θ j ) only if the Gaussian and the Simplified Laplace approximation disagree.

81 The Integrated nested Laplace-approximation (INLA) Assessing the error Overall check: Equivalent number of replicates Tool 3: Estimate the effective number of parameters From the Deviance Information Criteria: p D (θ) n trace ( Q prior (θ) Q post. (θ) 1) Compare with the number of observations: #observations/p D (θ) high ratio is good Theoretical justification

82 The Integrated nested Laplace-approximation (INLA) Assessing the error Overall check: Equivalent number of replicates Tool 3: Estimate the effective number of parameters From the Deviance Information Criteria: p D (θ) n trace ( Q prior (θ) Q post. (θ) 1) Compare with the number of observations: #observations/p D (θ) high ratio is good Theoretical justification

83 The Integrated nested Laplace-approximation (INLA) Assessing the error Overall check: Equivalent number of replicates Tool 3: Estimate the effective number of parameters From the Deviance Information Criteria: p D (θ) n trace ( Q prior (θ) Q post. (θ) 1) Compare with the number of observations: #observations/p D (θ) high ratio is good Theoretical justification

84 Examples Examples 10:30-11:30: Andrea Riebler: Performance of INLA analysing bivariate meta-regression and age-period-cohort models. 12:30-13:30: Birgit Schrödle: Spatio-temporal disease mapping using INLA. 13:45-14:45: Virgilio Gómez-Rubio: Approximate Bayesian Inference for Small Area Estimation 15:00-16:00: Lea Fortunato: About the Rapid Inquiry Facility, and spatial analyses with WinBUGS and INLA. 09:00-10:00: Rupali Akerkar: Approximate Bayesian Inference for Survival models. 10:15-11:30: Ingelin Steinsland & Anna Marie Holand: Animal Model and INLA.

85 Extensions Marginal likelihood Marginal likelihood Marginal likelihood is the normalising constant for π(θ y)

86 Extensions DIC Deviance Information Criteria D(x; θ) = 2 i log(y i x i, θ) DIC = 2 Mean (D(x; θ)) D(Mean(x); θ )

87 Cross-validation Cross-validation Based on we can compute π(x i y i, θ) π(x i y, θ) π(y i x i, θ) π(y i y i ) Similar with π(θ y i ) Keep the integration points {θ j } fixed. Detect surprising observations: Prob(y new i y i y i )

88 Cross-validation Cross-validation Based on we can compute π(x i y i, θ) π(x i y, θ) π(y i x i, θ) π(y i y i ) Similar with π(θ y i ) Keep the integration points {θ j } fixed. Detect surprising observations: Prob(y new i y i y i )

89 Cross-validation Cross-validation Based on we can compute π(x i y i, θ) π(x i y, θ) π(y i x i, θ) π(y i y i ) Similar with π(θ y i ) Keep the integration points {θ j } fixed. Detect surprising observations: Prob(y new i y i y i )

90 Cross-validation Cross-validation Based on we can compute π(x i y i, θ) π(x i y, θ) π(y i x i, θ) π(y i y i ) Similar with π(θ y i ) Keep the integration points {θ j } fixed. Detect surprising observations: Prob(y new i y i y i )

91 Recent developments Recent developments in (R-)INLA Improving the code: speedups and improved parallel performance Improving the R-interface Extending the building-blocks: prior-models and likelihood-models Lot of things still todo... Documentation Worked out examples Webpage +++

92 Recent developments Improving the code The two largest changes the last year (summer 2008) Change the way inla works in a multi-core environment. Especially important for more than dual-core (xmas 2008) Implement a more efficient rank-one update formula (Thanks Birgit!). Especially important for models with many constraints, but gave a huge speedup also for models with a single constraint.

93 Recent developments Improving the code The two largest changes the last year (summer 2008) Change the way inla works in a multi-core environment. Especially important for more than dual-core (xmas 2008) Implement a more efficient rank-one update formula (Thanks Birgit!). Especially important for models with many constraints, but gave a huge speedup also for models with a single constraint.

94 Recent developments inla & multi-core Gradient and Hessian computations are done in parallel π(θ y) θ i and 2 θ i θ j π(θ y) All computations of π(x i θ j, y) is now done in parallel wrt j, and not i as before.

95 Recent developments Multi-core example inla-example 7: CANCER-INCIDENCE, n = 7138, dim(θ) = 3 Number of cores Time used s s 4 9.1s 6 8.6s 8 8.3s Linear algebra (Cholesky/Solve/Inverse): 50% Administration: 50%

96 Recent developments Recent developments in (R-)INLA Improved R-interface inla-binary is now bundled with the R-package. Make use of model.matrix() in R: formula ~... + x*z will expand as formula ~... Allow for formula ~... + x + z + x:z + offset(a) + offset(b) defining a+b to be fixed offset in formula Larger change for survival-models (more on this tomorrow...)

97 Recent developments Recent developments in (R-)INLA Improved R-interface inla-binary is now bundled with the R-package. Make use of model.matrix() in R: formula ~... + x*z will expand as formula ~... Allow for formula ~... + x + z + x:z + offset(a) + offset(b) defining a+b to be fixed offset in formula Larger change for survival-models (more on this tomorrow...)

98 Recent developments Recent developments in (R-)INLA Improved R-interface inla-binary is now bundled with the R-package. Make use of model.matrix() in R: formula ~... + x*z will expand as formula ~... Allow for formula ~... + x + z + x:z + offset(a) + offset(b) defining a+b to be fixed offset in formula Larger change for survival-models (more on this tomorrow...)

99 Recent developments Recent developments in (R-)INLA Improved R-interface inla-binary is now bundled with the R-package. Make use of model.matrix() in R: formula ~... + x*z will expand as formula ~... Allow for formula ~... + x + z + x:z + offset(a) + offset(b) defining a+b to be fixed offset in formula Larger change for survival-models (more on this tomorrow...)

100 Recent developments Building blocks [13] "3diidwishartpart0" "3diidwishartpart1" "3diidwishartpart2" > names(inla.models()$models) [1] "iid" "rw1" "rw2" [4] "crw2" "seasonal" "besag" [7] "ar1" "generic" "2diid" [10] "2diidwishart" "2diidwishartpart0" "2diidwishartpart1" [16] "z" "rw2d" "matern2d"

101 Recent developments The z -model The z-models is η =... + Zz +... where Z is a n k matrix and z N k (0, τi).

102 Recent developments Likelihood models > names(inla.models()$lmodels) [1] "poisson" "binomial" [3] "exponential" "piecewise.constant" [5] "piecewise.linear" "gaussian" [7] "laplace" "weibull" [9] "weibullcure" "stochvol_t" [11] "zeroinflated_poisson_0" "zeroinflated_poisson_1" [13] "zeroinflated_binomial_0" "zeroinflated_binomial_1" [15] "T" "stochvol.nig"

103 Recent developments Zero-inflated models The (type 0) likelihood is defined as Prob(y...) = p 1 [y=0] + (1 p) Poission(y y > 0) The (type 1) likelihood is defined as Prob(y...) = p 1 [y=0] + (1 p) Poission(y) [Birgit: Mixture over p is still on the list... sorry.]

104 Recent developments The TODO-list Better support for spatial (GMRF) models. More on this tomorrow (Finn.L). Alternative spline-models, B-splines? (Thomas.K?) Simultaneous credibility intervals Documentation Worked out examples Webpage +++

105 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source

106 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source

107 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source

108 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source

109 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source

110 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source

111 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source

112 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source

113 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source

114 Summary Summary (I) Latent Gaussian models unifies many models into the same framework: unified approach towards Bayesian inference Use the same computer code Near optimal numerical algorithms for the sparse matrices Achieve nice speed-ups in a multi-core environment Practically exact results Prototype implementation inla-program: Inference for structured additive regression models GAM-like R-interface to the inla-program Available for Linux/Mac/Windows Open source

115 Summary Summary (II) It is our view that the prospects of this work are more important than this work itself: Make latent Gaussian models, more applicable, useful and appealing for the end user Allow us to use latent Gaussian models as baseline models, even in cases where more complex models are more appropriate

116 Summary Summary (II) It is our view that the prospects of this work are more important than this work itself: Make latent Gaussian models, more applicable, useful and appealing for the end user Allow us to use latent Gaussian models as baseline models, even in cases where more complex models are more appropriate

117 Summary Summary (II) It is our view that the prospects of this work are more important than this work itself: Make latent Gaussian models, more applicable, useful and appealing for the end user Allow us to use latent Gaussian models as baseline models, even in cases where more complex models are more appropriate

INLA: Integrated Nested Laplace Approximations

INLA: Integrated Nested Laplace Approximations INLA: Integrated Nested Laplace Approximations John Paige Statistics Department University of Washington October 10, 2017 1 The problem Markov Chain Monte Carlo (MCMC) takes too long in many settings.

More information

Monte Carlo for Spatial Models

Monte Carlo for Spatial Models Monte Carlo for Spatial Models Murali Haran Department of Statistics Penn State University Penn State Computational Science Lectures April 2007 Spatial Models Lots of scientific questions involve analyzing

More information

Ludwig Fahrmeir Gerhard Tute. Statistical odelling Based on Generalized Linear Model. íecond Edition. . Springer

Ludwig Fahrmeir Gerhard Tute. Statistical odelling Based on Generalized Linear Model. íecond Edition. . Springer Ludwig Fahrmeir Gerhard Tute Statistical odelling Based on Generalized Linear Model íecond Edition. Springer Preface to the Second Edition Preface to the First Edition List of Examples List of Figures

More information

arxiv: v1 [stat.co] 13 Feb 2017

arxiv: v1 [stat.co] 13 Feb 2017 Spatial Models with the Integrated Nested Laplace Approximation within Markov Chain Monte Carlo Virgilio Gómez-Rubio a,, Francisco Palmí-Perales a arxiv:1702.03891v1 [stat.co] 13 Feb 2017 a Department

More information

Introduction to the R-INLA R package. Elias T. Krainski

Introduction to the R-INLA R package. Elias T. Krainski Introduction to the R-INLA R package Elias T. Krainski 2 Outline Overview Tokyo example GAG urine data Close look at GAG urine example R-INLA details 3 What is R-INLA? The algorithm name: Integrated Nested

More information

Global modelling of air pollution using multiple data sources

Global modelling of air pollution using multiple data sources Global modelling of air pollution using multiple data sources Matthew Thomas SAMBa, University of Bath Email: M.L.Thomas@bath.ac.uk November 11, 015 1/ 3 OUTLINE Motivation Data Sources Existing Approaches

More information

GAMs semi-parametric GLMs. Simon Wood Mathematical Sciences, University of Bath, U.K.

GAMs semi-parametric GLMs. Simon Wood Mathematical Sciences, University of Bath, U.K. GAMs semi-parametric GLMs Simon Wood Mathematical Sciences, University of Bath, U.K. Generalized linear models, GLM 1. A GLM models a univariate response, y i as g{e(y i )} = X i β where y i Exponential

More information

A Fast and Exact Simulation Algorithm for General Gaussian Markov Random Fields

A Fast and Exact Simulation Algorithm for General Gaussian Markov Random Fields A Fast and Exact Simulation Algorithm for General Gaussian Markov Random Fields HÅVARD RUE DEPARTMENT OF MATHEMATICAL SCIENCES NTNU, NORWAY FIRST VERSION: FEBRUARY 23, 1999 REVISED: APRIL 23, 1999 SUMMARY

More information

Time Series Analysis by State Space Methods

Time Series Analysis by State Space Methods Time Series Analysis by State Space Methods Second Edition J. Durbin London School of Economics and Political Science and University College London S. J. Koopman Vrije Universiteit Amsterdam OXFORD UNIVERSITY

More information

Generalized Additive Models

Generalized Additive Models :p Texts in Statistical Science Generalized Additive Models An Introduction with R Simon N. Wood Contents Preface XV 1 Linear Models 1 1.1 A simple linear model 2 Simple least squares estimation 3 1.1.1

More information

Global modelling of air pollution using multiple data sources

Global modelling of air pollution using multiple data sources Global modelling of air pollution using multiple data sources Matthew Thomas M.L.Thomas@bath.ac.uk Supervised by Dr. Gavin Shaddick In collaboration with IHME and WHO June 14, 2016 1/ 1 MOTIVATION Air

More information

Estimation of Item Response Models

Estimation of Item Response Models Estimation of Item Response Models Lecture #5 ICPSR Item Response Theory Workshop Lecture #5: 1of 39 The Big Picture of Estimation ESTIMATOR = Maximum Likelihood; Mplus Any questions? answers Lecture #5:

More information

DATA ANALYSIS USING HIERARCHICAL GENERALIZED LINEAR MODELS WITH R

DATA ANALYSIS USING HIERARCHICAL GENERALIZED LINEAR MODELS WITH R DATA ANALYSIS USING HIERARCHICAL GENERALIZED LINEAR MODELS WITH R Lee, Rönnegård & Noh LRN@du.se Lee, Rönnegård & Noh HGLM book 1 / 25 Overview 1 Background to the book 2 A motivating example from my own

More information

DATA ANALYSIS USING HIERARCHICAL GENERALIZED LINEAR MODELS WITH R

DATA ANALYSIS USING HIERARCHICAL GENERALIZED LINEAR MODELS WITH R DATA ANALYSIS USING HIERARCHICAL GENERALIZED LINEAR MODELS WITH R Lee, Rönnegård & Noh LRN@du.se Lee, Rönnegård & Noh HGLM book 1 / 24 Overview 1 Background to the book 2 Crack growth example 3 Contents

More information

A spatio-temporal model for extreme precipitation simulated by a climate model.

A spatio-temporal model for extreme precipitation simulated by a climate model. A spatio-temporal model for extreme precipitation simulated by a climate model. Jonathan Jalbert Joint work with Anne-Catherine Favre, Claude Bélisle and Jean-François Angers STATMOS Workshop: Climate

More information

Package INLABMA. February 4, 2017

Package INLABMA. February 4, 2017 Version 0.1-8 Date 2017-02-03 Encoding latin1 Title Bayesian Model Averaging with INLA Package INLABMA February 4, 2017 Author, Roger Bivand Maintainer Depends R(>= 2.15.0), parallel,

More information

08 An Introduction to Dense Continuous Robotic Mapping

08 An Introduction to Dense Continuous Robotic Mapping NAVARCH/EECS 568, ROB 530 - Winter 2018 08 An Introduction to Dense Continuous Robotic Mapping Maani Ghaffari March 14, 2018 Previously: Occupancy Grid Maps Pose SLAM graph and its associated dense occupancy

More information

Bayesian nonparametric regression and density estimation using integrated nested Laplace approximations

Bayesian nonparametric regression and density estimation using integrated nested Laplace approximations Cleveland Clinic Lerner Research Institute From the SelectedWorks of Xiaofeng Wang 2013 Bayesian nonparametric regression and density estimation using integrated nested Laplace approximations Xiaofeng

More information

Theoretical Concepts of Machine Learning

Theoretical Concepts of Machine Learning Theoretical Concepts of Machine Learning Part 2 Institute of Bioinformatics Johannes Kepler University, Linz, Austria Outline 1 Introduction 2 Generalization Error 3 Maximum Likelihood 4 Noise Models 5

More information

NORGES TEKNISK-NATURVITENSKAPELIGE UNIVERSITET. Approximating Hidden Gaussian Markov Random Fields PREPRINT STATISTICS NO. 1/2003

NORGES TEKNISK-NATURVITENSKAPELIGE UNIVERSITET. Approximating Hidden Gaussian Markov Random Fields PREPRINT STATISTICS NO. 1/2003 NORGES TEKNISK-NATURVITENSKAPELIGE UNIVERSITET Approximating Hidden Gaussian Markov Random Fields by Håvard Rue, Ingelin Steinsland & Sveinung Erland PREPRINT STATISTICS NO. 1/2003 NORWEGIAN UNIVERSITY

More information

Variability in Annual Temperature Profiles

Variability in Annual Temperature Profiles Variability in Annual Temperature Profiles A Multivariate Spatial Analysis of Regional Climate Model Output Tamara Greasby, Stephan Sain Institute for Mathematics Applied to Geosciences, National Center

More information

Big Data in Environmental Research Pre-World Congress Meeting of New Researchers in Statistics and Probability Problem sheet

Big Data in Environmental Research Pre-World Congress Meeting of New Researchers in Statistics and Probability Problem sheet Big Data in Environmental Research Pre-World Congress Meeting of New Researchers in Statistics and Probability Problem sheet Website: http://www.stat.ubc.ca/~gavin/stepibooknewstyle/course_toronto.html

More information

Package acebayes. R topics documented: November 21, Type Package

Package acebayes. R topics documented: November 21, Type Package Type Package Package acebayes November 21, 2018 Title Optimal Bayesian Experimental Design using the ACE Algorithm Version 1.5.2 Date 2018-11-21 Author Antony M. Overstall, David C. Woods & Maria Adamou

More information

Nested Sampling: Introduction and Implementation

Nested Sampling: Introduction and Implementation UNIVERSITY OF TEXAS AT SAN ANTONIO Nested Sampling: Introduction and Implementation Liang Jing May 2009 1 1 ABSTRACT Nested Sampling is a new technique to calculate the evidence, Z = P(D M) = p(d θ, M)p(θ

More information

Quantitative Biology II!

Quantitative Biology II! Quantitative Biology II! Lecture 3: Markov Chain Monte Carlo! March 9, 2015! 2! Plan for Today!! Introduction to Sampling!! Introduction to MCMC!! Metropolis Algorithm!! Metropolis-Hastings Algorithm!!

More information

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)

More information

Bayesian Analysis of Differential Gene Expression

Bayesian Analysis of Differential Gene Expression Bayesian Analysis of Differential Gene Expression Biostat Journal Club Chuan Zhou chuan.zhou@vanderbilt.edu Department of Biostatistics Vanderbilt University Bayesian Modeling p. 1/1 Lewin et al., 2006

More information

STAT 705 Introduction to generalized additive models

STAT 705 Introduction to generalized additive models STAT 705 Introduction to generalized additive models Timothy Hanson Department of Statistics, University of South Carolina Stat 705: Data Analysis II 1 / 22 Generalized additive models Consider a linear

More information

Image analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis

Image analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis 7 Computer Vision and Classification 413 / 458 Computer Vision and Classification The k-nearest-neighbor method The k-nearest-neighbor (knn) procedure has been used in data analysis and machine learning

More information

Local spatial-predictor selection

Local spatial-predictor selection University of Wollongong Research Online Centre for Statistical & Survey Methodology Working Paper Series Faculty of Engineering and Information Sciences 2013 Local spatial-predictor selection Jonathan

More information

CHAPTER 1 INTRODUCTION

CHAPTER 1 INTRODUCTION Introduction CHAPTER 1 INTRODUCTION Mplus is a statistical modeling program that provides researchers with a flexible tool to analyze their data. Mplus offers researchers a wide choice of models, estimators,

More information

Markov Chain Monte Carlo (part 1)

Markov Chain Monte Carlo (part 1) Markov Chain Monte Carlo (part 1) Edps 590BAY Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Spring 2018 Depending on the book that you select for

More information

Machine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme, Nicolas Schilling

Machine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme, Nicolas Schilling Machine Learning B. Unsupervised Learning B.1 Cluster Analysis Lars Schmidt-Thieme, Nicolas Schilling Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University of Hildesheim,

More information

Calibration and emulation of TIE-GCM

Calibration and emulation of TIE-GCM Calibration and emulation of TIE-GCM Serge Guillas School of Mathematics Georgia Institute of Technology Jonathan Rougier University of Bristol Big Thanks to Crystal Linkletter (SFU-SAMSI summer school)

More information

CSC412: Stochastic Variational Inference. David Duvenaud

CSC412: Stochastic Variational Inference. David Duvenaud CSC412: Stochastic Variational Inference David Duvenaud Admin A3 will be released this week and will be shorter Motivation for REINFORCE Class projects Class Project ideas Develop a generative model for

More information

Mixture Models and the EM Algorithm

Mixture Models and the EM Algorithm Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine c 2017 1 Finite Mixture Models Say we have a data set D = {x 1,..., x N } where x i is

More information

Bootstrapping Methods

Bootstrapping Methods Bootstrapping Methods example of a Monte Carlo method these are one Monte Carlo statistical method some Bayesian statistical methods are Monte Carlo we can also simulate models using Monte Carlo methods

More information

Bayesian model selection and diagnostics

Bayesian model selection and diagnostics Bayesian model selection and diagnostics A typical Bayesian analysis compares a handful of models. Example 1: Consider the spline model for the motorcycle data, how many basis functions? Example 2: Consider

More information

Applying multiple imputation on waterbird census data Comparing two imputation methods

Applying multiple imputation on waterbird census data Comparing two imputation methods Applying multiple imputation on waterbird census data Comparing two imputation methods ISEC, Montpellier, 1 july 2014 Thierry Onkelinx, Koen Devos & Paul Quataert Thierry.Onkelinx@inbo.be Contents 1 Introduction

More information

Generative and discriminative classification techniques

Generative and discriminative classification techniques Generative and discriminative classification techniques Machine Learning and Category Representation 2014-2015 Jakob Verbeek, November 28, 2014 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.14.15

More information

Deep Generative Models Variational Autoencoders

Deep Generative Models Variational Autoencoders Deep Generative Models Variational Autoencoders Sudeshna Sarkar 5 April 2017 Generative Nets Generative models that represent probability distributions over multiple variables in some way. Directed Generative

More information

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 10: Learning with Partially Observed Data Theo Rekatsinas 1 Partially Observed GMs Speech recognition 2 Partially Observed GMs Evolution 3 Partially Observed

More information

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXX 23 An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework Ji Won Yoon arxiv:37.99v [cs.lg] 3 Jul 23 Abstract In order to cluster

More information

Monte Carlo Methods and Statistical Computing: My Personal E

Monte Carlo Methods and Statistical Computing: My Personal E Monte Carlo Methods and Statistical Computing: My Personal Experience Department of Mathematics & Statistics Indian Institute of Technology Kanpur November 29, 2014 Outline Preface 1 Preface 2 3 4 5 6

More information

Machine Learning A WS15/16 1sst KU Version: January 11, b) [1 P] For the probability distribution P (A, B, C, D) with the factorization

Machine Learning A WS15/16 1sst KU Version: January 11, b) [1 P] For the probability distribution P (A, B, C, D) with the factorization Machine Learning A 708.064 WS15/16 1sst KU Version: January 11, 2016 Exercises Problems marked with * are optional. 1 Conditional Independence I [3 P] a) [1 P] For the probability distribution P (A, B,

More information

Modeling Criminal Careers as Departures From a Unimodal Population Age-Crime Curve: The Case of Marijuana Use

Modeling Criminal Careers as Departures From a Unimodal Population Age-Crime Curve: The Case of Marijuana Use Modeling Criminal Careers as Departures From a Unimodal Population Curve: The Case of Marijuana Use Donatello Telesca, Elena A. Erosheva, Derek A. Kreader, & Ross Matsueda April 15, 2014 extends Telesca

More information

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves Machine Learning A 708.064 11W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence I [2 P] a) [1 P] Give an example for a probability distribution P (A, B, C) that disproves

More information

1 Methods for Posterior Simulation

1 Methods for Posterior Simulation 1 Methods for Posterior Simulation Let p(θ y) be the posterior. simulation. Koop presents four methods for (posterior) 1. Monte Carlo integration: draw from p(θ y). 2. Gibbs sampler: sequentially drawing

More information

How to Price a House

How to Price a House How to Price a House An Interpretable Bayesian Approach Dustin Lennon dustin@inferentialist.com Inferentialist Consulting Seattle, WA April 9, 2014 Introduction Project to tie up loose ends / came out

More information

Bayesian Modelling with JAGS and R

Bayesian Modelling with JAGS and R Bayesian Modelling with JAGS and R Martyn Plummer International Agency for Research on Cancer Rencontres R, 3 July 2012 CRAN Task View Bayesian Inference The CRAN Task View Bayesian Inference is maintained

More information

Robot Mapping. Least Squares Approach to SLAM. Cyrill Stachniss

Robot Mapping. Least Squares Approach to SLAM. Cyrill Stachniss Robot Mapping Least Squares Approach to SLAM Cyrill Stachniss 1 Three Main SLAM Paradigms Kalman filter Particle filter Graphbased least squares approach to SLAM 2 Least Squares in General Approach for

More information

Graphbased. Kalman filter. Particle filter. Three Main SLAM Paradigms. Robot Mapping. Least Squares Approach to SLAM. Least Squares in General

Graphbased. Kalman filter. Particle filter. Three Main SLAM Paradigms. Robot Mapping. Least Squares Approach to SLAM. Least Squares in General Robot Mapping Three Main SLAM Paradigms Least Squares Approach to SLAM Kalman filter Particle filter Graphbased Cyrill Stachniss least squares approach to SLAM 1 2 Least Squares in General! Approach for

More information

BART STAT8810, Fall 2017

BART STAT8810, Fall 2017 BART STAT8810, Fall 2017 M.T. Pratola November 1, 2017 Today BART: Bayesian Additive Regression Trees BART: Bayesian Additive Regression Trees Additive model generalizes the single-tree regression model:

More information

What is machine learning?

What is machine learning? Machine learning, pattern recognition and statistical data modelling Lecture 12. The last lecture Coryn Bailer-Jones 1 What is machine learning? Data description and interpretation finding simpler relationship

More information

Package glmmml. R topics documented: March 25, Encoding UTF-8 Version Date Title Generalized Linear Models with Clustering

Package glmmml. R topics documented: March 25, Encoding UTF-8 Version Date Title Generalized Linear Models with Clustering Encoding UTF-8 Version 1.0.3 Date 2018-03-25 Title Generalized Linear Models with Clustering Package glmmml March 25, 2018 Binomial and Poisson regression for clustered data, fixed and random effects with

More information

Machine Learning A W 1sst KU. b) [1 P] For the probability distribution P (A, B, C, D) with the factorization

Machine Learning A W 1sst KU. b) [1 P] For the probability distribution P (A, B, C, D) with the factorization Machine Learning A 708.064 13W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence a) [1 P] For the probability distribution P (A, B, C, D) with the factorization P (A, B,

More information

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant

More information

CS Introduction to Data Mining Instructor: Abdullah Mueen

CS Introduction to Data Mining Instructor: Abdullah Mueen CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen LECTURE 8: ADVANCED CLUSTERING (FUZZY AND CO -CLUSTERING) Review: Basic Cluster Analysis Methods (Chap. 10) Cluster Analysis: Basic Concepts

More information

Expectation-Maximization Methods in Population Analysis. Robert J. Bauer, Ph.D. ICON plc.

Expectation-Maximization Methods in Population Analysis. Robert J. Bauer, Ph.D. ICON plc. Expectation-Maximization Methods in Population Analysis Robert J. Bauer, Ph.D. ICON plc. 1 Objective The objective of this tutorial is to briefly describe the statistical basis of Expectation-Maximization

More information

MCMC Diagnostics. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) MCMC Diagnostics MATH / 24

MCMC Diagnostics. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) MCMC Diagnostics MATH / 24 MCMC Diagnostics Yingbo Li Clemson University MATH 9810 Yingbo Li (Clemson) MCMC Diagnostics MATH 9810 1 / 24 Convergence to Posterior Distribution Theory proves that if a Gibbs sampler iterates enough,

More information

Missing Data Analysis for the Employee Dataset

Missing Data Analysis for the Employee Dataset Missing Data Analysis for the Employee Dataset 67% of the observations have missing values! Modeling Setup Random Variables: Y i =(Y i1,...,y ip ) 0 =(Y i,obs, Y i,miss ) 0 R i =(R i1,...,r ip ) 0 ( 1

More information

Unified Methods for Censored Longitudinal Data and Causality

Unified Methods for Censored Longitudinal Data and Causality Mark J. van der Laan James M. Robins Unified Methods for Censored Longitudinal Data and Causality Springer Preface v Notation 1 1 Introduction 8 1.1 Motivation, Bibliographic History, and an Overview of

More information

Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010

Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010 Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010 Tim Hanson, Ph.D. University of South Carolina T. Hanson (USC) Stat 704: Data Analysis I, Fall 2010 1 / 26 Additive predictors

More information

Statistics (STAT) Statistics (STAT) 1. Prerequisites: grade in C- or higher in STAT 1200 or STAT 1300 or STAT 1400

Statistics (STAT) Statistics (STAT) 1. Prerequisites: grade in C- or higher in STAT 1200 or STAT 1300 or STAT 1400 Statistics (STAT) 1 Statistics (STAT) STAT 1200: Introductory Statistical Reasoning Statistical concepts for critically evaluation quantitative information. Descriptive statistics, probability, estimation,

More information

Bayesian Computation: Overview and Methods for Low-Dimensional Models

Bayesian Computation: Overview and Methods for Low-Dimensional Models Bayesian Computation: Overview and Methods for Low-Dimensional Models Tom Loredo Dept. of Astronomy, Cornell University http://www.astro.cornell.edu/staff/loredo/bayes/ IAC Winter School, 3 4 Nov 2014

More information

Inference and Representation

Inference and Representation Inference and Representation Rachel Hodos New York University Lecture 5, October 6, 2015 Rachel Hodos Lecture 5: Inference and Representation Today: Learning with hidden variables Outline: Unsupervised

More information

K-Means and Gaussian Mixture Models

K-Means and Gaussian Mixture Models K-Means and Gaussian Mixture Models David Rosenberg New York University June 15, 2015 David Rosenberg (New York University) DS-GA 1003 June 15, 2015 1 / 43 K-Means Clustering Example: Old Faithful Geyser

More information

Acknowledgments. Acronyms

Acknowledgments. Acronyms Acknowledgments Preface Acronyms xi xiii xv 1 Basic Tools 1 1.1 Goals of inference 1 1.1.1 Population or process? 1 1.1.2 Probability samples 2 1.1.3 Sampling weights 3 1.1.4 Design effects. 5 1.2 An introduction

More information

Package ICsurv. February 19, 2015

Package ICsurv. February 19, 2015 Package ICsurv February 19, 2015 Type Package Title A package for semiparametric regression analysis of interval-censored data Version 1.0 Date 2014-6-9 Author Christopher S. McMahan and Lianming Wang

More information

Multiple imputation using chained equations: Issues and guidance for practice

Multiple imputation using chained equations: Issues and guidance for practice Multiple imputation using chained equations: Issues and guidance for practice Ian R. White, Patrick Royston and Angela M. Wood http://onlinelibrary.wiley.com/doi/10.1002/sim.4067/full By Gabrielle Simoneau

More information

Preface to the Second Edition. Preface to the First Edition. 1 Introduction 1

Preface to the Second Edition. Preface to the First Edition. 1 Introduction 1 Preface to the Second Edition Preface to the First Edition vii xi 1 Introduction 1 2 Overview of Supervised Learning 9 2.1 Introduction... 9 2.2 Variable Types and Terminology... 9 2.3 Two Simple Approaches

More information

An introduction to SPSS

An introduction to SPSS An introduction to SPSS To open the SPSS software using U of Iowa Virtual Desktop... Go to https://virtualdesktop.uiowa.edu and choose SPSS 24. Contents NOTE: Save data files in a drive that is accessible

More information

Structured Learning. Jun Zhu

Structured Learning. Jun Zhu Structured Learning Jun Zhu Supervised learning Given a set of I.I.D. training samples Learn a prediction function b r a c e Supervised learning (cont d) Many different choices Logistic Regression Maximum

More information

Bayesian Estimation for Skew Normal Distributions Using Data Augmentation

Bayesian Estimation for Skew Normal Distributions Using Data Augmentation The Korean Communications in Statistics Vol. 12 No. 2, 2005 pp. 323-333 Bayesian Estimation for Skew Normal Distributions Using Data Augmentation Hea-Jung Kim 1) Abstract In this paper, we develop a MCMC

More information

Hierarchical Modelling for Large Spatial Datasets

Hierarchical Modelling for Large Spatial Datasets Hierarchical Modelling for Large Spatial Datasets Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics,

More information

The glmmml Package. August 20, 2006

The glmmml Package. August 20, 2006 The glmmml Package August 20, 2006 Version 0.65-1 Date 2006/08/20 Title Generalized linear models with clustering A Maximum Likelihood and bootstrap approach to mixed models. License GPL version 2 or newer.

More information

SGN (4 cr) Chapter 11

SGN (4 cr) Chapter 11 SGN-41006 (4 cr) Chapter 11 Clustering Jussi Tohka & Jari Niemi Department of Signal Processing Tampere University of Technology February 25, 2014 J. Tohka & J. Niemi (TUT-SGN) SGN-41006 (4 cr) Chapter

More information

Bayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19

Bayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19 Bayesian Methods David Rosenberg New York University April 11, 2017 David Rosenberg (New York University) DS-GA 1003 April 11, 2017 1 / 19 Classical Statistics Classical Statistics David Rosenberg (New

More information

GAMs, GAMMs and other penalized GLMs using mgcv in R. Simon Wood Mathematical Sciences, University of Bath, U.K.

GAMs, GAMMs and other penalized GLMs using mgcv in R. Simon Wood Mathematical Sciences, University of Bath, U.K. GAMs, GAMMs and other penalied GLMs using mgcv in R Simon Wood Mathematical Sciences, University of Bath, U.K. Simple eample Consider a very simple dataset relating the timber volume of cherry trees to

More information

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov ECE521: Week 11, Lecture 20 27 March 2017: HMM learning/inference With thanks to Russ Salakhutdinov Examples of other perspectives Murphy 17.4 End of Russell & Norvig 15.2 (Artificial Intelligence: A Modern

More information

Summary: A Tutorial on Learning With Bayesian Networks

Summary: A Tutorial on Learning With Bayesian Networks Summary: A Tutorial on Learning With Bayesian Networks Markus Kalisch May 5, 2006 We primarily summarize [4]. When we think that it is appropriate, we comment on additional facts and more recent developments.

More information

Package mcemglm. November 29, 2015

Package mcemglm. November 29, 2015 Type Package Package mcemglm November 29, 2015 Title Maximum Likelihood Estimation for Generalized Linear Mixed Models Version 1.1 Date 2015-11-28 Author Felipe Acosta Archila Maintainer Maximum likelihood

More information

Scalable Multidimensional Hierarchical Bayesian Modeling on Spark

Scalable Multidimensional Hierarchical Bayesian Modeling on Spark Scalable Multidimensional Hierarchical Bayesian Modeling on Spark Robert Ormandi, Hongxia Yang and Quan Lu Yahoo! Sunnyvale, CA 2015 Click-Through-Rate (CTR) Prediction Estimating the probability of click

More information

Discussion on Bayesian Model Selection and Parameter Estimation in Extragalactic Astronomy by Martin Weinberg

Discussion on Bayesian Model Selection and Parameter Estimation in Extragalactic Astronomy by Martin Weinberg Discussion on Bayesian Model Selection and Parameter Estimation in Extragalactic Astronomy by Martin Weinberg Phil Gregory Physics and Astronomy Univ. of British Columbia Introduction Martin Weinberg reported

More information

A Bayesian General Linear Modeling Approach to Cortical Surface fmri Data Analysis

A Bayesian General Linear Modeling Approach to Cortical Surface fmri Data Analysis A Bayesian General Linear Modeling Approach to Cortical Surface fmri Data Analysis Amanda F. Mejia a, Yu Ryan Yue b, David Bolin c, Finn Lindgren d and Martin A. Lindquist e arxiv:1706.00959v1 [stat.ap]

More information

3D Human Motion Analysis and Manifolds

3D Human Motion Analysis and Manifolds D E P A R T M E N T O F C O M P U T E R S C I E N C E U N I V E R S I T Y O F C O P E N H A G E N 3D Human Motion Analysis and Manifolds Kim Steenstrup Pedersen DIKU Image group and E-Science center Motivation

More information

Linear Modeling with Bayesian Statistics

Linear Modeling with Bayesian Statistics Linear Modeling with Bayesian Statistics Bayesian Approach I I I I I Estimate probability of a parameter State degree of believe in specific parameter values Evaluate probability of hypothesis given the

More information

low bias high variance high bias low variance error test set training set high low Model Complexity Typical Behaviour Lecture 11:

low bias high variance high bias low variance error test set training set high low Model Complexity Typical Behaviour Lecture 11: Lecture 11: Overfitting and Capacity Control high bias low variance Typical Behaviour low bias high variance Sam Roweis error test set training set November 23, 4 low Model Complexity high Generalization,

More information

Introduction to Mobile Robotics

Introduction to Mobile Robotics Introduction to Mobile Robotics Gaussian Processes Wolfram Burgard Cyrill Stachniss Giorgio Grisetti Maren Bennewitz Christian Plagemann SS08, University of Freiburg, Department for Computer Science Announcement

More information

Splines and penalized regression

Splines and penalized regression Splines and penalized regression November 23 Introduction We are discussing ways to estimate the regression function f, where E(y x) = f(x) One approach is of course to assume that f has a certain shape,

More information

Wildfire Chances and Probabilistic Risk Assessment

Wildfire Chances and Probabilistic Risk Assessment Wildfire Chances and Probabilistic Risk Assessment D R Brillinger, H K Preisler and H M Naderi Statistics Department University of California Berkeley, CA, 97-386 Pacific Southwest Research Station USDA

More information

Issues in MCMC use for Bayesian model fitting. Practical Considerations for WinBUGS Users

Issues in MCMC use for Bayesian model fitting. Practical Considerations for WinBUGS Users Practical Considerations for WinBUGS Users Kate Cowles, Ph.D. Department of Statistics and Actuarial Science University of Iowa 22S:138 Lecture 12 Oct. 3, 2003 Issues in MCMC use for Bayesian model fitting

More information

Machine Learning / Jan 27, 2010

Machine Learning / Jan 27, 2010 Revisiting Logistic Regression & Naïve Bayes Aarti Singh Machine Learning 10-701/15-781 Jan 27, 2010 Generative and Discriminative Classifiers Training classifiers involves learning a mapping f: X -> Y,

More information

Spatial Statistics using R-INLA and Gaussian Markov random fields

Spatial Statistics using R-INLA and Gaussian Markov random fields Spatial Statistics using R-INLA and Gaussian Markov random fields David Bolin and Johan Lindström 1 Introduction In this lab we will look at an example of how to use the SPDE models in the INLA package.

More information

Infectious Disease Models. Angie Dobbs. A Thesis. Presented to. The University of Guelph. In partial fulfilment of requirements.

Infectious Disease Models. Angie Dobbs. A Thesis. Presented to. The University of Guelph. In partial fulfilment of requirements. Issues of Computational Efficiency and Model Approximation for Individual-Level Infectious Disease Models by Angie Dobbs A Thesis Presented to The University of Guelph In partial fulfilment of requirements

More information

Missing Data Analysis for the Employee Dataset

Missing Data Analysis for the Employee Dataset Missing Data Analysis for the Employee Dataset 67% of the observations have missing values! Modeling Setup For our analysis goals we would like to do: Y X N (X, 2 I) and then interpret the coefficients

More information

Regression. Dr. G. Bharadwaja Kumar VIT Chennai

Regression. Dr. G. Bharadwaja Kumar VIT Chennai Regression Dr. G. Bharadwaja Kumar VIT Chennai Introduction Statistical models normally specify how one set of variables, called dependent variables, functionally depend on another set of variables, called

More information

Spatial Interpolation & Geostatistics

Spatial Interpolation & Geostatistics (Z i Z j ) 2 / 2 Spatial Interpolation & Geostatistics Lag Lag Mean Distance between pairs of points 1 Tobler s Law All places are related, but nearby places are related more than distant places Corollary:

More information

ECE521 Lecture 18 Graphical Models Hidden Markov Models

ECE521 Lecture 18 Graphical Models Hidden Markov Models ECE521 Lecture 18 Graphical Models Hidden Markov Models Outline Graphical models Conditional independence Conditional independence after marginalization Sequence models hidden Markov models 2 Graphical

More information

Computer vision: models, learning and inference. Chapter 10 Graphical Models

Computer vision: models, learning and inference. Chapter 10 Graphical Models Computer vision: models, learning and inference Chapter 10 Graphical Models Independence Two variables x 1 and x 2 are independent if their joint probability distribution factorizes as Pr(x 1, x 2 )=Pr(x

More information