Optimal designs for comparing curves

Size: px
Start display at page:

Download "Optimal designs for comparing curves"

Transcription

1 Optimal designs for comparing curves Holger Dette, Ruhr-Universität Bochum Maria Konstantinou, Ruhr-Universität Bochum Kirsten Schorning, Ruhr-Universität Bochum FP7 HEALTH

2 Outline 1 Motivation 2 Testing for similarity 3 Optimal designs for comparing curves 4 Robust designs 5 Conclusions

3 Holger Dette Optimal designs for comparing curves 1 / 36 Motivation: four examples Populations of different geographic regions may bear differences in efficacy (or safety) dose response Objective: Ability to extrapolate study results Demonstrating equivalence of curves becomes an issue Comparison of dose response relationships for two regimens For example, demonstrate that once-daily and twice-daily applications of a drug are similar over a given dose range

4 Holger Dette Optimal designs for comparing curves 1 / 36 Motivation: four examples Populations of different geographic regions may bear differences in efficacy (or safety) dose response Objective: Ability to extrapolate study results Demonstrating equivalence of curves becomes an issue Comparison of dose response relationships for two regimens For example, demonstrate that once-daily and twice-daily applications of a drug are similar over a given dose range

5 Holger Dette Optimal designs for comparing curves 2 / 36 Motivation: four examples Comparison of different drugs containing the same active substance in order to claim bioequivalence. Traditional approaches based on AUC or Cmax may be misleading. Example: Both curves have the same AUC and Cmax It might be better to compare the curves directly

6 Holger Dette Optimal designs for comparing curves 2 / 36 Motivation: four examples Comparison of different drugs containing the same active substance in order to claim bioequivalence. Traditional approaches based on AUC or Cmax may be misleading. Example: Both curves have the same AUC and Cmax It might be better to compare the curves directly

7 Holger Dette Optimal designs for comparing curves 3 / 36 Motivation: four examples Comparison of dissolution profiles (correlated observations) In vitro dissolution profile comparison of two formulations (test vs. reference product) in order to demonstrate bioequivalence Figure: twelve tablets per product, each measured at six time points % Dissolved Time (min)

8 Comparing curves Two dose response curves from different populations (e.g. European, Japanese) Are the two curves similar? If they are: extrapolating information from one population to the other might be possible Dose Response Dose Response Holger Dette Optimal designs for comparing curves 4 / 36

9 Holger Dette Optimal designs for comparing curves 5 / 36 Comparing curves Two dose response models (from two samples) Y ijk = m i (x ij, ϑ i ) + ε ijk ; i = 1, 2; j = 1,..., l i ; k = 1,..., n ij, ε ijk independent N (0, σi 2 ) (i = 1, 2) x ij X (dose levels) X : design space (dose range) ϑ i R d i (parameter in model m i )

10 Holger Dette Optimal designs for comparing curves 6 / 36 Problem of similarity: Problem: Are the dose response curves m 1 and m 2 similar? If they are: Information from one population can be transferred to the other! We are particularly interested in optimal designs for a test for similarity (in order to minimize the sample size) IDEAL - Integrated DEsign and AnaLysis of small population group trials (EU-Project 7th Framework Program)

11 Holger Dette Optimal designs for comparing curves 7 / 36 Measures of similarity d: a metric measuring the distance between m 1 and m 2. Hypothesis of similarity: H 0 : d = d(m 1, m 2 ) versus H 1 : d = d(m 1, m 2 ) < (here is a pre-specified constant). In this talk: maximum difference d = d (m 1, m 2 ) = max x X m 1(x, ϑ 1 ) m 2 (x, ϑ 2 )

12 Holger Dette Optimal designs for comparing curves 7 / 36 Measures of similarity d: a metric measuring the distance between m 1 and m 2. Hypothesis of similarity: H 0 : d = d(m 1, m 2 ) versus H 1 : d = d(m 1, m 2 ) < (here is a pre-specified constant). In this talk: maximum difference d = d (m 1, m 2 ) = max x X m 1(x, ϑ 1 ) m 2 (x, ϑ 2 )

13 Holger Dette Optimal designs for comparing curves 8 / 36 Example Emax and Loglinear model (X = [0, 1]) m 1 (x, ϑ 1 ) = x x, m 2(x, ϑ 2 ) = log(x+0.2) dose effect EMAX model loglinear model maximum distance dose range Maximum deviation between Emax and loglinear model

14 Holger Dette Optimal designs for comparing curves 9 / 36 Testing for similarity (Gsteiger, Bretz, and Liu, 2011) Use the maximum distance d = d (m 1, m 2 ) = max x X m 1(x, ϑ 1 ) m 2 (x, ϑ 2 ) Construct a confidence band for the difference m 1 (, ϑ 1 ) m 2 (, ϑ 2 ) Reject and decide for similarity H 0 : d H 1 : d < if the confidence band is contained in the rectangle X [, ]

15 Holger Dette Optimal designs for comparing curves 9 / 36 Testing for similarity (Gsteiger, Bretz, and Liu, 2011) Use the maximum distance d = d (m 1, m 2 ) = max x X m 1(x, ϑ 1 ) m 2 (x, ϑ 2 ) Construct a confidence band for the difference m 1 (, ϑ 1 ) m 2 (, ϑ 2 ) Reject and decide for similarity H 0 : d H 1 : d < if the confidence band is contained in the rectangle X [, ]

16 Holger Dette Optimal designs for comparing curves 9 / 36 Testing for similarity (Gsteiger, Bretz, and Liu, 2011) Use the maximum distance d = d (m 1, m 2 ) = max x X m 1(x, ϑ 1 ) m 2 (x, ϑ 2 ) Construct a confidence band for the difference m 1 (, ϑ 1 ) m 2 (, ϑ 2 ) Reject and decide for similarity H 0 : d H 1 : d < if the confidence band is contained in the rectangle X [, ]

17 Holger Dette Optimal designs for comparing curves 10 / 36 Testing for similarity - example H 0 : d = 1 versus H 1 : d < = 1 Difference = 1 Difference = Design Space Design Space H 0 : is rejected! H 0 : is not rejected!

18 Holger Dette Optimal designs for comparing curves 11 / 36 Improvements Optimal designs In this talk we will improve this test: Efficient designs for the construction of confidence bands Dette, H., Schorning, K. Optimal designs for comparing curves (Annals of Statistics, 2016) Idea: Minimize the width of the confidence band by the construction of good designs (increase in power) This is a non standard optimal design problem!

19 Holger Dette Optimal designs for comparing curves 11 / 36 Improvements Optimal designs In this talk we will improve this test: Efficient designs for the construction of confidence bands Dette, H., Schorning, K. Optimal designs for comparing curves (Annals of Statistics, 2016) Idea: Minimize the width of the confidence band by the construction of good designs (increase in power) This is a non standard optimal design problem!

20 Holger Dette Optimal designs for comparing curves 12 / 36 Testing for similarity - more details More details for Gsteiger, Bretz, Liu (2011): Two designs for two samples (of sizes n 1 and n 2 ) ( x11 x 1l1 ξ 1 = n 11 n 1 n 1l1 n 1 ) ( x21 x 2l2 ξ 2 = n 21 n 2 n 2l2 n 2 ) Note: The design for the ith sample is called ξ i (i = 1, 2) n i is the sample size of sample i (i = 1, 2) l i is the number of different dose levels in sample i x ij is the jth dose level in the ith sample (j = 1,..., l i ) n ij is the number of patients treated assigned to dose level x ij n ij is the relative proportion of patients assigned to dose level n i x ij

21 Holger Dette Optimal designs for comparing curves 13 / 36 Testing for similarity - more details ˆϑ i : maximum likelihood estimate of ϑ i (i = 1, 2) where (i = 1, 2) ˆϑ i M i (ξ i, ϑ i ) = l i n ij j=1 a N (ϑi, σ2 i n i M 1 i (ξ i, ϑ i )) n i ( )( ϑ i m i (x ij, ϑ i ) ϑ i m i (x ij, ϑ i ) is the information matrix and depends on the design ξ i for the ith sample Note: Other estimates could be considered as well ) T

22 Holger Dette Optimal designs for comparing curves 13 / 36 Testing for similarity - more details ˆϑ i : maximum likelihood estimate of ϑ i (i = 1, 2) where (i = 1, 2) ˆϑ i M i (ξ i, ϑ i ) = l i n ij j=1 a N (ϑi, σ2 i n i M 1 i (ξ i, ϑ i )) n i ( )( ϑ i m i (x ij, ϑ i ) ϑ i m i (x ij, ϑ i ) is the information matrix and depends on the design ξ i for the ith sample Note: Other estimates could be considered as well ) T

23 Holger Dette Optimal designs for comparing curves 14 / 36 Testing for similarity - more details Confidence band { x X where m 1(x, ˆϑ 1 ) m 2 (x, ˆϑ 2 ) (m 1 (x, ϑ 1 ) m 2 (x, ϑ 2 )) } D ˆγ(x, ξ 1, ξ 2 ) ˆγ 2 (x, ξ 1, ξ 2 ) = ˆσ2 1 n 1 x m 1(x, ˆϑ 1 )M 1 1 (ξ 1, ˆϑ 1 ) ( x m 1(x, ˆϑ 1 ) ) T + ˆσ2 2 n 2 x m 2(x, ˆϑ 2 )M 1 2 (ξ 2, ˆϑ 2 ) ( x m 2(x, ˆϑ 2 ) ) T is an estimate of the variance of the predicted difference Critical value D is determined by parametric bootstrap Note: the width of the band at the point x depends on the designs ξ 1 and ξ 2 and is approximately given by 2Dˆγ(x, ξ 1, ξ 2 )

24 Holger Dette Optimal designs for comparing curves 15 / 36 Example Emax and loglinear model (X = [0, 1]) m 1(x, ϑ 1) = x, m2(x, ϑ2) = log(x + 0.2) x Hypotheses d 2.5 versus H 1 : d < 2.5 dose effect EMAX model loglinear model maximum distance dose range Emax and loglinear model.

25 Holger Dette Optimal designs for comparing curves 16 / 36 Example Simulated confidence bands (n 1 = n 2 = 100, σ 1 = σ 2 = 1.478) using standard designs ( ) ξ 1 = ξ 2 = Difference Design Space Confidence bands for the difference of the Emax and loglinear model. The test is very conservative! Can we do better by choosing a nice design?

26 Holger Dette Optimal designs for comparing curves 17 / 36 Approximate design theory Two designs ( x11 x 1l1 ξ 1 = n 11 n 1 n 1l1 n 1 ) ( x21 x 2l2 ξ 2 = n 21 n 2 n 2l2 n 2 ) Approximate design theory: replace n ij n i by w ij (0, 1), Approximate designs ξ 1 = ( x11 x 1l1 w 11 w 1l1 ) n 1 n 1 + n 2 by λ (0, 1) ξ 2 = ( x21 x 2l2 w 21 w 2l2 )

27 Holger Dette Optimal designs for comparing curves 18 / 36 Approximate design theory Note: two designs ξ 1, ξ 2 have to be determined (λ is fixed) Recall: the width of the band at the point x is proportional to γ(x, ξ 1, ξ 2 ) = γ(x, ϑ 1, ϑ 2, ξ 1, ξ 2 ) An optimal pair of designs minimizes the maximum width of the confidence band for the difference m 1 (x, ϑ 1 ) m 2 (x, ϑ 2 )! More mathematically: an optimal pair of designs (ξ 1, ξ 2 ) minimizes where γ 2 (ξ 1, ξ 2 ) = max x X γ2 (x, ξ 1, ξ 2 ) γ 2 (x, ξ 1, ξ 2 ) = σ2 1 λ 1 ϑ 1 m 1 (x, ϑ 1 )M 1 1 (ξ 1, ϑ 1 ) ( ϑ 2 m 1 (x, ϑ 1 ) ) T + σ2 2 λ 2 ϑ 2 m 2 (x, ϑ 2 )M 1 2 (ξ 2, ϑ 2 ) ( ϑ 2 m 2 (x, ϑ 2 ) ) T

28 Holger Dette Optimal designs for comparing curves 18 / 36 Approximate design theory Note: two designs ξ 1, ξ 2 have to be determined (λ is fixed) Recall: the width of the band at the point x is proportional to γ(x, ξ 1, ξ 2 ) = γ(x, ϑ 1, ϑ 2, ξ 1, ξ 2 ) An optimal pair of designs minimizes the maximum width of the confidence band for the difference m 1 (x, ϑ 1 ) m 2 (x, ϑ 2 )! More mathematically: an optimal pair of designs (ξ 1, ξ 2 ) minimizes where γ 2 (ξ 1, ξ 2 ) = max x X γ2 (x, ξ 1, ξ 2 ) γ 2 (x, ξ 1, ξ 2 ) = σ2 1 λ 1 ϑ 1 m 1 (x, ϑ 1 )M 1 1 (ξ 1, ϑ 1 ) ( ϑ 2 m 1 (x, ϑ 1 ) ) T + σ2 2 λ 2 ϑ 2 m 2 (x, ϑ 2 )M 1 2 (ξ 2, ϑ 2 ) ( ϑ 2 m 2 (x, ϑ 2 ) ) T

29 Holger Dette Optimal designs for comparing curves 19 / 36 Approximate design theory Note: we have to find l 1, l 2 (number of different dose levels) x ij (dose levels) w ij (proportions of patients allocated to each dose level) Optimal designs depend on parameters ϑ 1, ϑ 2 In this talk: we mainly assume some prior knowledge for the parameters and determine (locally optimal designs) Alternative designs can be constructed by standard methodology Bayesian optimal designs (later) (standardized) minimax designs Adaptive designs

30 Holger Dette Optimal designs for comparing curves 19 / 36 Approximate design theory Note: we have to find l 1, l 2 (number of different dose levels) x ij (dose levels) w ij (proportions of patients allocated to each dose level) Optimal designs depend on parameters ϑ 1, ϑ 2 In this talk: we mainly assume some prior knowledge for the parameters and determine (locally optimal designs) Alternative designs can be constructed by standard methodology Bayesian optimal designs (later) (standardized) minimax designs Adaptive designs

31 Holger Dette Optimal designs for comparing curves 19 / 36 Approximate design theory Note: we have to find l 1, l 2 (number of different dose levels) x ij (dose levels) w ij (proportions of patients allocated to each dose level) Optimal designs depend on parameters ϑ 1, ϑ 2 In this talk: we mainly assume some prior knowledge for the parameters and determine (locally optimal designs) Alternative designs can be constructed by standard methodology Bayesian optimal designs (later) (standardized) minimax designs Adaptive designs

32 Holger Dette Optimal designs for comparing curves 20 / 36 Approximate design theory Locally optimal designs have to be found numerically - Particle Swarm Optimization (PSO - metaheuristic) Check the optimality of the calculated designs Theorem (equivalence theorem) (ξ1, ξ 2 ) is optimal if and only if there exists a measure µ supported on the set {z X γ 2 (z, ξ 1, ξ 2) = max x X γ2 (x, ξ 1, ξ 2)} such that for all x 1, x 2 X L(x 1, x 2 ) := h(x 1, x 2, ξ1, ξ2, z)µ (dz) 0 X

33 Holger Dette Optimal designs for comparing curves 20 / 36 Approximate design theory Locally optimal designs have to be found numerically - Particle Swarm Optimization (PSO - metaheuristic) Check the optimality of the calculated designs Theorem (equivalence theorem) (ξ1, ξ 2 ) is optimal if and only if there exists a measure µ supported on the set {z X γ 2 (z, ξ 1, ξ 2) = max x X γ2 (x, ξ 1, ξ 2)} such that for all x 1, x 2 X L(x 1, x 2 ) := h(x 1, x 2, ξ1, ξ2, z)µ (dz) 0 X

34 Holger Dette Optimal designs for comparing curves 21 / 36 Approximate design theory The function h is complicated, but known: h(x 1, x 2, ξ1, ξ2 (, z) = m 1 (x 1, ϑ 1 )M 1 1 (ξ1, ϑ 1 ) m 1 (z, ϑ 2 ) ϑ 1 ϑ 1 ( + m 2 (x 2, ϑ 2 )M 1 2 (ξ2, ϑ 2 ) m 2 (z, ϑ 2 ) ϑ 2 ϑ 2 γ(ξ1, ξ2 ) µ is not known (the optimality criterion is not differentiable - solution of a minimax problem) ) )

35 Holger Dette Optimal designs for comparing curves 22 / 36 Example Emax and Log-linear model (X = [0, 1]) m 1 (x, ϑ 1 ) = x x, m 2(x, ϑ 2 ) = log(x+0.2) dose effect EMAX model loglinear model maximum distance dose range Emax and loglinear model.

36 Holger Dette Optimal designs for comparing curves 23 / 36 Optimal designs for Emax- and Log-linear model Emax- and Log-linear model m 1 (x, ϑ 1 ) = x, x [0, 1] x m 2 (x, ϑ 2 ) = log(x + 0.2), x [0, 1] d (m 1, m 2 ) = Numerically calculated optimal designs ( ) ( ξ = ξ = )

37 Holger Dette Optimal designs for comparing curves 24 / 36 Equivalence theorem Checking condition: L(x 1, x 2 ) = X h(x 1, x 2, ξ 1, ξ 2, z)µ (dz) Design space Design space The designs ξ 1 and ξ 2 are optimal

38 Holger Dette Optimal designs for comparing curves 25 / 36 Example (continued) Confidence bands for the difference of the Emax and loglinear model using a standard design (left panel) and the optimal design (right panel).

39 Holger Dette Optimal designs for comparing curves 26 / 36 Power of the test of Gsteiger et al (2011) Simulated power of (n 1 = n 2 = 100; 100 runs) of the test for the hypothesis H 0 : d versus H 1 : d < optimal design standard design

40 Holger Dette Optimal designs for comparing curves 27 / 36 Sensitivity Emax and Exponential model with assumed parameters Emax (m 1 ) ϑ 1 + ϑ2t t+ϑ 3 (0.2, 0.7, 0.2) exponential (m 2 ) ϑ 1 + ϑ 2 exp(t/ϑ 3 ) (0.183, 0.017, 0.28) Simulated confidence bands obtained from the standard design (dashed lines) and optimal design Difference Difference Design space Design space ϑ Emax = (0.1, 0.35, 0.1) ϑ Emax = (0.4, 0.7, 0.4) ϑ exp = (0.1, 0.05, 0.167) ϑ exp = (0.4, 0.2, 0.66)

41 Holger Dette Optimal designs for comparing curves 28 / 36 Sensitivity Simulated confidence bands obtained from the standard design (dashed lines) and optimal design Difference Difference Design space Design space ϑ Emax = (0.1, 0.35, 0.1) ϑ Emax = (0.4, 0.7, 0.4) ϑ exp = (0.1, 0.05, 0.167) ϑ exp = (0.4, 0.2, 0.66) An improvement can be obtained, even if the parameters are misspecified Can we do even better, incorporating the uncertainty about the parameters in the construction of the optimal designs?

42 Holger Dette Optimal designs for comparing curves 29 / 36 Bayesian optimal designs Recall: the width of the band at the point x is proportional to γ(x, ϑ 1, ϑ 2, ξ 1, ξ 2 ), where γ 2 (x, ϑ 1, ϑ 2, ξ 1, ξ 2 ) = σ2 1 n 1 x m 1(x, ϑ 1 )M 1 1 (ξ 1, ϑ 1 ) ( x m 1(x, ϑ 1 ) ) T + σ2 2 n 2 x m 2(x, ϑ 2 )M 1 2 (ξ 2, ϑ 2 ) ( x m 2(x, ϑ 2 ) ) T Local optimal designs: fix ϑ 1 and ϑ 2 and minimize max x X γ2 (x, ϑ 1, ϑ 2, ξ 1, ξ 2 ) Bayesian optimal designs: put prior distributions π 1 and π 2 on ϑ 1 and ϑ 2, respectively, and minimize max x X γ2 (x, ϑ 1, ϑ 2, ξ 1, ξ 2 )π 1 (dϑ 1 )π 2 (dϑ 2 )

43 Holger Dette Optimal designs for comparing curves 30 / 36 Bayesian optimal designs Usually the designs are sensitive with respect to the model assumption, e.g. An optimal design for comparing two Emax models might be inefficient, if the true models are an exponential and a loglinear model Model uncertainty can be addressed as well assume that m 1,..., m p are competing models assign to each pair (m i, m j ) a nonnegative weight α ij, which reflects the experimenters belief about the pair (m i, m j ) for group 1 and 2

44 Holger Dette Optimal designs for comparing curves 31 / 36 Bayesian optimal designs Model robust Bayesian optimal designs minimize p α ij i,j=1 max x X γ2 ij(x, ϑ i, ϑ j, ξ 1, ξ 2 )π i (dϑ i )π j (dϑ j ) where π 1,..., π p are prior distributions for the parameters ϑ 1,..., ϑ p in the models m 1,..., m p, respectively. Similar results are available An algorithm to calculate the model robust Bayesian optimal designs An equivalence theorem, to check the optimality of the numerical results

45 Holger Dette Optimal designs for comparing curves 32 / 36 Model robust Bayesian optimal designs - an example Commonly used dose response models model m(t, ϑ) parameters Emax (m 1 ) ϑ 1 + ϑ2t t+ϑ 3 (0.2, 0.7, 0.2) exponential (m 2 ) ϑ 1 + ϑ 2 exp(t/ϑ 3 ) (0.183, 0.017, 0.28) loginear (m 3 ) ϑ 1 + ϑ 2 log(t + ϑ 3 ) (0.74, 0.33, 0.2) We are interested in a comparison of two Emax or two exponential or two loglinear models, i.e α 11 = α 22 = α 33 = 1/3 and α ij = 0 else We fix parameters for each model (prior distribution concentrated in one point)

46 Holger Dette Optimal designs for comparing curves 33 / 36 Model robust Bayesian optimal designs - an example Model robust Bayesian optimal designs First sample: allocate 30.27%, 30.72% 19.56%, 19.44% of the patients to dose levels 0.00, 0.17, 0.77, 1.00 Second sample: allocate 29.96%, 30.18%, 19.38%, 20.48% at the points 0.00, 0.17, 0.76, 1.00 how good are the designs? model 1 / model 2 Emax / Emax exp / exp loglin / loglin standard design 67.98% 42.87% 74.29% robust optimal design 90.40% 78.45% 90.79%

47 Holger Dette Optimal designs for comparing curves 34 / 36 Model robust Bayesian optimal designs - an example Model robust Bayesian optimal designs First sample: allocate 30.27%, 30.72% 19.56%, 19.44% of the patients to dose levels 0.00, 0.17, 0.77, 1.00 Second sample: allocate 29.96%, 30.18%, 19.38%, 20.48% at the points 0.00, 0.17, 0.76, 1.00 how good are the designs if we use other models? model 1 / model 2 loglin/exp loglin/emax exp/emax standard design 58.85% 72.83% 59.00% D-optimal designs (Emax) 2.21% 93.81% 2.24% D-optimal designs (loglinear) 7.31% 92.44% 7.40% D-optimal designs (exponential) 15.08% 3.72% 4.29% robust optimal design 67.30% 89.65% 69.57%

48 Holger Dette Optimal designs for comparing curves 35 / 36 Conclusions The problem of testing for the similarity of two (parametric) curves is fundamental interest in biostatistics The power of the test of Gsteiger et al. (2011) can be increased by optimal designs which produce narrower confidence bands The designs can be constructed such that they are robust with respect to model misspecification with respect to misspecification of model parameters Standard designs are not efficient in this context, e.g. uniform design individual D-optimal designs The resulting procedure is substantially more powerful, but it is still not a good test!

49 Holger Dette Optimal designs for comparing curves 35 / 36 Conclusions The problem of testing for the similarity of two (parametric) curves is fundamental interest in biostatistics The power of the test of Gsteiger et al. (2011) can be increased by optimal designs which produce narrower confidence bands The designs can be constructed such that they are robust with respect to model misspecification with respect to misspecification of model parameters Standard designs are not efficient in this context, e.g. uniform design individual D-optimal designs The resulting procedure is substantially more powerful, but it is still not a good test!

50 Holger Dette Optimal designs for comparing curves 35 / 36 Conclusions The problem of testing for the similarity of two (parametric) curves is fundamental interest in biostatistics The power of the test of Gsteiger et al. (2011) can be increased by optimal designs which produce narrower confidence bands The designs can be constructed such that they are robust with respect to model misspecification with respect to misspecification of model parameters Standard designs are not efficient in this context, e.g. uniform design individual D-optimal designs The resulting procedure is substantially more powerful, but it is still not a good test!

51 Holger Dette Optimal designs for comparing curves 35 / 36 Conclusions The problem of testing for the similarity of two (parametric) curves is fundamental interest in biostatistics The power of the test of Gsteiger et al. (2011) can be increased by optimal designs which produce narrower confidence bands The designs can be constructed such that they are robust with respect to model misspecification with respect to misspecification of model parameters Standard designs are not efficient in this context, e.g. uniform design individual D-optimal designs The resulting procedure is substantially more powerful, but it is still not a good test!

52 Holger Dette Optimal designs for comparing curves 35 / 36 Conclusions The problem of testing for the similarity of two (parametric) curves is fundamental interest in biostatistics The power of the test of Gsteiger et al. (2011) can be increased by optimal designs which produce narrower confidence bands The designs can be constructed such that they are robust with respect to model misspecification with respect to misspecification of model parameters Standard designs are not efficient in this context, e.g. uniform design individual D-optimal designs The resulting procedure is substantially more powerful, but it is still not a good test!

53 Holger Dette Optimal designs for comparing curves 36 / 36 Power of the test of Gsteiger et al (2011) Simulated power of (n 1 = n 2 = 100; 100 runs) of the test for the hypothesis H 0 : d versus H 1 : d < optimal design standard design Note: d (m 1, m 2 ) = (rather extreme alternatives!) In the next talk a more efficient test will be constructed Optimal designs for this test can be constructed as well

54 Holger Dette Optimal designs for comparing curves 36 / 36 Power of the test of Gsteiger et al (2011) Simulated power of (n 1 = n 2 = 100; 100 runs) of the test for the hypothesis H 0 : d versus H 1 : d < optimal design standard design Note: d (m 1, m 2 ) = (rather extreme alternatives!) In the next talk a more efficient test will be constructed Optimal designs for this test can be constructed as well

Equivalence of dose-response curves

Equivalence of dose-response curves Equivalence of dose-response curves Holger Dette, Ruhr-Universität Bochum Kathrin Möllenhoff, Ruhr-Universität Bochum Stanislav Volgushev, University of Toronto Frank Bretz, Novartis Basel FP7 HEALTH 2013-602552

More information

Use of Extreme Value Statistics in Modeling Biometric Systems

Use of Extreme Value Statistics in Modeling Biometric Systems Use of Extreme Value Statistics in Modeling Biometric Systems Similarity Scores Two types of matching: Genuine sample Imposter sample Matching scores Enrolled sample 0.95 0.32 Probability Density Decision

More information

Optimal Two-Stage Adaptive Enrichment Designs, using Sparse Linear Programming

Optimal Two-Stage Adaptive Enrichment Designs, using Sparse Linear Programming Optimal Two-Stage Adaptive Enrichment Designs, using Sparse Linear Programming Michael Rosenblum Johns Hopkins Bloomberg School of Public Health Joint work with Xingyuan (Ethan) Fang and Han Liu Princeton

More information

Z-TEST / Z-STATISTIC: used to test hypotheses about. µ when the population standard deviation is unknown

Z-TEST / Z-STATISTIC: used to test hypotheses about. µ when the population standard deviation is unknown Z-TEST / Z-STATISTIC: used to test hypotheses about µ when the population standard deviation is known and population distribution is normal or sample size is large T-TEST / T-STATISTIC: used to test hypotheses

More information

Physics 736. Experimental Methods in Nuclear-, Particle-, and Astrophysics. - Statistical Methods -

Physics 736. Experimental Methods in Nuclear-, Particle-, and Astrophysics. - Statistical Methods - Physics 736 Experimental Methods in Nuclear-, Particle-, and Astrophysics - Statistical Methods - Karsten Heeger heeger@wisc.edu Course Schedule and Reading course website http://neutrino.physics.wisc.edu/teaching/phys736/

More information

Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis

Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis Xavier Le Faucheur a, Brani Vidakovic b and Allen Tannenbaum a a School of Electrical and Computer Engineering, b Department of Biomedical

More information

Monte Carlo for Spatial Models

Monte Carlo for Spatial Models Monte Carlo for Spatial Models Murali Haran Department of Statistics Penn State University Penn State Computational Science Lectures April 2007 Spatial Models Lots of scientific questions involve analyzing

More information

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization. Wolfram Burgard

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization. Wolfram Burgard Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Wolfram Burgard 1 Motivation Recall: Discrete filter Discretize the continuous state space High memory complexity

More information

Convexization in Markov Chain Monte Carlo

Convexization in Markov Chain Monte Carlo in Markov Chain Monte Carlo 1 IBM T. J. Watson Yorktown Heights, NY 2 Department of Aerospace Engineering Technion, Israel August 23, 2011 Problem Statement MCMC processes in general are governed by non

More information

Chapter 7: Dual Modeling in the Presence of Constant Variance

Chapter 7: Dual Modeling in the Presence of Constant Variance Chapter 7: Dual Modeling in the Presence of Constant Variance 7.A Introduction An underlying premise of regression analysis is that a given response variable changes systematically and smoothly due to

More information

Splines and penalized regression

Splines and penalized regression Splines and penalized regression November 23 Introduction We are discussing ways to estimate the regression function f, where E(y x) = f(x) One approach is of course to assume that f has a certain shape,

More information

Chapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea

Chapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea Chapter 3 Bootstrap 3.1 Introduction The estimation of parameters in probability distributions is a basic problem in statistics that one tends to encounter already during the very first course on the subject.

More information

Data Mining Classification: Bayesian Decision Theory

Data Mining Classification: Bayesian Decision Theory Data Mining Classification: Bayesian Decision Theory Lecture Notes for Chapter 2 R. O. Duda, P. E. Hart, and D. G. Stork, Pattern classification, 2nd ed. New York: Wiley, 2001. Lecture Notes for Chapter

More information

Statistics 202: Data Mining. c Jonathan Taylor. Outliers Based in part on slides from textbook, slides of Susan Holmes.

Statistics 202: Data Mining. c Jonathan Taylor. Outliers Based in part on slides from textbook, slides of Susan Holmes. Outliers Based in part on slides from textbook, slides of Susan Holmes December 2, 2012 1 / 1 Concepts What is an outlier? The set of data points that are considerably different than the remainder of the

More information

Lab 5 - Risk Analysis, Robustness, and Power

Lab 5 - Risk Analysis, Robustness, and Power Type equation here.biology 458 Biometry Lab 5 - Risk Analysis, Robustness, and Power I. Risk Analysis The process of statistical hypothesis testing involves estimating the probability of making errors

More information

Monte Carlo Methods and Statistical Computing: My Personal E

Monte Carlo Methods and Statistical Computing: My Personal E Monte Carlo Methods and Statistical Computing: My Personal Experience Department of Mathematics & Statistics Indian Institute of Technology Kanpur November 29, 2014 Outline Preface 1 Preface 2 3 4 5 6

More information

The Cross-Entropy Method for Mathematical Programming

The Cross-Entropy Method for Mathematical Programming The Cross-Entropy Method for Mathematical Programming Dirk P. Kroese Reuven Y. Rubinstein Department of Mathematics, The University of Queensland, Australia Faculty of Industrial Engineering and Management,

More information

Nested Sampling: Introduction and Implementation

Nested Sampling: Introduction and Implementation UNIVERSITY OF TEXAS AT SAN ANTONIO Nested Sampling: Introduction and Implementation Liang Jing May 2009 1 1 ABSTRACT Nested Sampling is a new technique to calculate the evidence, Z = P(D M) = p(d θ, M)p(θ

More information

Slides 11: Verification and Validation Models

Slides 11: Verification and Validation Models Slides 11: Verification and Validation Models Purpose and Overview The goal of the validation process is: To produce a model that represents true behaviour closely enough for decision making purposes.

More information

DS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University

DS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University DS 4400 Machine Learning and Data Mining I Alina Oprea Associate Professor, CCIS Northeastern University September 20 2018 Review Solution for multiple linear regression can be computed in closed form

More information

Network Traffic Measurements and Analysis

Network Traffic Measurements and Analysis DEIB - Politecnico di Milano Fall, 2017 Sources Hastie, Tibshirani, Friedman: The Elements of Statistical Learning James, Witten, Hastie, Tibshirani: An Introduction to Statistical Learning Andrew Ng:

More information

Automatic Selection of Compiler Options Using Non-parametric Inferential Statistics

Automatic Selection of Compiler Options Using Non-parametric Inferential Statistics Automatic Selection of Compiler Options Using Non-parametric Inferential Statistics Masayo Haneda Peter M.W. Knijnenburg Harry A.G. Wijshoff LIACS, Leiden University Motivation An optimal compiler optimization

More information

- with application to cluster and significance analysis

- with application to cluster and significance analysis Selection via - with application to cluster and significance of gene expression Rebecka Jörnsten Department of Statistics, Rutgers University rebecka@stat.rutgers.edu, http://www.stat.rutgers.edu/ rebecka

More information

Nonparametric Mixed-Effects Models for Longitudinal Data

Nonparametric Mixed-Effects Models for Longitudinal Data Nonparametric Mixed-Effects Models for Longitudinal Data Zhang Jin-Ting Dept of Stat & Appl Prob National University of Sinagpore University of Seoul, South Korea, 7 p.1/26 OUTLINE The Motivating Data

More information

Lecture 14. December 19, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Lecture 14. December 19, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University. geometric Lecture 14 Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University December 19, 2007 geometric 1 2 3 4 geometric 5 6 7 geometric 1 Review about logs

More information

MATH 2400, Analytic Geometry and Calculus 3

MATH 2400, Analytic Geometry and Calculus 3 MATH 2400, Analytic Geometry and Calculus 3 List of important Definitions and Theorems 1 Foundations Definition 1. By a function f one understands a mathematical object consisting of (i) a set X, called

More information

Splines. Patrick Breheny. November 20. Introduction Regression splines (parametric) Smoothing splines (nonparametric)

Splines. Patrick Breheny. November 20. Introduction Regression splines (parametric) Smoothing splines (nonparametric) Splines Patrick Breheny November 20 Patrick Breheny STA 621: Nonparametric Statistics 1/46 Introduction Introduction Problems with polynomial bases We are discussing ways to estimate the regression function

More information

Summer School in Statistics for Astronomers & Physicists June 15-17, Cluster Analysis

Summer School in Statistics for Astronomers & Physicists June 15-17, Cluster Analysis Summer School in Statistics for Astronomers & Physicists June 15-17, 2005 Session on Computational Algorithms for Astrostatistics Cluster Analysis Max Buot Department of Statistics Carnegie-Mellon University

More information

Chapter 3: Supervised Learning

Chapter 3: Supervised Learning Chapter 3: Supervised Learning Road Map Basic concepts Evaluation of classifiers Classification using association rules Naïve Bayesian classification Naïve Bayes for text classification Summary 2 An example

More information

Chapters 5-6: Statistical Inference Methods

Chapters 5-6: Statistical Inference Methods Chapters 5-6: Statistical Inference Methods Chapter 5: Estimation (of population parameters) Ex. Based on GSS data, we re 95% confident that the population mean of the variable LONELY (no. of days in past

More information

Nonparametric and Simulation-Based Tests. Stat OSU, Autumn 2018 Dalpiaz

Nonparametric and Simulation-Based Tests. Stat OSU, Autumn 2018 Dalpiaz Nonparametric and Simulation-Based Tests Stat 3202 @ OSU, Autumn 2018 Dalpiaz 1 What is Parametric Testing? 2 Warmup #1, Two Sample Test for p 1 p 2 Ohio Issue 1, the Drug and Criminal Justice Policies

More information

Segmentation: Clustering, Graph Cut and EM

Segmentation: Clustering, Graph Cut and EM Segmentation: Clustering, Graph Cut and EM Ying Wu Electrical Engineering and Computer Science Northwestern University, Evanston, IL 60208 yingwu@northwestern.edu http://www.eecs.northwestern.edu/~yingwu

More information

Lecture 10: SVM Lecture Overview Support Vector Machines The binary classification problem

Lecture 10: SVM Lecture Overview Support Vector Machines The binary classification problem Computational Learning Theory Fall Semester, 2012/13 Lecture 10: SVM Lecturer: Yishay Mansour Scribe: Gitit Kehat, Yogev Vaknin and Ezra Levin 1 10.1 Lecture Overview In this lecture we present in detail

More information

Lecture 27, April 24, Reading: See class website. Nonparametric regression and kernel smoothing. Structured sparse additive models (GroupSpAM)

Lecture 27, April 24, Reading: See class website. Nonparametric regression and kernel smoothing. Structured sparse additive models (GroupSpAM) School of Computer Science Probabilistic Graphical Models Structured Sparse Additive Models Junming Yin and Eric Xing Lecture 7, April 4, 013 Reading: See class website 1 Outline Nonparametric regression

More information

DATA MINING AND MACHINE LEARNING. Lecture 6: Data preprocessing and model selection Lecturer: Simone Scardapane

DATA MINING AND MACHINE LEARNING. Lecture 6: Data preprocessing and model selection Lecturer: Simone Scardapane DATA MINING AND MACHINE LEARNING Lecture 6: Data preprocessing and model selection Lecturer: Simone Scardapane Academic Year 2016/2017 Table of contents Data preprocessing Feature normalization Missing

More information

An Introduction to the Bootstrap

An Introduction to the Bootstrap An Introduction to the Bootstrap Bradley Efron Department of Statistics Stanford University and Robert J. Tibshirani Department of Preventative Medicine and Biostatistics and Department of Statistics,

More information

Equivalence Tests for Two Means in a 2x2 Cross-Over Design using Differences

Equivalence Tests for Two Means in a 2x2 Cross-Over Design using Differences Chapter 520 Equivalence Tests for Two Means in a 2x2 Cross-Over Design using Differences Introduction This procedure calculates power and sample size of statistical tests of equivalence of the means of

More information

Resources for statistical assistance. Quantitative covariates and regression analysis. Methods for predicting continuous outcomes.

Resources for statistical assistance. Quantitative covariates and regression analysis. Methods for predicting continuous outcomes. Resources for statistical assistance Quantitative covariates and regression analysis Carolyn Taylor Applied Statistics and Data Science Group (ASDa) Department of Statistics, UBC January 24, 2017 Department

More information

Chapter 6: Examples 6.A Introduction

Chapter 6: Examples 6.A Introduction Chapter 6: Examples 6.A Introduction In Chapter 4, several approaches to the dual model regression problem were described and Chapter 5 provided expressions enabling one to compute the MSE of the mean

More information

Optimal Techniques and Sequences for Population Modeling of Bioavailability

Optimal Techniques and Sequences for Population Modeling of Bioavailability Optimal Techniques and Sequences for Population Modeling of Bioavailability Roger W. Jelliffe, M.D Professor of Medicine Laboratory of Applied Pharmacokinetics, USC Keck School of Medicine www.usc.edu/hsc/lab_apk/,

More information

Predictive Analysis: Evaluation and Experimentation. Heejun Kim

Predictive Analysis: Evaluation and Experimentation. Heejun Kim Predictive Analysis: Evaluation and Experimentation Heejun Kim June 19, 2018 Evaluation and Experimentation Evaluation Metrics Cross-Validation Significance Tests Evaluation Predictive analysis: training

More information

Mixed Effects Models. Biljana Jonoska Stojkova Applied Statistics and Data Science Group (ASDa) Department of Statistics, UBC.

Mixed Effects Models. Biljana Jonoska Stojkova Applied Statistics and Data Science Group (ASDa) Department of Statistics, UBC. Mixed Effects Models Biljana Jonoska Stojkova Applied Statistics and Data Science Group (ASDa) Department of Statistics, UBC March 6, 2018 Resources for statistical assistance Department of Statistics

More information

Spatial Outlier Detection

Spatial Outlier Detection Spatial Outlier Detection Chang-Tien Lu Department of Computer Science Northern Virginia Center Virginia Tech Joint work with Dechang Chen, Yufeng Kou, Jiang Zhao 1 Spatial Outlier A spatial data point

More information

GAMs semi-parametric GLMs. Simon Wood Mathematical Sciences, University of Bath, U.K.

GAMs semi-parametric GLMs. Simon Wood Mathematical Sciences, University of Bath, U.K. GAMs semi-parametric GLMs Simon Wood Mathematical Sciences, University of Bath, U.K. Generalized linear models, GLM 1. A GLM models a univariate response, y i as g{e(y i )} = X i β where y i Exponential

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 13: The bootstrap (v3) Ramesh Johari ramesh.johari@stanford.edu 1 / 30 Resampling 2 / 30 Sampling distribution of a statistic For this lecture: There is a population model

More information

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization 6th WSEAS International Conference on SYSTEM SCIENCE and SIMULATION in ENGINEERING, Venice, Italy, November 21-23, 2007 18 Performance Evaluation of an Interior Point Filter Line Search Method for Constrained

More information

Part I. Hierarchical clustering. Hierarchical Clustering. Hierarchical clustering. Produces a set of nested clusters organized as a

Part I. Hierarchical clustering. Hierarchical Clustering. Hierarchical clustering. Produces a set of nested clusters organized as a Week 9 Based in part on slides from textbook, slides of Susan Holmes Part I December 2, 2012 Hierarchical Clustering 1 / 1 Produces a set of nested clusters organized as a Hierarchical hierarchical clustering

More information

Bootstrap Confidence Interval of the Difference Between Two Process Capability Indices

Bootstrap Confidence Interval of the Difference Between Two Process Capability Indices Int J Adv Manuf Technol (2003) 21:249 256 Ownership and Copyright 2003 Springer-Verlag London Limited Bootstrap Confidence Interval of the Difference Between Two Process Capability Indices J.-P. Chen 1

More information

Metrics for Performance Evaluation How to evaluate the performance of a model? Methods for Performance Evaluation How to obtain reliable estimates?

Metrics for Performance Evaluation How to evaluate the performance of a model? Methods for Performance Evaluation How to obtain reliable estimates? Model Evaluation Metrics for Performance Evaluation How to evaluate the performance of a model? Methods for Performance Evaluation How to obtain reliable estimates? Methods for Model Comparison How to

More information

Passive Differential Matched-field Depth Estimation of Moving Acoustic Sources

Passive Differential Matched-field Depth Estimation of Moving Acoustic Sources Lincoln Laboratory ASAP-2001 Workshop Passive Differential Matched-field Depth Estimation of Moving Acoustic Sources Shawn Kraut and Jeffrey Krolik Duke University Department of Electrical and Computer

More information

Outline. Bayesian Data Analysis Hierarchical models. Rat tumor data. Errandum: exercise GCSR 3.11

Outline. Bayesian Data Analysis Hierarchical models. Rat tumor data. Errandum: exercise GCSR 3.11 Outline Bayesian Data Analysis Hierarchical models Helle Sørensen May 15, 2009 Today: More about the rat tumor data: model, derivation of posteriors, the actual computations in R. : a hierarchical normal

More information

A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection (Kohavi, 1995)

A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection (Kohavi, 1995) A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection (Kohavi, 1995) Department of Information, Operations and Management Sciences Stern School of Business, NYU padamopo@stern.nyu.edu

More information

Introduction to hypothesis testing

Introduction to hypothesis testing Introduction to hypothesis testing Mark Johnson Macquarie University Sydney, Australia February 27, 2017 1 / 38 Outline Introduction Hypothesis tests and confidence intervals Classical hypothesis tests

More information

The Perils of Unfettered In-Sample Backtesting

The Perils of Unfettered In-Sample Backtesting The Perils of Unfettered In-Sample Backtesting Tyler Yeats June 8, 2015 Abstract When testing a financial investment strategy, it is common to use what is known as a backtest, or a simulation of how well

More information

Probabilistic Robotics

Probabilistic Robotics Probabilistic Robotics Discrete Filters and Particle Filters Models Some slides adopted from: Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Kai Arras and Probabilistic Robotics Book SA-1 Probabilistic

More information

Optimal Denoising of Natural Images and their Multiscale Geometry and Density

Optimal Denoising of Natural Images and their Multiscale Geometry and Density Optimal Denoising of Natural Images and their Multiscale Geometry and Density Department of Computer Science and Applied Mathematics Weizmann Institute of Science, Israel. Joint work with Anat Levin (WIS),

More information

Bootstrapping Methods

Bootstrapping Methods Bootstrapping Methods example of a Monte Carlo method these are one Monte Carlo statistical method some Bayesian statistical methods are Monte Carlo we can also simulate models using Monte Carlo methods

More information

Bootstrapping Method for 14 June 2016 R. Russell Rhinehart. Bootstrapping

Bootstrapping Method for  14 June 2016 R. Russell Rhinehart. Bootstrapping Bootstrapping Method for www.r3eda.com 14 June 2016 R. Russell Rhinehart Bootstrapping This is extracted from the book, Nonlinear Regression Modeling for Engineering Applications: Modeling, Model Validation,

More information

Simulation Supported POD Methodology and Validation for Automated Eddy Current Procedures

Simulation Supported POD Methodology and Validation for Automated Eddy Current Procedures 4th International Symposium on NDT in Aerospace 2012 - Th.1.A.1 Simulation Supported POD Methodology and Validation for Automated Eddy Current Procedures Anders ROSELL, Gert PERSSON Volvo Aero Corporation,

More information

The Optimal Discovery Procedure: A New Approach to Simultaneous Significance Testing

The Optimal Discovery Procedure: A New Approach to Simultaneous Significance Testing UW Biostatistics Working Paper Series 9-6-2005 The Optimal Discovery Procedure: A New Approach to Simultaneous Significance Testing John D. Storey University of Washington, jstorey@u.washington.edu Suggested

More information

In this computer exercise we will work with the analysis of variance in R. We ll take a look at the following topics:

In this computer exercise we will work with the analysis of variance in R. We ll take a look at the following topics: UPPSALA UNIVERSITY Department of Mathematics Måns Thulin, thulin@math.uu.se Analysis of regression and variance Fall 2011 COMPUTER EXERCISE 2: One-way ANOVA In this computer exercise we will work with

More information

Today. Lecture 4: Last time. The EM algorithm. We examine clustering in a little more detail; we went over it a somewhat quickly last time

Today. Lecture 4: Last time. The EM algorithm. We examine clustering in a little more detail; we went over it a somewhat quickly last time Today Lecture 4: We examine clustering in a little more detail; we went over it a somewhat quickly last time The CAD data will return and give us an opportunity to work with curves (!) We then examine

More information

New Results of Fully Bayesian

New Results of Fully Bayesian of Fully Bayesian UCI February 7, 2012 of Fully Bayesian Calibration Samples Principle Component Analysis Model Building Three source parameter sampling schemes Simulation Quasar data sets New data sets

More information

Data transformation in multivariate quality control

Data transformation in multivariate quality control Motto: Is it normal to have normal data? Data transformation in multivariate quality control J. Militký and M. Meloun The Technical University of Liberec Liberec, Czech Republic University of Pardubice

More information

Bayesian Sequential Sampling Policies and Sufficient Conditions for Convergence to a Global Optimum

Bayesian Sequential Sampling Policies and Sufficient Conditions for Convergence to a Global Optimum Bayesian Sequential Sampling Policies and Sufficient Conditions for Convergence to a Global Optimum Peter Frazier Warren Powell 2 Operations Research & Information Engineering, Cornell University 2 Operations

More information

Physics 736. Experimental Methods in Nuclear-, Particle-, and Astrophysics. Lecture 14

Physics 736. Experimental Methods in Nuclear-, Particle-, and Astrophysics. Lecture 14 Physics 736 Experimental Methods in Nuclear-, Particle-, and Astrophysics Lecture 14 Karsten Heeger heeger@wisc.edu Course Schedule and Reading course website http://neutrino.physics.wisc.edu/teaching/phys736/

More information

Image Quality Assessment Techniques: An Overview

Image Quality Assessment Techniques: An Overview Image Quality Assessment Techniques: An Overview Shruti Sonawane A. M. Deshpande Department of E&TC Department of E&TC TSSM s BSCOER, Pune, TSSM s BSCOER, Pune, Pune University, Maharashtra, India Pune

More information

Modeling and Performance Analysis with Discrete-Event Simulation

Modeling and Performance Analysis with Discrete-Event Simulation Simulation Modeling and Performance Analysis with Discrete-Event Simulation Chapter 10 Verification and Validation of Simulation Models Contents Model-Building, Verification, and Validation Verification

More information

Random projection for non-gaussian mixture models

Random projection for non-gaussian mixture models Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,

More information

Bagging for One-Class Learning

Bagging for One-Class Learning Bagging for One-Class Learning David Kamm December 13, 2008 1 Introduction Consider the following outlier detection problem: suppose you are given an unlabeled data set and make the assumptions that one

More information

Local Selection for Heuristic Algorithms as a Factor in Accelerating Optimum Search

Local Selection for Heuristic Algorithms as a Factor in Accelerating Optimum Search Local Selection for Heuristic Algorithms as a Factor in Accelerating Optimum Search Danuta Jama Institute of Mathematics Silesian University of Technology Kaszubska 23, 44-100 Gliwice, Poland Email: Danuta.Jama@polsl.pl

More information

Assessing the Quality of the Natural Cubic Spline Approximation

Assessing the Quality of the Natural Cubic Spline Approximation Assessing the Quality of the Natural Cubic Spline Approximation AHMET SEZER ANADOLU UNIVERSITY Department of Statisticss Yunus Emre Kampusu Eskisehir TURKEY ahsst12@yahoo.com Abstract: In large samples,

More information

Beyond the Assumption of Constant Hazard Rate in Estimating Incidence Rate on Current Status Data with Applications to Phase IV Cancer Trial

Beyond the Assumption of Constant Hazard Rate in Estimating Incidence Rate on Current Status Data with Applications to Phase IV Cancer Trial Beyond the Assumption of Constant Hazard Rate in Estimating Incidence Rate on Current Status Data with Applications to Phase IV Cancer Trial Deokumar Srivastava, Ph.D. Member Department of Biostatistics

More information

Analysis of guifi.net s Topology

Analysis of guifi.net s Topology Analysis of guifi.net s Topology Extension of Results Llorenç Cerdà-Alabern Universitat Politècnica de Catalunya, Computer Architecture Dep. Jordi Girona -, E-, Barcelona, Spain llorenc@ac.upc.edu May,

More information

Charged Particle Reconstruction in HIC Detectors

Charged Particle Reconstruction in HIC Detectors Charged Particle Reconstruction in HIC Detectors Ralf-Arno Tripolt, Qiyan Li [http://de.wikipedia.org/wiki/marienburg_(mosel)] H-QM Lecture Week on Introduction to Heavy Ion Physics Kloster Marienburg/Mosel,

More information

Support Vector Machines

Support Vector Machines Support Vector Machines Chapter 9 Chapter 9 1 / 50 1 91 Maximal margin classifier 2 92 Support vector classifiers 3 93 Support vector machines 4 94 SVMs with more than two classes 5 95 Relationshiop to

More information

BART STAT8810, Fall 2017

BART STAT8810, Fall 2017 BART STAT8810, Fall 2017 M.T. Pratola November 1, 2017 Today BART: Bayesian Additive Regression Trees BART: Bayesian Additive Regression Trees Additive model generalizes the single-tree regression model:

More information

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles INTERNATIONAL JOURNAL OF MATHEMATICS MODELS AND METHODS IN APPLIED SCIENCES Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles M. Fernanda P.

More information

Analyzing Longitudinal Data Using Regression Splines

Analyzing Longitudinal Data Using Regression Splines Analyzing Longitudinal Data Using Regression Splines Zhang Jin-Ting Dept of Stat & Appl Prob National University of Sinagpore August 18, 6 DSAP, NUS p.1/16 OUTLINE Motivating Longitudinal Data Parametric

More information

Topics in Machine Learning-EE 5359 Model Assessment and Selection

Topics in Machine Learning-EE 5359 Model Assessment and Selection Topics in Machine Learning-EE 5359 Model Assessment and Selection Ioannis D. Schizas Electrical Engineering Department University of Texas at Arlington 1 Training and Generalization Training stage: Utilizing

More information

Analysis of variance - ANOVA

Analysis of variance - ANOVA Analysis of variance - ANOVA Based on a book by Julian J. Faraway University of Iceland (UI) Estimation 1 / 50 Anova In ANOVAs all predictors are categorical/qualitative. The original thinking was to try

More information

Artificial Intelligence for Robotics: A Brief Summary

Artificial Intelligence for Robotics: A Brief Summary Artificial Intelligence for Robotics: A Brief Summary This document provides a summary of the course, Artificial Intelligence for Robotics, and highlights main concepts. Lesson 1: Localization (using Histogram

More information

Clustering. Mihaela van der Schaar. January 27, Department of Engineering Science University of Oxford

Clustering. Mihaela van der Schaar. January 27, Department of Engineering Science University of Oxford Department of Engineering Science University of Oxford January 27, 2017 Many datasets consist of multiple heterogeneous subsets. Cluster analysis: Given an unlabelled data, want algorithms that automatically

More information

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)

More information

Camera Motion Identification in the Rough Indexing Paradigm

Camera Motion Identification in the Rough Indexing Paradigm Camera Motion Identification in the Rough Indexing aradigm etra KRÄMER and Jenny BENOIS-INEAU LaBRI University Bordeaux I, France {petra.kraemer,jenny.benois}@labri.fr Camera Motion Identification in the

More information

An imputation approach for analyzing mixed-mode surveys

An imputation approach for analyzing mixed-mode surveys An imputation approach for analyzing mixed-mode surveys Jae-kwang Kim 1 Iowa State University June 4, 2013 1 Joint work with S. Park and S. Kim Ouline Introduction Proposed Methodology Application to Private

More information

Modelling and Quantitative Methods in Fisheries

Modelling and Quantitative Methods in Fisheries SUB Hamburg A/553843 Modelling and Quantitative Methods in Fisheries Second Edition Malcolm Haddon ( r oc) CRC Press \ y* J Taylor & Francis Croup Boca Raton London New York CRC Press is an imprint of

More information

Model-based Recursive Partitioning for Subgroup Analyses

Model-based Recursive Partitioning for Subgroup Analyses EBPI Epidemiology, Biostatistics and Prevention Institute Model-based Recursive Partitioning for Subgroup Analyses Torsten Hothorn; joint work with Heidi Seibold and Achim Zeileis 2014-12-03 Subgroup analyses

More information

Taking into account sampling variability of model selection indices: A parametric bootstrap approach. Sunthud Pornprasertmanit, Wei Wu, Todd D.

Taking into account sampling variability of model selection indices: A parametric bootstrap approach. Sunthud Pornprasertmanit, Wei Wu, Todd D. Parametric Bootstrap in Nonnested Model Comparison 1 Taking into account sampling variability of model selection indices: A parametric bootstrap approach Sunthud Pornprasertmanit, Wei Wu, Todd D. Little

More information

Outline. No Free Lunch Theorems SMTWTP. Outline DM812 METAHEURISTICS

Outline. No Free Lunch Theorems SMTWTP. Outline DM812 METAHEURISTICS DM812 METAHEURISTICS Outline Lecture 9 Marco Chiarandini 1. Department of Mathematics and Computer Science University of Southern Denmark, Odense, Denmark 2. Outline 1. 2. Linear permutations

More information

Approximate Bayesian Computation using Auxiliary Models

Approximate Bayesian Computation using Auxiliary Models Approximate Bayesian Computation using Auxiliary Models Tony Pettitt Co-authors Chris Drovandi, Malcolm Faddy Queensland University of Technology Brisbane MCQMC February 2012 Tony Pettitt () ABC using

More information

Digital Image Processing Laboratory: MAP Image Restoration

Digital Image Processing Laboratory: MAP Image Restoration Purdue University: Digital Image Processing Laboratories 1 Digital Image Processing Laboratory: MAP Image Restoration October, 015 1 Introduction This laboratory explores the use of maximum a posteriori

More information

Minitab 17 commands Prepared by Jeffrey S. Simonoff

Minitab 17 commands Prepared by Jeffrey S. Simonoff Minitab 17 commands Prepared by Jeffrey S. Simonoff Data entry and manipulation To enter data by hand, click on the Worksheet window, and enter the values in as you would in any spreadsheet. To then save

More information

Estimating Noise and Dimensionality in BCI Data Sets: Towards Illiteracy Comprehension

Estimating Noise and Dimensionality in BCI Data Sets: Towards Illiteracy Comprehension Estimating Noise and Dimensionality in BCI Data Sets: Towards Illiteracy Comprehension Claudia Sannelli, Mikio Braun, Michael Tangermann, Klaus-Robert Müller, Machine Learning Laboratory, Dept. Computer

More information

Week 4: Simple Linear Regression II

Week 4: Simple Linear Regression II Week 4: Simple Linear Regression II Marcelo Coca Perraillon University of Colorado Anschutz Medical Campus Health Services Research Methods I HSMP 7607 2017 c 2017 PERRAILLON ARR 1 Outline Algebraic properties

More information

Macros and ODS. SAS Programming November 6, / 89

Macros and ODS. SAS Programming November 6, / 89 Macros and ODS The first part of these slides overlaps with last week a fair bit, but it doesn t hurt to review as this code might be a little harder to follow. SAS Programming November 6, 2014 1 / 89

More information

2016 Stat-Ease, Inc. & CAMO Software

2016 Stat-Ease, Inc. & CAMO Software Multivariate Analysis and Design of Experiments in practice using The Unscrambler X Frank Westad CAMO Software fw@camo.com Pat Whitcomb Stat-Ease pat@statease.com Agenda Goal: Part 1: Part 2: Show how

More information

The Bootstrap and Jackknife

The Bootstrap and Jackknife The Bootstrap and Jackknife Summer 2017 Summer Institutes 249 Bootstrap & Jackknife Motivation In scientific research Interest often focuses upon the estimation of some unknown parameter, θ. The parameter

More information

SIMULTANEOUS COMPUTATION OF MODEL ORDER AND PARAMETER ESTIMATION FOR ARX MODEL BASED ON MULTI- SWARM PARTICLE SWARM OPTIMIZATION

SIMULTANEOUS COMPUTATION OF MODEL ORDER AND PARAMETER ESTIMATION FOR ARX MODEL BASED ON MULTI- SWARM PARTICLE SWARM OPTIMIZATION SIMULTANEOUS COMPUTATION OF MODEL ORDER AND PARAMETER ESTIMATION FOR ARX MODEL BASED ON MULTI- SWARM PARTICLE SWARM OPTIMIZATION Kamil Zakwan Mohd Azmi, Zuwairie Ibrahim and Dwi Pebrianti Faculty of Electrical

More information

Strategies for Modeling Two Categorical Variables with Multiple Category Choices

Strategies for Modeling Two Categorical Variables with Multiple Category Choices 003 Joint Statistical Meetings - Section on Survey Research Methods Strategies for Modeling Two Categorical Variables with Multiple Category Choices Christopher R. Bilder Department of Statistics, University

More information