The partial Package. R topics documented: October 16, Version 0.1. Date Title partial package. Author Andrea Lehnert-Batar
|
|
- Martin Tucker
- 5 years ago
- Views:
Transcription
1 The partial Package October 16, 2006 Version 0.1 Date Title partial package Author Andrea Lehnert-Batar Maintainer Andrea Lehnert-Batar Depends R (>= 2.0.1),e1071 Description partial implements (partial) attributable risk estimates, corresponding variance estimates and confidence intervals. License GPL (version 2 or later) R topics documented: AR PartialAR boot icu Index 11 AR (Adjusted) Attributable Risk Estimates, Variances and Confidence Intervals Description Usage AR derives crude, adjusted, crude joint or adjusted joint attributable risk estimates for one or multiple exposure factors of primary interest adjusted for secondary confounding variables together with different variance estimates and confidence intervals. AR( D, x, C = NULL, model = NULL, fmla, w = NULL, Var = c("none","delta","boot","bayes","jackknife"), CI = c("none","normal","logit","percentile","bca"), alpha = 0.05, B = 500) 1
2 2 AR Arguments D x C w model fmla Var CI alpha B a vector containing a dichotomous indicator variable for the disease status. a matrix containing a dichotomous indicator variable for the exposure status. If ncol(x)=1, the crude or adjusted attributable risk for the risk factor in x is computed. If ncol(x)>1, the joint attributable risk of the multiple risk factors in x is returned. a matrix containing one or multiple confounding variables for adjusting the attributable risk. Every column of C must be dichotomous. If C=NULL, the crude attributable risk for the exposure in x is computed. a weight vector which is used to define the resampling technique. If a nonparametric or bayesian bootstrap or the jackknife is used, w can be ignored by the user as it is regulated by the input parameter Var. If else the user wants to use a different resampling method, w can individually be changed. if model=true, the attributable risk is computed by use of coefficients from a logistic regression model. If model=null, the attributable risk is computed with probabilities directly estimated from the contingency tables of the data set. if model=true, fmla defines the desired form of the logistic regression model and is an obligatory parameter. a character string indicating the method of variance estimation: Var="delta" indicates a variance estimate derived over the delta method, Var="boot" means application of a nonparametric bootstrap, Var="bayes" indicates the Bayesian Bootstrap and Var="jackknife" the Jackknife. If the default Var="none" is selected, only the point estimate of the attributable risk is returned. Var="none" is the default! a character string indicating the method of confidence interval estimation: if CI="normal" a confidence interval constructed by using percentiles from a standard normal distribution is computed. If CI="logit" a logit-transformation of the attributable risk is used. If the logit-transformation is used together with variance estimation based on resampling methods (bootstrap or jackknife) moments of a truncated normal distribution are used for the construction of the confidence interval if the empirical distribution of the attributable risk contains negative values. "percentile" and CI="BCa" yields confidence intervals based on the simple percentile method and the BCa method, respectively (only possible when Var="boot" or Var="bayes"). CI="none" is the default! the probability of error for the estimation of confidence intervals, yielding a 1 α confidence level. number of replications for resampling methods. Details Point estimates of the crude attributable risk are the same wether the model free or the model based approach is applied. For adjusted attributable risk estimates, the model free approach yields estimates based on the case load weighting approach, where the attributable risk is written as a weighted-sum over all adjustment levels of the confounding variables in C. If the model based approach is applied based on a main effects model (for instance fmla = "D~x1+c1+c2") the point estimate is based on the Mantel-Haenszel approach. This means that an adjusted Odds Ratio is plugged into the formula of the attributable risk. By using a fully saturated model for estimation (thus fmla="d~(x1+c1+c2)^3"), the point estimate again equals the model free estimate, as the interaction structure within the data is totally considered. If adjusted attributable risks are computed for large data sets containing no sparse strata, the model free approach is recommended. But
3 AR 3 if many of the strata defined by the adjustment levels of the confounding variables contain only few observations, the model based approach should be used. The main benefit of the model based approach is its flexibility, as a logistic regression model perfectly mapping the interaction structure of the dataset can be determined over the parameter fmla. All supplied variants of variance estimation yield asymptotic estimates. The variance estimate based on the delta method is computationally least expensive. It is based on a expansion of the attributable risk about its mean by taking a one step Taylor approximation. The nonparametric bootstrap (Var="boot") is based on Efron s bootstrap. Samples are taken out of the underlying data set and the attributable risk is repeatedly computed for each of those samples. Rubin s bayesian bootstrap (Var="bayes") is a weighted method of the bootstrap. A weight vector is generated from a Dirichlet distribution. For each sampled weight vector, the attributable risk is computed with weighted observations. The jackknife can be seen as an approximation to Efron s nonparametric bootstrap. Replications of the attributable risk are computed by leaving out every single observation once at a time. If the data set contains far more observations as the number of replications normally computed for a bayesian or nonparametric bootstrap, the computation of a bootstrap is recommended. For the number of replications, B=500 is usually assumed to suffice. The number of replications should be increased if BCa-intervals are computed. BCa-intervals thus are computationally expensive, but yield very stable results in many situations. Attention: Results of simulation studies showed that the application of the simple percentile method can yield unsatisfactory results when applied to the adjusted AR. Value If attributable risk estimates are computed a single value for the point estimate is returned. If Var!="none" a list containing the point estimate together with its corresponding variance estimate is returned. If CI!="none" a list containing the point estimate, the variance estimate and the corresponding confidence interval is returned. Note The variables D, x and C have to be dichotomous, but it has to be ensured that they are not defined as factors. D has to be a vector, whereas x and C have to be matrices. If the model based approach is used, the colnames of x and C have to be used in fmla! Note that the implemented estimates of the attributable risk are only valid if the data has been obtained under a multinomial sampling model! Author(s) Andrea Lehnert-Batar References Basu S., Landis J.R. (1995) Model-based Estimation of Population Attributable Risk under Crosssectional Sampling. American Journal of Epidemiology, 142, Benichou J., Gail M.H. (1989) A Delta Method for Implicitly Defined Random Variables. The American Statistician, 43, Benichou J. (2001) A review of adjusted estimators of attributable risk. Statistical Methods in Medical Research, 10, Efron B. (1979) Bootstrap methods: another look at the jackknife. Annals of Statistics, 7,1-26. Gefeller O. (1992) An annotated bibliography on the attributable risk. Biometrical Journal, 34,
4 4 PartialAR Lehnert-Batar A., Pfahlberg A., Gefeller O. (2006) Confidence Intervals for Attributable Risk Estimates under Multinomial Sampling. Biometrical Journal, to appear. Quenouille M. (1949) Approximation tests of correlation in time series. Journal of the Royal Statistical Society, Series B, 11, Rubin D.B. (1981) The Bayesian Bootstrap. The Annals of Statistics, 9, See Also PartialAR, boot Examples data(icu) attach(data.frame(icu)) ### Computation of crude AR for INF model-free and ### ### model-based with variance estimates ### Exp <- matrix(inf,ncol=1) colnames(exp) <- "Infection" AR(D = STA, x = Exp, Var = "delta") AR(D = STA, x = Exp, model = TRUE, fmla = "STA~Infection", Var="delta") ### Computation of adjusted AR model-free adjusted for SEX and RACE ### ### First coerce variable RACE into dummy matrix as it is not dichotomous! ### RACENEW <- model.matrix(~as.factor(race)-1) AR(D = STA, x = Exp, C = cbind(sex,racenew), Var="delta", CI="normal") ### Computation of joint attributable risk for exposure factors INF and TYP ### ### adjusted for RACE and SEX ### Exp2 <- cbind(inf,typ) Conf <- cbind(racenew,sex) AR(D = STA, x = Exp2, C = Conf, Var="delta") ### Computation of model-based AR for the exposure factor INF, ###\n ### adjusted for TYP, LOC and SEX ### Conf <- cbind(racenew[,-3],sex) colnames(conf) <- c("white","black","sex") AR(D = STA, x = Exp, C = Conf, model = TRUE, fmla = "STA~Infection+white+black+sex", Var = "delta", CI = "normal") PartialAR Partial Attributable Risk Estimates, Variances and Confidence Intervals Description PartialAR derives estimates for partial attributable risks for multiple exposure factors. Variance estimates and confidence intervals can optionally be returned. Variance estimates are available based on either resampling methods or the delta method. Confidence intervals based on the BCa method or alternatively constructed with percentiles of the standard normal distribution are available.
5 PartialAR 5 Usage PartialAR( D, x, w = NULL, model = NULL, fmla, Var = c("none","delta","boot","bayes","jackknife"), CI = c("none","normal","percentile","bca"), alpha = 0.05, B = 500) Arguments D x w model fmla Var CI alpha B a vector containing a dichotomous indicator variable for the disease status. a matrix containing different exposure variables. Each column of x indicates the exposure status of one exposure variable, which has to be dichotomous. Categorial risk factors have first to be transformed into dummy variables. a weight vector which is used to define the resampling technique. If a nonparametric or bayesian bootstrap or the jackknife is used, w can be ignored by the user as it is regulated by the input parameter Var. If else the user wants to use a different resampling method, w can individually be changed. if model=true, the conditional probabilities necessary for the partial attributable risks are computed by use of estimated coefficients from a logistic regression. If model=null, the conditional probabilities are directly estimated from the data. if model=true, fmla defines the desired form of the logistic regression model and is an obligatory parameter. a character string indicating the method of variance estimation: Var="delta" indicates a variance estimated derived over the delta method, Var="boot" and Var="bayes" means a nonparametric bootstrap and bayesian bootstrap, respectively, Var="jackknife" the Jackknife. Var="none" is the default! a character string indicating the method of confidence interval estimation: if CI="normal" a confidence interval constructed by using percentiles from a standard normal distribution is computed. CI="percentile" and CI="BCa" yields confidence intervals based on the simple percentile and the BCa method, respectively (only possible when Var="boot" or Var="bayes"). CI="none" is the default! the probability of error for the estimation of confidence intervals, yielding a 1 α confidence level. number of replications for resampling methods. Details The partial attributable risk additively decomposes the joint attributable risk of the risk factors in p into risk shares for each single factor. The parameter conceptually corresponds to the Shapley value from cooperative game theory. The partial attributable risk estimates give the possibility of ranking the risk factors according to their individual relevance for the disease load within the population under study. If model=null, the conditional probabilities necessary for the computation of the partial attributable risks are directly estimated from the data set. Else if model=true, a logistic regression is used for estimating the conditional probabilities. If the model based approach is used, a formula defining the form of the logistic regression model has to be given in fmla. By choosing the model based approach, the matrix p must contain colnames which are used in fmla (see example!). All supplied variants of variance estimation yield asymptotic estimates. The variance estimate based on the delta method is computationally least expensive. It is based on an expansion of the
6 6 PartialAR Value partial attributable risk about its mean by taking a one step Taylor approximation. If the sample size of the data set is small, the delta method can yield insufficient results. The nonparametric bootstrap (Var="boot") is based on Efron s bootstrap. Samples are taken out of the underlying data set and the attributable risk is repeatedly computed for each of those samples. Rubin s bayesian bootstrap (Var="bayes") is a weighted method of the bootstrap. A weight vector is generated from a Dirichlet distribution. For each sampled weight vector, the partial attributable risks are computed with weighted observations. The jackknife can be seen as an approximation to Efron s nonparametric bootstrap. Replications of the partial attributable risks are computed by leaving out every single observation once at a time. If the data set contains far more observations as the number of replications normally computed for a bayesian or nonparametric bootstrap, the computation of a bootstrap is recommended. For the number of replications, B=500 is normally assumed to suffice. The number of replications should be increased if BCa-intervals are computed. BCa-intervals thus are computationally expensive, but yield very stable results in many situations. Results of a simulation study showed that the bayesian bootstrap tends to overestimate the variance. In situations where a sufficient amount of observations is available within the different strata defined by the status of disease and exposures, the use of variance estimates derived over the delta method is advisable, as the computation is computationally least expensive. If sparse data situations occur, the nonparametric bootstrap combined with the BCa-method should be chosen. If only partial attributable risk estimates are computed a vector of length identical to the number of exposure variables in p is returned. If Var=TRUE, a list containing the vector of partial attributable risks and the vector with corresponding variance estimates is returned. If CI=TRUE, a list containing the vector of partial attributable risks, their variance estimates and a matrix with endpoints of confidence intervals is returned. Note The variables in D and x have to be dichotomous, but it has to be ensured that they are not defined as factors. D has to be a vector, whereas x has to be a matrix. If the model based approach is used, the colnames of x have to be used in fmla! Note that the implemented estimates of the attributable risk are only valid if the data has been obtained under a multinomial sampling model! Author(s) Andrea Lehnert-Batar References Cox L A JR. (1985), A new measure of attributable risk for public health applications. Management Science, 31, Efron B. (1979) Bootstrap methods: another look at the jackknife. Annals of Statistics, 7,1-26. Eide G E, Gefeller O. (1995), Sequential and average attributable fractions as aids in the selection of preventive strategies. Journal of Clinical Epidemiology, 48, Grömping U, Weimann U. (2004), The asymptotic distribution of the partial attributable risk in cross-sectional studies. Statistics, 38, Land M, Vogel C, Gefeller O. (2001), Partitioning methods for multi-factorial risk attribution. Statistical Methods in Medical Research, 10, Quenouille M. (1949) Approximation tests of correlation in time series. Journal of the Royal Statistical Society, Series B, 11,
7 boot 7 See Also AR, boot Examples data(icu) attach(data.frame(icu)) ### Computation of partial attributable risks together with corresponing ### ### confidence intervals based on variance estimates derived from delta method ### PartialAR(STA,cbind(CAN,INF,TYP,LOC),Var="delta",CI="normal") ### Compuation of partial attributable risks by model-based approach ### ### using a simple main-effects model ### Exp <- cbind(can,inf,typ,loc) colnames(exp) <- c("cancer","infection","admission","coma") PartialAR(D = STA, x = Exp, model = TRUE, fmla = "STA~cancer+infection+admission+coma", Var = "delta", CI = "normal") boot Bootstrap and Jackknife Replications for the Attributable Risk or Partial Attributable Risks Description Usage boot computes replications of the (adjusted) attributable risk or partial attributable risks in order to obtain the empirical distribution of the estimated parameter. The computation of replications is either based on the nonparametric bootstrap, the bayesian bootstrap or the jackknife. boot(d, x, C = NULL, param = c("ar","par"), model = NULL, fmla = fmla, type = c("boot","bayes","jackknife"), B = 500) Arguments D x C param model fmla a vector containing a dichotomous indicator variable for the disease status. a matrix containing different exposure variables. Each column of x indicates the exposure status of one exposure variable, which has to be dichotomous. Categorial risk factors have first to be transformed into dummy variables. a matrix containing one or multiple confounding variables for adjusting the attributable risk. Every column of C must be dichotomous. If C=NULL, replications of the crude attributable risk for the exposure in x are computed. This input parameter is only valid if param="ar". defines, whether replications for the (adjusted) attributable risk (param="ar") or for partial attributable risks (param="par") are computed. if model=true the model based approach of estimating the (partial) attributable risk is used for the computation of replications. if the model based approach is used, fmla determines the logistic regression model.
8 8 boot B type Number of replications. A character string indicating the resampling method: type="boot" computes replications based on nonparametric bootstrap, type="jackknife" uses the jackknife. Details Value Note The nonparametric bootstrap (type="boot") is based on Efron s bootstrap. Samples are taken out of the underlying data set and the attributable risk is repeatedly computed for each of those samples. Rubin s bayesian bootstrap is a weighted method of the bootstrap. A weight vector is generated from a Dirichlet distribution. For each sampled weight vector, the attributable risk is computed with weighted observations. The jackknife can be seen as an approximation to Efron s nonparametric bootstrap. Replications of the attributable risk are computed by leaving out every singel observation once at a time. If the data set contains far more observations as the number of replications normally computed for a bayesian or nonparametric bootstrap, the computation of a bootstrap is recommended. For the number of replications, B=500 is usually assumed to suffice. If param="ar", a vector of length B containing replications of the attributable risk is returned. Else if param="par", a matrix with B rows and ncol(x) columns with replications of the partial attributable risks is returned. The variables in D and x have to be dichotomous, but it has to be ensured that they are not defined as factors. D has to be a vector, whereas x has to be a matrix. If the model based approach is used, the colnames of x have to be used in fmla! Note that the implemented estimates of the attributable risk are only valid if the data has been obtained under a multinomial sampling model! Author(s) Andrea Lehnert-Batar References Basu S., Landis J.R. (1995) Model-based Estimation of Population Attributable Risk under Crosssectional Sampling. American Journal of Epidemiology, 142, Benichou J., Gail M.H. (1989) A Delta Method for Implicitly Defined Random Variables. The American Statistician, 43, Benichou J. (2001) A review of adjusted estimators of attributable risk. Statistical Methods in Medical Research, 10, Cox L A JR. (1985), A new measure of attributable risk for public health applications. Management Science, 31, Efron B. (1979) Bootstrap methods: another look at the jackknife. Annals of Statistics, 7,1-26. Eide G E, Gefeller O. (1995), Sequential and average attributable fractions as aids in the selection of preventive strategies. Journal of Clinical Epidemiology, 48, Lehnert-Batar A., Pfahlberg A., Gefeller O. (2006) Confidence Intervals for Attributable Risk Estimates under Multinomial Sampling. Biometrical Journal, to appear. Quenouille M. (1949) Approximation tests of correlation in time series. Journal of the Royal Statistical Society, Series B, 11,
9 icu 9 See Also Rubin D.B. (1981) The Bayesian Bootstrap. The Annals of Statistics, 9, AR Examples data(icu) attach(data.frame(icu)) ### Computation of nonparametric bootstrap replications for the ### ### adjusted AR of INF adjusted for SEX ### boot(d = STA,x = INF, C = SEX, param = "AR", type="boot") ### Computation of nonparametric bootstrap replications ### ### of partial attributable risks ### boot(d = STA,x = cbind(can,inf,typ,loc), param = "PAR", type="boot") icu ICU Data Description Usage Format Details Source The ICU data frame has 200 rows and 7 columns. data(icu) This data frame contains the following columns: STA factor, vital status (0 = Lived, 1 = Died). SEX factor, sex (0 = male, 1 = female). RACE factor, Race (1 = white, 2 = black, 3 = other). CAN factor, Cancer part of present problem (0 = No, 1 = Yes). INF factor, Infection probable at ICU admission (0 = No, 1 = Yes). TYP factor, Type of admission (0 = Elective, 1 = Emergency). LOC factor, Level of consciousness at ICU admission (0 = no coma, 1 = deep stupor or coma). The ICU data set consists of a sample of 200 subjects who were part of a much larger study on survival of patients following admission to an adult intensive care unit (ICU). The data set presented here contains all variables which significantly explained the vital status at hospital discharge. The data were collected at Baystate Medical Center in Springfield, Massachusetts. These data are copyrighted by John Wiley & Sons Inc. and must be acknowledged and used accordingly. Hosmer and Lemeshow (2000), Applied Logistic Regression: Second Edition.
10 10 icu References D.W. Hosmer and S. Lemeshow (2000), Applied Logistic Regression. New York: Wiley Series in Probability and Mathematical Statistics
11 Index Topic datasets icu, 9 Topic manip AR, 1 boot, 7 PartialAR, 4 AR, 1, 7, 9 boot, 4, 7, 7 icu, 9 PartialAR, 4, 4 11
The Bootstrap and Jackknife
The Bootstrap and Jackknife Summer 2017 Summer Institutes 249 Bootstrap & Jackknife Motivation In scientific research Interest often focuses upon the estimation of some unknown parameter, θ. The parameter
More informationCalculating measures of biological interaction
European Journal of Epidemiology (2005) 20: 575 579 Ó Springer 2005 DOI 10.1007/s10654-005-7835-x METHODS Calculating measures of biological interaction Tomas Andersson 1, Lars Alfredsson 1,2, Henrik Ka
More informationAn Introduction to the Bootstrap
An Introduction to the Bootstrap Bradley Efron Department of Statistics Stanford University and Robert J. Tibshirani Department of Preventative Medicine and Biostatistics and Department of Statistics,
More informationCOPYRIGHTED MATERIAL CONTENTS
PREFACE ACKNOWLEDGMENTS LIST OF TABLES xi xv xvii 1 INTRODUCTION 1 1.1 Historical Background 1 1.2 Definition and Relationship to the Delta Method and Other Resampling Methods 3 1.2.1 Jackknife 6 1.2.2
More informationEvaluating generalization (validation) Harvard-MIT Division of Health Sciences and Technology HST.951J: Medical Decision Support
Evaluating generalization (validation) Harvard-MIT Division of Health Sciences and Technology HST.951J: Medical Decision Support Topics Validation of biomedical models Data-splitting Resampling Cross-validation
More informationPackage rereg. May 30, 2018
Title Recurrent Event Regression Version 1.1.4 Package rereg May 30, 2018 A collection of regression models for recurrent event process and failure time. Available methods include these from Xu et al.
More information3.6 Sample code: yrbs_data <- read.spss("yrbs07.sav",to.data.frame=true)
InJanuary2009,CDCproducedareportSoftwareforAnalyisofYRBSdata, describingtheuseofsas,sudaan,stata,spss,andepiinfoforanalyzingdatafrom theyouthriskbehaviorssurvey. ThisreportprovidesthesameinformationforRandthesurveypackage.Thetextof
More informationApplied Survey Data Analysis Module 2: Variance Estimation March 30, 2013
Applied Statistics Lab Applied Survey Data Analysis Module 2: Variance Estimation March 30, 2013 Approaches to Complex Sample Variance Estimation In simple random samples many estimators are linear estimators
More informationPackage DSBayes. February 19, 2015
Type Package Title Bayesian subgroup analysis in clinical trials Version 1.1 Date 2013-12-28 Copyright Ravi Varadhan Package DSBayes February 19, 2015 URL http: //www.jhsph.edu/agingandhealth/people/faculty_personal_pages/varadhan.html
More informationPackage samplesizelogisticcasecontrol
Package samplesizelogisticcasecontrol February 4, 2017 Title Sample Size Calculations for Case-Control Studies Version 0.0.6 Date 2017-01-31 Author Mitchell H. Gail To determine sample size for case-control
More informationCross-validation and the Bootstrap
Cross-validation and the Bootstrap In the section we discuss two resampling methods: cross-validation and the bootstrap. 1/44 Cross-validation and the Bootstrap In the section we discuss two resampling
More informationCorrectly Compute Complex Samples Statistics
SPSS Complex Samples 15.0 Specifications Correctly Compute Complex Samples Statistics When you conduct sample surveys, use a statistics package dedicated to producing correct estimates for complex sample
More informationCross-validation and the Bootstrap
Cross-validation and the Bootstrap In the section we discuss two resampling methods: cross-validation and the bootstrap. These methods refit a model of interest to samples formed from the training set,
More informationBootstrap Confidence Interval of the Difference Between Two Process Capability Indices
Int J Adv Manuf Technol (2003) 21:249 256 Ownership and Copyright 2003 Springer-Verlag London Limited Bootstrap Confidence Interval of the Difference Between Two Process Capability Indices J.-P. Chen 1
More informationin this course) ˆ Y =time to event, follow-up curtailed: covered under ˆ Missing at random (MAR) a
Chapter 3 Missing Data 3.1 Types of Missing Data ˆ Missing completely at random (MCAR) ˆ Missing at random (MAR) a ˆ Informative missing (non-ignorable non-response) See 1, 38, 59 for an introduction to
More informationAcknowledgments. Acronyms
Acknowledgments Preface Acronyms xi xiii xv 1 Basic Tools 1 1.1 Goals of inference 1 1.1.1 Population or process? 1 1.1.2 Probability samples 2 1.1.3 Sampling weights 3 1.1.4 Design effects. 5 1.2 An introduction
More informationModelling and Quantitative Methods in Fisheries
SUB Hamburg A/553843 Modelling and Quantitative Methods in Fisheries Second Edition Malcolm Haddon ( r oc) CRC Press \ y* J Taylor & Francis Croup Boca Raton London New York CRC Press is an imprint of
More informationCorrectly Compute Complex Samples Statistics
PASW Complex Samples 17.0 Specifications Correctly Compute Complex Samples Statistics When you conduct sample surveys, use a statistics package dedicated to producing correct estimates for complex sample
More informationStatistics (STAT) Statistics (STAT) 1. Prerequisites: grade in C- or higher in STAT 1200 or STAT 1300 or STAT 1400
Statistics (STAT) 1 Statistics (STAT) STAT 1200: Introductory Statistical Reasoning Statistical concepts for critically evaluation quantitative information. Descriptive statistics, probability, estimation,
More informationPackage PTE. October 10, 2017
Type Package Title Personalized Treatment Evaluator Version 1.6 Date 2017-10-9 Package PTE October 10, 2017 Author Adam Kapelner, Alina Levine & Justin Bleich Maintainer Adam Kapelner
More informationStatistical Methods for the Analysis of Repeated Measurements
Charles S. Davis Statistical Methods for the Analysis of Repeated Measurements With 20 Illustrations #j Springer Contents Preface List of Tables List of Figures v xv xxiii 1 Introduction 1 1.1 Repeated
More informationPackage samplesizecmh
Package samplesizecmh Title Power and Sample Size Calculation for the Cochran-Mantel-Haenszel Test Date 2017-12-13 Version 0.0.0 Copyright Spectrum Health, Grand Rapids, MI December 21, 2017 Calculates
More informationSTATISTICS (STAT) Statistics (STAT) 1
Statistics (STAT) 1 STATISTICS (STAT) STAT 2013 Elementary Statistics (A) Prerequisites: MATH 1483 or MATH 1513, each with a grade of "C" or better; or an acceptable placement score (see placement.okstate.edu).
More informationPackage FHtest. November 8, 2017
Type Package Package FHtest November 8, 2017 Title Tests for Right and Interval-Censored Survival Data Based on the Fleming-Harrington Class Version 1.4 Date 2017-11-8 Author Ramon Oller, Klaus Langohr
More informationBayesian Inference for Two-Phase Studies with Categorical Covariates
Bayesian Inference for Two-Phase Studies with Categorical Covariates Michelle Ross and Jon Wakefield, Department of Biostatistics, University of Washington, Seattle, WA 98195. Working Paper no. 123 Center
More informationBootstrap and multiple imputation under missing data in AR(1) models
EUROPEAN ACADEMIC RESEARCH Vol. VI, Issue 7/ October 2018 ISSN 2286-4822 www.euacademic.org Impact Factor: 3.4546 (UIF) DRJI Value: 5.9 (B+) Bootstrap and multiple imputation under missing ELJONA MILO
More informationbook 2014/5/6 15:21 page v #3 List of figures List of tables Preface to the second edition Preface to the first edition
book 2014/5/6 15:21 page v #3 Contents List of figures List of tables Preface to the second edition Preface to the first edition xvii xix xxi xxiii 1 Data input and output 1 1.1 Input........................................
More informationEstimating survival from Gray s flexible model. Outline. I. Introduction. I. Introduction. I. Introduction
Estimating survival from s flexible model Zdenek Valenta Department of Medical Informatics Institute of Computer Science Academy of Sciences of the Czech Republic I. Introduction Outline II. Semi parametric
More informationPaper PO-06. Gone are the days when social and behavioral science researchers should simply report obtained test statistics (e.g.
Paper PO-06 CI_MEDIATE: A SAS Macro for Computing Point and Interval Estimates of Effect Sizes Associated with Mediation Analysis Thanh V. Pham, University of South Florida, Tampa, FL Eun Kyeng Baek, University
More informationTitle. Syntax. svy: tabulate oneway One-way tables for survey data. Basic syntax. svy: tabulate varname. Full syntax
Title svy: tabulate oneway One-way tables for survey data Syntax Basic syntax svy: tabulate varname Full syntax svy [ vcetype ] [, svy options ] : tabulate varname [ if ] [ in ] [, tabulate options display
More informationSmoothing Dissimilarities for Cluster Analysis: Binary Data and Functional Data
Smoothing Dissimilarities for Cluster Analysis: Binary Data and unctional Data David B. University of South Carolina Department of Statistics Joint work with Zhimin Chen University of South Carolina Current
More informationPurposeful Selection of Variables in Logistic Regression: Macro and Simulation Results
ection on tatistical Computing urposeful election of Variables in Logistic Regression: Macro and imulation Results Zoran ursac 1, C. Heath Gauss 1, D. Keith Williams 1, David Hosmer 2 1 iostatistics, University
More informationThe ctest Package. January 3, 2000
R objects documented: The ctest Package January 3, 2000 bartlett.test....................................... 1 binom.test........................................ 2 cor.test.........................................
More informationEstimation and Comparison of Receiver Operating Characteristic Curves
UW Biostatistics Working Paper Series 1-22-2008 Estimation and Comparison of Receiver Operating Characteristic Curves Margaret Pepe University of Washington, Fred Hutch Cancer Research Center, mspepe@u.washington.edu
More information(R) / / / / / / / / / / / / Statistics/Data Analysis
(R) / / / / / / / / / / / / Statistics/Data Analysis help incroc (version 1.0.2) Title incroc Incremental value of a marker relative to a list of existing predictors. Evaluation is with respect to receiver
More informationSampling Size Calculations for Estimating the Proportion of False Positive
Sampling Size Calculations for Estimating the Proportion of False Positive and False Negative CLABSIs in the State of Texas J. Charles Huber Jr, PhD Associate Professor of Biostatistics Ryan Hollingsworth
More informationA noninformative Bayesian approach to small area estimation
A noninformative Bayesian approach to small area estimation Glen Meeden School of Statistics University of Minnesota Minneapolis, MN 55455 glen@stat.umn.edu September 2001 Revised May 2002 Research supported
More informationHealth Disparities (HD): It s just about comparing two groups
A review of modern methods of estimating the size of health disparities May 24, 2017 Emil Coman 1 Helen Wu 2 1 UConn Health Disparities Institute, 2 UConn Health Modern Modeling conference, May 22-24,
More informationCategorical Data in a Designed Experiment Part 2: Sizing with a Binary Response
Categorical Data in a Designed Experiment Part 2: Sizing with a Binary Response Authored by: Francisco Ortiz, PhD Version 2: 19 July 2018 Revised 18 October 2018 The goal of the STAT COE is to assist in
More informationPackage spcadjust. September 29, 2016
Version 1.1 Date 2015-11-20 Title Functions for Calibrating Control Charts Package spcadjust September 29, 2016 Author Axel Gandy and Jan Terje Kvaloy . Maintainer
More informationBrief Guide on Using SPSS 10.0
Brief Guide on Using SPSS 10.0 (Use student data, 22 cases, studentp.dat in Dr. Chang s Data Directory Page) (Page address: http://www.cis.ysu.edu/~chang/stat/) I. Processing File and Data To open a new
More informationGene signature selection to predict survival benefits from adjuvant chemotherapy in NSCLC patients
1 Gene signature selection to predict survival benefits from adjuvant chemotherapy in NSCLC patients 1,2 Keyue Ding, Ph.D. Nov. 8, 2014 1 NCIC Clinical Trials Group, Kingston, Ontario, Canada 2 Dept. Public
More informationPackage ICsurv. February 19, 2015
Package ICsurv February 19, 2015 Type Package Title A package for semiparametric regression analysis of interval-censored data Version 1.0 Date 2014-6-9 Author Christopher S. McMahan and Lianming Wang
More informationPackage deming. June 19, 2018
Package deming June 19, 2018 Title Deming, Thiel-Sen and Passing-Bablock Regression Maintainer Terry Therneau Generalized Deming regression, Theil-Sen regression and Passing-Bablock
More informationMultiple imputation using chained equations: Issues and guidance for practice
Multiple imputation using chained equations: Issues and guidance for practice Ian R. White, Patrick Royston and Angela M. Wood http://onlinelibrary.wiley.com/doi/10.1002/sim.4067/full By Gabrielle Simoneau
More informationPackage MPCI. October 25, 2015
Package MPCI October 25, 2015 Type Package Title Multivariate Process Capability Indices (MPCI) Version 1.0.7 Date 2015-10-23 Depends R (>= 3.1.0), graphics, stats, utils Author Edgar Santos-Fernandez,
More informationPackage compeir. February 19, 2015
Type Package Package compeir February 19, 2015 Title Event-specific incidence rates for competing risks data Version 1.0 Date 2011-03-09 Author Nadine Grambauer, Andreas Neudecker Maintainer Nadine Grambauer
More informationHierarchical Mixture Models for Nested Data Structures
Hierarchical Mixture Models for Nested Data Structures Jeroen K. Vermunt 1 and Jay Magidson 2 1 Department of Methodology and Statistics, Tilburg University, PO Box 90153, 5000 LE Tilburg, Netherlands
More informationProduct Catalog. AcaStat. Software
Product Catalog AcaStat Software AcaStat AcaStat is an inexpensive and easy-to-use data analysis tool. Easily create data files or import data from spreadsheets or delimited text files. Run crosstabulations,
More informationInstall RStudio from - use the standard installation.
Session 1: Reading in Data Before you begin: Install RStudio from http://www.rstudio.com/ide/download/ - use the standard installation. Go to the course website; http://faculty.washington.edu/kenrice/rintro/
More informationStatistical Analysis Using Combined Data Sources: Discussion JPSM Distinguished Lecture University of Maryland
Statistical Analysis Using Combined Data Sources: Discussion 2011 JPSM Distinguished Lecture University of Maryland 1 1 University of Michigan School of Public Health April 2011 Complete (Ideal) vs. Observed
More informationPackage addhaz. September 26, 2018
Package addhaz September 26, 2018 Title Binomial and Multinomial Additive Hazard Models Version 0.5 Description Functions to fit the binomial and multinomial additive hazard models and to estimate the
More informationMissing Data: What Are You Missing?
Missing Data: What Are You Missing? Craig D. Newgard, MD, MPH Jason S. Haukoos, MD, MS Roger J. Lewis, MD, PhD Society for Academic Emergency Medicine Annual Meeting San Francisco, CA May 006 INTRODUCTION
More information100 Myung Hwan Na log-hazard function. The discussion section of Abrahamowicz, et al.(1992) contains a good review of many of the papers on the use of
J. KSIAM Vol.3, No.2, 99-106, 1999 SPLINE HAZARD RATE ESTIMATION USING CENSORED DATA Myung Hwan Na Abstract In this paper, the spline hazard rate model to the randomly censored data is introduced. The
More informationNotes on Simulations in SAS Studio
Notes on Simulations in SAS Studio If you are not careful about simulations in SAS Studio, you can run into problems. In particular, SAS Studio has a limited amount of memory that you can use to write
More informationPackage abe. October 30, 2017
Package abe October 30, 2017 Type Package Title Augmented Backward Elimination Version 3.0.1 Date 2017-10-25 Author Rok Blagus [aut, cre], Sladana Babic [ctb] Maintainer Rok Blagus
More informationResampling Methods. Levi Waldron, CUNY School of Public Health. July 13, 2016
Resampling Methods Levi Waldron, CUNY School of Public Health July 13, 2016 Outline and introduction Objectives: prediction or inference? Cross-validation Bootstrap Permutation Test Monte Carlo Simulation
More informationSTAT 311 (3 CREDITS) VARIANCE AND REGRESSION ANALYSIS ELECTIVE: ALL STUDENTS. CONTENT Introduction to Computer application of variance and regression
STAT 311 (3 CREDITS) VARIANCE AND REGRESSION ANALYSIS ELECTIVE: ALL STUDENTS. CONTENT Introduction to Computer application of variance and regression analysis. Analysis of Variance: one way classification,
More informationThe Use of Sample Weights in Hot Deck Imputation
Journal of Official Statistics, Vol. 25, No. 1, 2009, pp. 21 36 The Use of Sample Weights in Hot Deck Imputation Rebecca R. Andridge 1 and Roderick J. Little 1 A common strategy for handling item nonresponse
More informationPackage epitab. July 4, 2018
Type Package Package epitab July 4, 2018 Title Flexible Contingency Tables for Epidemiology Version 0.2.2 Author Stuart Lacy Maintainer Stuart Lacy Builds contingency tables that
More informationnquery Sample Size & Power Calculation Software Validation Guidelines
nquery Sample Size & Power Calculation Software Validation Guidelines Every nquery sample size table, distribution function table, standard deviation table, and tablespecific side table has been tested
More informationPackage binomlogit. February 19, 2015
Type Package Title Efficient MCMC for Binomial Logit Models Version 1.2 Date 2014-03-12 Author Agnes Fussl Maintainer Agnes Fussl Package binomlogit February 19, 2015 Description The R package
More informationDealing with Categorical Data Types in a Designed Experiment
Dealing with Categorical Data Types in a Designed Experiment Part II: Sizing a Designed Experiment When Using a Binary Response Best Practice Authored by: Francisco Ortiz, PhD STAT T&E COE The goal of
More informationAnnotated multitree output
Annotated multitree output A simplified version of the two high-threshold (2HT) model, applied to two experimental conditions, is used as an example to illustrate the output provided by multitree (version
More informationPoisson Regressions for Complex Surveys
Poisson Regressions for Complex Surveys Overview Researchers often use sample survey methodology to obtain information about a large population by selecting and measuring a sample from that population.
More informationComputation of the variance-covariance matrix
Computation of the variance-covariance matrix An example with the Countr package. Tarak Kharrat 1 and Georgi N. Boshnakov 2 1 Salford Business School, University of Salford, UK. 2 School of Mathematics,
More informationPackage OrderedList. December 31, 2017
Title Similarities of Ordered Gene Lists Version 1.50.0 Date 2008-07-09 Package OrderedList December 31, 2017 Author Xinan Yang, Stefanie Scheid, Claudio Lottaz Detection of similarities between ordered
More informationSTATISTICS (STAT) 200 Level Courses. 300 Level Courses. Statistics (STAT) 1
Statistics (STAT) 1 STATISTICS (STAT) 200 Level Courses STAT 250: Introductory Statistics I. 3 credits. Elementary introduction to statistics. Topics include descriptive statistics, probability, and estimation
More informationPackage anidom. July 25, 2017
Type Package Package anidom July 25, 2017 Title Inferring Dominance Hierarchies and Estimating Uncertainty Version 0.1.2 Date 2017-07-25 Author Damien R. Farine and Alfredo Sanchez-Tojar Maintainer Damien
More informationTelephone Survey Response: Effects of Cell Phones in Landline Households
Telephone Survey Response: Effects of Cell Phones in Landline Households Dennis Lambries* ¹, Michael Link², Robert Oldendick 1 ¹University of South Carolina, ²Centers for Disease Control and Prevention
More informationMultiple Imputation for Missing Data. Benjamin Cooper, MPH Public Health Data & Training Center Institute for Public Health
Multiple Imputation for Missing Data Benjamin Cooper, MPH Public Health Data & Training Center Institute for Public Health Outline Missing data mechanisms What is Multiple Imputation? Software Options
More informationModelling Personalized Screening: a Step Forward on Risk Assessment Methods
Modelling Personalized Screening: a Step Forward on Risk Assessment Methods Validating Prediction Models Inmaculada Arostegui Universidad del País Vasco UPV/EHU Red de Investigación en Servicios de Salud
More informationMean Tests & X 2 Parametric vs Nonparametric Errors Selection of a Statistical Test SW242
Mean Tests & X 2 Parametric vs Nonparametric Errors Selection of a Statistical Test SW242 Creation & Description of a Data Set * 4 Levels of Measurement * Nominal, ordinal, interval, ratio * Variable Types
More informationStatistical Matching using Fractional Imputation
Statistical Matching using Fractional Imputation Jae-Kwang Kim 1 Iowa State University 1 Joint work with Emily Berg and Taesung Park 1 Introduction 2 Classical Approaches 3 Proposed method 4 Application:
More informationPackage citools. October 20, 2018
Type Package Package citools October 20, 2018 Title Confidence or Prediction Intervals, Quantiles, and Probabilities for Statistical Models Version 0.5.0 Maintainer John Haman Functions
More informationHow to use the rbsurv Package
How to use the rbsurv Package HyungJun Cho, Sukwoo Kim, Soo-heang Eo, and Jaewoo Kang April 30, 2018 Contents 1 Introduction 1 2 Robust likelihood-based survival modeling 2 3 Algorithm 2 4 Example: Glioma
More informationAnalysis of Complex Survey Data with SAS
ABSTRACT Analysis of Complex Survey Data with SAS Christine R. Wells, Ph.D., UCLA, Los Angeles, CA The differences between data collected via a complex sampling design and data collected via other methods
More informationWeek 4: Simple Linear Regression III
Week 4: Simple Linear Regression III Marcelo Coca Perraillon University of Colorado Anschutz Medical Campus Health Services Research Methods I HSMP 7607 2017 c 2017 PERRAILLON ARR 1 Outline Goodness of
More informationThe relaimpo Package
The relaimpo Package October 1, 2007 Title Relative importance of regressors in linear models Version 1.2-2 Date 2007-09-30 Author Ulrike Groemping Description relaimpo provides several metrics for assessing
More informationLecture 12. August 23, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.
Lecture 12 Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University August 23, 2007 1 2 3 4 5 1 2 Introduce the bootstrap 3 the bootstrap algorithm 4 Example
More informationResearch with Large Databases
Research with Large Databases Key Statistical and Design Issues and Software for Analyzing Large Databases John Ayanian, MD, MPP Ellen P. McCarthy, PhD, MPH Society of General Internal Medicine Chicago,
More informationAPPENDIX B EXCEL BASICS 1
APPENDIX B EXCEL BASICS 1 Microsoft Excel is a powerful application for education researchers and students studying educational statistics. Excel worksheets can hold data for a variety of uses and therefore
More informationPackage oaxaca. January 3, Index 11
Type Package Title Blinder-Oaxaca Decomposition Version 0.1.4 Date 2018-01-01 Package oaxaca January 3, 2018 Author Marek Hlavac Maintainer Marek Hlavac
More informationStatCalc User Manual. Version 9 for Mac and Windows. Copyright 2018, AcaStat Software. All rights Reserved.
StatCalc User Manual Version 9 for Mac and Windows Copyright 2018, AcaStat Software. All rights Reserved. http://www.acastat.com Table of Contents Introduction... 4 Getting Help... 4 Uninstalling StatCalc...
More informationPackage biglars. February 19, 2015
Package biglars February 19, 2015 Type Package Title Scalable Least-Angle Regression and Lasso Version 1.0.2 Date Tue Dec 27 15:06:08 PST 2011 Author Mark Seligman, Chris Fraley, Tim Hesterberg Maintainer
More informationSTAT 113: Lab 9. Colin Reimer Dawson. Last revised November 10, 2015
STAT 113: Lab 9 Colin Reimer Dawson Last revised November 10, 2015 We will do some of the following together. The exercises with a (*) should be done and turned in as part of HW9. Before we start, let
More informationPackage MIICD. May 27, 2017
Type Package Package MIICD May 27, 2017 Title Multiple Imputation for Interval Censored Data Version 2.4 Depends R (>= 2.13.0) Date 2017-05-27 Maintainer Marc Delord Implements multiple
More informationStatistical matching: conditional. independence assumption and auxiliary information
Statistical matching: conditional Training Course Record Linkage and Statistical Matching Mauro Scanu Istat scanu [at] istat.it independence assumption and auxiliary information Outline The conditional
More informationLecture 1: Statistical Reasoning 2. Lecture 1. Simple Regression, An Overview, and Simple Linear Regression
Lecture Simple Regression, An Overview, and Simple Linear Regression Learning Objectives In this set of lectures we will develop a framework for simple linear, logistic, and Cox Proportional Hazards Regression
More informationHow to Use the Cancer-Rates.Info/NJ
How to Use the Cancer-Rates.Info/NJ Web- Based Incidence and Mortality Mapping and Inquiry Tool to Obtain Statewide and County Cancer Statistics for New Jersey Cancer Incidence and Mortality Inquiry System
More informationPackage bootlr. July 13, 2015
Type Package Package bootlr July 13, 2015 Title Bootstrapped Confidence Intervals for (Negative) Likelihood Ratio Tests Version 1.0 Date 2015-07-10 Author Keith A. Marill and Ari B. Friedman Maintainer
More informationHILDA PROJECT TECHNICAL PAPER SERIES No. 2/08, February 2008
HILDA PROJECT TECHNICAL PAPER SERIES No. 2/08, February 2008 HILDA Standard Errors: A Users Guide Clinton Hayes The HILDA Project was initiated, and is funded, by the Australian Government Department of
More informationRESAMPLING METHODS. Chapter 05
1 RESAMPLING METHODS Chapter 05 2 Outline Cross Validation The Validation Set Approach Leave-One-Out Cross Validation K-fold Cross Validation Bias-Variance Trade-off for k-fold Cross Validation Cross Validation
More informationFitting latency models using B-splines in EPICURE for DOS
Fitting latency models using B-splines in EPICURE for DOS Michael Hauptmann, Jay Lubin January 11, 2007 1 Introduction Disease latency refers to the interval between an increment of exposure and a subsequent
More informationData Analysis and Solver Plugins for KSpread USER S MANUAL. Tomasz Maliszewski
Data Analysis and Solver Plugins for KSpread USER S MANUAL Tomasz Maliszewski tmaliszewski@wp.pl Table of Content CHAPTER 1: INTRODUCTION... 3 1.1. ABOUT DATA ANALYSIS PLUGIN... 3 1.3. ABOUT SOLVER PLUGIN...
More informationPackage snn. August 23, 2015
Type Package Title Stabilized Nearest Neighbor Classifier Version 1.1 Date 2015-08-22 Author Package snn August 23, 2015 Maintainer Wei Sun Implement K-nearest neighbor classifier,
More informationPackage lol. R topics documented: December 13, Type Package Title Lots Of Lasso Version Date Author Yinyin Yuan
Type Package Title Lots Of Lasso Version 1.30.0 Date 2011-04-02 Author Package lol December 13, 2018 Maintainer Various optimization methods for Lasso inference with matri warpper
More informationSimulating from the Polya posterior by Glen Meeden, March 06
1 Introduction Simulating from the Polya posterior by Glen Meeden, glen@stat.umn.edu March 06 The Polya posterior is an objective Bayesian approach to finite population sampling. In its simplest form it
More informationPackage DTRreg. August 30, 2017
Package DTRreg August 30, 2017 Type Package Title DTR Estimation and Inference via G-Estimation, Dynamic WOLS, and Q-Learning Version 1.3 Date 2017-08-30 Author Michael Wallace, Erica E M Moodie, David
More informationUpdates and Errata for Statistical Data Analytics (1st edition, 2015)
Updates and Errata for Statistical Data Analytics (1st edition, 2015) Walter W. Piegorsch University of Arizona c 2018 The author. All rights reserved, except where previous rights exist. CONTENTS Preface
More information