Enhancing population pharmacokinetic modeling efficiency and quality using an integrated workflow

Size: px
Start display at page:

Download "Enhancing population pharmacokinetic modeling efficiency and quality using an integrated workflow"

Transcription

1 DOI /s ORIGINAL PAPER Enhancing population pharmacokinetic modeling efficiency and quality using an integrated workflow Henning Schmidt Andrijana Radivojevic Received: 16 May 2014 / Accepted: 8 July 2014 Ó Springer Science+Business Media New York 2014 Abstract Population pharmacokinetic (poppk) analyses are at the core of Pharmacometrics and need to be performed regularly. Although these analyses are relatively standard, a large variability can be observed in both the time (efficiency) and the way they are performed (quality). Main reasons for this variability include the level of experience of a modeler, personal preferences and tools. This paper aims to examine how the process of poppk model building can be supported in order to increase its efficiency and quality. The presented approach to the conduct of poppk analyses is centered around three key components: (1) identification of most common and important poppk model features, (2) required information content and formatting of the data for modeling, and (3) methodology, workflow and workflow supporting tools. This approach has been used in several poppk modeling projects and a documented example is provided in the supplementary material. Efficiency of model building is improved by avoiding repetitive coding and other laborintensive tasks and by putting the emphasis on a fit-forpurpose model. Quality is improved by ensuring that the workflow and tools are in alignment with a poppk modeling guidance which is established within an organization. The main conclusion of this paper is that workflow based approaches to poppk modeling are feasible and have significant potential to ameliorate its various aspects. However, the implementation of such an approach in a pharmacometric organization requires openness towards innovation and Electronic supplementary material The online version of this article (doi: /s ) contains supplementary material, which is available to authorized users. H. Schmidt (&) A. Radivojevic Advanced Quantitative Sciences, Novartis Pharma AG, Postfach, 4002 Basel, Switzerland henning.schmidt@novartis.com change the key ingredient for evolution of integrative and quantitative drug development in the pharmaceutical industry. Keywords Population pharmacokinetics Modeling Workflow Efficiency Quality Dataset specification Introduction Population pharmacokinetic modeling constitutes a substantial area of Pharmacometrics [1] and is an integral part of model-based drug development [2]. In general terms, the main purpose of poppk modeling is to identify the central tendencies, the uncertainty and variability in the pharmacokinetic (PK) parameters, and the influence of covariates [3, 4]. PopPK modeling activities need to be performed regularly during drug development from first-in-human dose selection to submission and beyond [5 7]. In this paper it is assumed that the rational for a specific poppk modeling activity is well defined and the focus is set on the efficient, streamlined, and systematic conduct of poppk analyses. The planning and reporting parts are outside of scope. Anecdotal evidence indicates that lack of meeting the time constraints often limits the extent of pharmacometric involvement in the industry [8]. Although this quote is not specifically addressing poppk analyses, it is well known that these analyses can take a considerable amount of time, binding resources that are then not available to support other aspects of model-based drug development [9, 10]. It is not implicitly assumed that each poppk analysis that requires a long time is not done efficiently and it is not in scope of this paper to assess the manifold reasons why sometimes such analyses take longer. However, one of the stated aims of this

2 Table 1 Relative frequency of poppk model elements, based on internal survey and not considering binding models (N = 52) Model features Number of compartments Other 0 Model outputs Single 94 Multiple (e.g., plasma and tissue concentrations, 6 metabolites) Elimination Linear, from central compartment, time 92 independent elimination rate constant Saturable or linear? saturable, from central 8 compartment, time independent elimination rate constant, saturable part modelled by Michaelis Menten term Other 0 Bioavailability Linear (covariates on F are allowed) 44 Nonlinear (F as nonlinear function of dose or 2 concentration, etc.) Not considered in the model 52 Other 2 Absorption First-order, time independent 71 Zero-order, time independent 0 Mixed zero/first order, time independent 8 Not considered in the model 21 Other 0 Absorption delay Not considered in the model 86 Lag time 10 Transit compartment 4 Other 0 Distribution Linear, time independent 100 Nonlinear or time dependent 0 Other nonlinearities None 96 Saturable red blood cell binding 2 Self-induction, self-inhibition 2 Other 0 Time dependent covariates No 100 Yes 0 Inter occasion variability No 85 Yes 15 Relative frequency (%) paper is to address improvement of efficiency in the conduct of poppk analyses. Here, efficiency means not only a potential decrease of time for turn-around of such analyses, but also that the conduct of these analyses should be supported by methodology and tools. This will allow the modeler to focus more on the interpretation of modeling results and to assess which model fits the purpose of interest best, rather than to spend a lot of time on getting the best model and performing overly repetitive tasks. Another challenge with poppk analyses lays in the relatively large inter-modeler variability of how such analyses might be conducted, all due to the modeler s experience, expertise, and employed tools. Byon and colleagues [2] already pointed out the need for standard procedures, at least across the same organization, in order to achieve a more consistent high quality output of poppk analyses. There are many aspects that contribute to the quality of these analyses. Here in this paper a minimum level of quality is considered and defined by the following requirements: All poppk modeling activities within the same organization are performed following the same internal poppk modeling guidance. Analysis results should be independent of the modeler who is conducting the analysis. Code, models, data that were used for the analysis should be easily accessible, self-explaining, and executable by everyone without the need for interaction with the original modeler. In this paper it is assumed that an internal poppk modeling guidance is in place, covering not only all aspects of the published guidances from health authorities [11, 12], but also more in depth methodology and modeling workflow. Efficiency and quality of pharmacometric analyses have already been considered throughout the literature. Bonate and colleagues [13] present detailed guidelines for quality control of population pharmacometric analyses, with a focus on NONMEM [14]. However, many of the discussed elements are equally applicable or adaptable with respect to other tools. An example for the consideration of efficiency is the work by Jönsson and Jonsson [10], which discusses important requirements for efficient conduct of general pharmacometric activities, including planning, execution, and reporting. Focusing more on the execution part of poppk modeling activities, other authors acknowledged the need for flexible integration of the extensive choice of software platforms and available tools (see, e.g., [15] ). Bergsma and colleagues [16] presented an implementation of such an integrated framework in R language [17] and Ridolfi and co-workers [18] proposed a web-based approach [18]. Both [16] and do not cover the aspect of a modeling guidance or

3 modeling methodology, which is critical when considering consistency and/or quality of results across an organization. Automatic poppk model generation is often associated with the term efficiency and some efforts were made in the past to achieve this target. Schaap et al. [19] used a genetic algorithm approach for automatic model building, still preserving the freedom of a modeler to take certain decision in between. However, main drawbacks of general optimization based approaches to searching the space of candidate models are the definition of the cost-function and the definition of the model space to search at a given time. A recent contribution from Byon and colleagues [2] presents a detailed poppk modeling guidance, focusing on systematic, streamlined, and standardized approaches to poppk modeling in order to optimize and harmonize the processes across their organization. However, the authors do not go into more detail regarding the support of their poppk workflow with associated workflow tools. It is important to realize, however, that a workflow based only on poppk guidance and definition of the methodology, might not be the most efficient if not supported by workflow tools. Equally, a platform that integrates all available tools might not lead to consistent high quality modeling results if no guidance exists on how to perform the analysis. This paper presents a general approach that addresses the linking of a modeling guidance with efficient workflow tools, to allow for an increase in efficiency and quality of the conduct of poppk analyses. This approach can be considered not only tool independent, but also independent on a specific modeling methodology or poppk guidance. Implementation is possible in different scripting languages, with different nonlinear mixed effect modeling tools at the back-end. Nevertheless, in order to demonstrate the usefulness of the approach, a specific platform had to be chosen for a first implementation of the workflow tools. In this case the platform of choice consists of MATLAB [20], SBPOP [21], MONOLIX [22] and NONMEM [14]. Depending on the preferences of an organization, modifications can be done on all parts. This paper is structured as follows: section Definition of considered poppk model features summarizes the results of an internal survey about the most common features in poppk models, allowing definition of the features that a workflow minimally needs to cover. A generalized dataset specification is presented in section Generalized dataset format, allowing to capture the modeling relevant information, without being tool or modeling task specific. The modeling workflow and associated workflow tools are presented in section PopPK modeling workflow and workflow tools.this is followed by a discussion and conclusion. A complete example poppk analysis on simulated data, using the presented workflow with both NONMEM and MONOLIX, is made available in the supplementary material, allowing the Table 2 Selected poppk model/modeling features that are considered in the workflow Model feature Type of model Number of compartments Model outputs Elimination Bioavailability Absorption Absorption delay Distribution Individual parameter distributions Residual error models Fixed and random effects Covariance model Covariate model Options considered in workflow Compartmental models (binding models out of scope) 1, 2, 3 Single Linear and linear? saturable (Michaelis Menten) from central compartment, time independent Linear First-order Lag-time Linear Logit-normal for relative bioavailability, lognormal for other PK parameters Additive, proportional, additive/proportional, exponential Allowing for estimation and fixing of values for selected parameters Diagonal, block-diagonal and full correlation matrices for random effects Allowing estimation and fixing of selected covariate coefficients interested reader to repeat the example analysis and to adapt it to own projects. The supplementary material is made available as a zip file. The contents can be accessed by opening the file Supplementary Material.html in a web browser of choice. Definition of considered poppk model features Pharmacometric modeling is often considered to be an art, and a typical argument against a more industrialized form of model building is that all models are different. However, more concrete data on how different poppk models can be is generally not available. Prior to embarking on the development of the presented poppk modeling workflow and associated tools it is important to characterize the poppk model space of interest and to select a certain subspace to be considered by the workflow. The chosen approach to determine the typical elements of poppk models was to conduct an internal poppk model survey by interviews with colleagues in the Advanced Quantitative Sciences group (AQS, former Modellng&Simulation). The limitation to an internal survey, rather than conducting a full literature review, was intentional and allows focusing on fit-for-purpose models that

4 Table 3 Format of generalized dataset Dataset section Column name Definition Type Identification of USUBJID Unique subject identifier String subject STUDY Study number Numerical CENTER Centre number Numerical SUBJECT Subject number Numerical Study information PART Part of study (if parts present, otherwise 0) Numerical INDICATION Indication flag Numerical INDICATION_NAME Indication name String Treatment group TRT Unique treatment identifier Numerical information TRT_NAME Analyst given treatment name String Visit information VISIT Visit number Numerical VISIT_NAME Visit name as coded in the clinical database String BASE Flag indicating assessments at baseline (= 0 for non-baseline, = 1 for first Numerical baseline, 2 = for second baseline, etc.) SCREEN Flag indicating assessments at screening (= 0 for non-screening, = 1 for first Numerical screening, 2 = for second screening, etc.) Event time DATE_DAY Date of event String information DATE_TIME Time of event String NOMINAL_TIME Planned time of event, based on protocol in the time unit, defined in TIME_UNIT Numerical column TIME Actual time of event in the time unit, defined in TIME_UNIT column. Time 0 Numerical defined as time of first active or placebo dose TIME_UNIT Unit of all numerical time definitions in the dataset (e.g., hours or days ) String Event value TYPE See Table 4 Numerical information TYPE_NAME See Table 4 String SUBTYPE See Table 4 Numerical VALUE See Table 4 Numerical VALUE_TEXT Text version of value (e.g. male or female for gender ) String UNIT Unit of the value reported in the VALUE column. For same event the same unit has String to be used across the dataset NAME See Table 4 String DURATION See Table 4 Numerical ULOQ See Table 4 Numerical LLOQ See Table 4 Numerical Dose event additional information ROUTE See Table 4 String INTERVAL Interval of dosing in time units, defined in the TIME_UNIT column allows for Numerical coding repeated dosing more efficiently NR_DOSES Number of doses given with the specified interval allows for coding repeated dosing more efficiently Numerical are developed on a day-to-day basis to support decision making in internal drug development projects. Interviews were conducted either personally, via phone, or . The focus was only on the models and their features and not on the modeling methodology nor on the tools used for the modeling. The survey resulted in data for 58 poppk modeling activities (43 different compounds, otherwise data from different development phases and/or different modeling approaches compartmental/binding), of which 6 were using binding models and 52 were compartmental models. Table 1 summarizes the results of the survey in terms of the relative frequency for poppk model elements for the 52 compartmental models. Additional observations from the survey: All used proportional, additive/proportional, or exponential residual error models. Individual PK parameters are always assumed to be lognormally distributed, except in some cases the bioavailability for which a logit-normal distribution was used.

5 Fixed and random effect parameters and covariate coefficients were estimated or fixed to specific values, based on assumptions or prior knowledge. Transit compartments were used only on data with very rich sampling in the absorption phase, which sometimes was available in phase I studies. Based on the survey, poppk model features to be considered for the development of the workflow and the associated workflow tools were identified and listed in Table 2. Based on the selection in Table 2, % of the possible model features in each category in Table 1 are considered. Although this might still seem somewhat limited, this selection itself does not impact the generality of the presented workflow approach. Focusing on the elements of main importance allowed a first development of the workflow and implementation of associated tools. Extension of the workflow to additional model features is straight-forward and can be considered on a need basis. Generalized dataset format Within AQS, modeling datasets for the conduct of pharmacometric analyses are usually prepared by the modeler him/herself or requested from a supporting programming group. The content and the structure of such datasets are defined in a dataset specification, which typically depends on the modeling activity, the modeling tool, the models to be assessed, and even on the modeler. In the context of model-based drug development often many different analyses, or pharmacometric activities, need to be performed within the same project. This includes, but is not limited to, graphical exploration, doseconcentration and concentration response modeling, where many different biomarkers or endpoints might need to be considered. Requesting a separate modeling dataset for each modeling activity considerably strains the resources of the modeler who prepares these datasets or of the supporting programming group, especially in the case when a certain level of validation is required. In this paper, a generalized dataset format is proposed (Table 3), covering PK as well as pharmacodynamic (PD) data. Requirements for this dataset format have been a welldefined structure that is project independent (same across projects, compounds, and indications), independent of the modeling activity to be performed, and independent of the modeling tool to be used. Additionally, the format had to have a minimum level of redundancy and include information that is typically not found within pharmacometric analysis datasets: names for variables, units, etc. This renders the dataset more understandable without additional documentation, reduces mistakes, and facilitates project hand-over processes. The presented dataset format can be seen as an extension to the notion of a master modeling dataset, discussed in [10]. Each row in this dataset corresponds to an event, i.e., to a given dose or an observation (compound or metabolite concentration, biomarker or endpoint for safety or efficacy, adverse event). Extensions to other types of events are possible but not in scope of this paper. The standardized structure allows for more efficient and less error-prone preparation of datasets, pooling across studies, easier understanding and re-use of the data, automatic generation of the most common graphical exploration plots, and automatic conversion to modeling activity and potentially tool dependent modeling datasets. Many of the data consistency checks that are suggested in [13] can be done automatically based on such a general data format. To seamlessly integrate this format into the presented poppk workflow, a minimum set of standards need to be defined (Table 4). The main drawback of such a master dataset approach is the non-negligible size in large drug development projects. It cannot be excluded that for very specific poppk analyses manual modifications of an automatically generated modeling dataset need to be done. However, all poppk modeling features considered in this paper (Table 2) are captured by the approach of a generalized dataset format. Additional features can be handled when needed, by revising and extending the format, including updating of the relevant workflow tools. PopPK modeling workflow and workflow tools Figure 1 shows a high level poppk workflow. The only difference to traditional approaches is the use of the generalized dataset format, rather than starting directly from a tool and modeling activity dependent modeling dataset. In this section, the different steps of the workflow and their support by workflow tools are discussed in more detail. Due to practicability, the workflow tools cannot be exemplified based on a complete poppk example throughout this paper. Pseudo-code and summary results will be shown instead, where applicable. For a complete overview of the workflow tools and their functioning, the reader is referred to the supplementary material. As with the model features in Table 1, a workflow tool based approach should not and cannot aim at covering 100 % of the possible analyses or features. However, this does not represent a limitation of the approach. At each stage the modeler can perform additional analyses, if these are needed. Over time, if an analysis that is not yet covered becomes more often used, it could be included into the workflow tool.

6 Table 4 Standards in the dataset required for the presented poppk workflow Column name Definitions and conventions Standards for poppk workflow TYPE TYPE_NAME Numerical value for grouping of certain types of events. Required groups are 0 (for dosing events) and 1 (for PK observations). Additional groups could consist of demographics, metabolites, target capture, mechanism of action (MoA) biomarkers, clinical endpoints, etc Textual representation of the corresponding TYPE value. No standard defined. For dosing events Dose might be used and for PK observation events PK TYPE = 0 corresponds to dosing events, and TYPE = 1to PK measurement events SUBTYPE Numerical identifier for a specific event within a group If TYPE = 0: SUBTYPE = 1 corresponds to dosing events of the first compound, SUBTYPE = 2 to the second, etc If TYPE = 1: SUBTYPE = 1 corresponds to the observed PK concentration of the first compound, SUBTYPE = 2 to the second, etc VALUE NAME DURATION ULOQ LLOQ ROUTE Value of the event, defined by TYPE and SUBTYPE. E.g., the given dose, the observed PK concentration, or the value of other readouts. The values need to be in the units, defined in the UNIT column Unique name for the event that is coded by TYPE and SUBTYPE. No standard defined. For dosing events Dose compound X and for PK observations X plasma concentration might be used Numeric value capturing the duration of an event, in the time units defined in the column TIME_UNIT Upper limit of quantification of event defined by TYPE and SUBTYPE (if applicable), empty if not Lower limit of quantification of event defined by TYPE and SUBTYPE (if applicable), empty if not For the conversion to a model specific modeling dataset the route of administration needs to be defined for dosing events None None None If TYPE = 0 (dosing event), this column needs to contain the duration of the administration (e.g., the infusion time). 0 thus means bolus administration None If TYPE = 1: lower limit of quantification for the concentration of compound defined by SUBTYPE The following identifiers are recognized by the current poppk workflow tools: IV Subcut Oral Graphical exploration Prior to any modeling activity the data needs to be explored and understood. As with the model features (Table 1), there are some graphical analyses that should always be carried out and there are others that are more rarely done or are very specific to a project. Modelers typically develop their own scripts which then are adapted to each new modeling dataset to generate the needed graphical exploration plots. Since each modeler has a favorite tool for implementation of these scripts, established coding preferences, personal aesthetic criteria (e.g., how to label the axes, which colors to use), etc., this approach is not the most efficient and leads to a considerable variability of resulting plots. A more efficient approach to graphical exploration is to: (1) define upfront the plots/analyses that at least are needed in every poppk modeling activity, (2) develop a workflow tool that allows producing all these plots with a single function call, including elements such as high quality representation, axes labelling, and storing the resulting output in a reasonably organized structure. A critical requirement to allow for automatic generation of these analyses is the use of a standardized dataset format, which here is the previously defined generalized dataset format. A pseudo-code example for the graphical exploration analysis is shown below: In this example, the input to the workflow tool/function (ExplorePopPKdata) is the dataset in the generalized dataset format (data), the names of components in the dataset that are to be considered as potential covariates (covariates), and a path to where to write the output

7 Fig. 1 High level poppk modeling workflow Generalized dataset format Graphical exploration Data cleaning Modeling dataset generation Model building Final model (outputpath). The output of this workflow tool is several files in the output folder. Examples for the generated information are: Summary plots of the individual data on linear and logarithmic Y axis. Individual plots of the individual data on linear and logarithmic Y axis, indicating dosing times and drug amounts. Each observation record is annotated by the index in the dataset and by its value and time after last dose. Summary statistics for the chosen covariates. Spaghetti diagrams stratified by treatment arm and study, dose normalized and non-dose normalized, and both over time since first dose (TIME) and since last dose (TAD). Histograms and correlation plots for the continuous covariates and correlation information for categorical and continuous covariates. Data cleaning Exploration of PK data often leads to the identification of PK samples that might be inconsistent with the rest of the data, resulting in the exclusion of single PK records or all records for certain subjects from consideration. Since the generalized dataset often does not only contain PK but also PD data, subjects receiving only placebo doses should be removed for poppk modeling. Covariates of interest are not always recorded in all subjects, requiring imputation of missing ones. Depending on the decision of a modeler, records with concentration samples that are below the limit of quantification might be removed from the dataset or treated in a different way. There are many additional elements of data cleaning that are not considered in here. Important is that the step of data cleaning is often done in a modeler-dependent fashion. Some modelers might request from a programmer to flag the records and subjects to be removed from the subsequent analysis within the dataset, leading to multiple iterations on the original dataset. Others might handle it themselves by scripting or manual removal. Some modelers might request imputation of covariates to be done during the dataset preparation; others do it on their own. Rules for imputation of covariates might also vary across modelers and programmers. All these different possible approaches can lead to considerable inter-modeler variability and a reduction of efficiency. A workflow tool that supports a modeling methodology/ guidance allows applying more consistent rules for data cleaning across an organization, leading to increased quality, and can allow for a faster turnaround of the data cleaning. The output of this workflow tool should be a cleaned dataset to which the desired transformations have been applied. Additionally, a record of the applied transformations should be generated for documentation purposes in a well-organized structure. After this step it might be useful to re-run the graphical exploration analyses on the cleaned dataset, for documentation and checking of the results. A pseudo-code example for the data cleaning step is shown below:

8 In this example, the input to the workflow tool/function (CleanPopPKdata) is the dataset in the generalized dataset format (data), the names of components in the dataset that are to be considered as potential covariates (covariates), the rules for imputation of missing covariates (imputationrules), the unique subject identifiers of the subjects to be removed and the reason for removal (removesubjects), the index of records in the dataset that are to be removed and the reason for removal (removerecords), a flag that indicates on how to treat data below the limit of quantification (FLAG_LLOQ) and a path to where to write the log files containing information about the cleaning steps (outputpath). The output argument of this workflow tool is the cleaned dataset (datacleaned). Different options for the FLAG_LLOQ might be available. In the actual implementation of the workflow tool, a value of 0 removes the data below the LLOQ from the dataset, while a value of 1 keeps them and creates a new column with information about censored data. Additional approaches to handle data below the LLOQ can be implemented in a similar manner, allowing efficient adaptation of the data to the needs of a modeler. Base model Fig. 2 Model building workflow Covariance model Covariate model Final model selection software used by the workflow tools. By employing the generalized dataset format this step can be completed transparently to the user, meaning that the modeler does not need to be an expert in creating datasets for specific NLME software. The minimum information required by the workflow tool for this step is the generalized dataset format and a list with names of readouts that should be included in the modeling dataset as candidate covariates. A pseudo-code example for the modeling dataset generation: Modeling dataset generation Up to this point in the poppk modeling workflow, the format of the dataset was independent of any non-linear mixed effect modeling (NLME) software. However, for the purpose of parameter estimation, the dataset needs to be converted into a format that is suitable for the NLME In this example, the input to the workflow tool/function (Convert2popPKdataset) is the cleaned dataset that has been returned by the data cleaning step (data), the names of components in the dataset that are to be added as columns in the modeling dataset (covariates), and a path with filename to where to write the generated modeling dataset (outputfile). The output argument

9 Fig. 3 Efficient approach to evaluate a certain poppk model sub-space of interest User defined model space One step model sub-space evaluation Model generation Parameter estimation Creation of GoF plots Creation of tables for comparison Results for selected model space (convertinfo) of this workflow tool might contain information that is required for model generation step and is passed as an input argument to this step (see next pseudo-code example). In the actual implementation of this workflow tool, the output consists of a string that describes the type of columns in the modeling dataset. Model building: general steps Figure 2 shows the four main steps of poppk model building, considered in this paper. As discussed above, the goal is not to define how these different steps should be done (this should be covered in the modeling guidance), but to show an efficient way of how a workflow tool can support the modeler in the exploration of the model space of interest. The base model building step considers the underlying structural model, the residual error model, and fixed and random effects to be estimated. Certain selected covariates might already be considered if there is a rational for doing so. The covariance model building step focuses on a refinement of the decision regarding the selection of random effects and the correlations between random effects to be estimated. Assessment of statistical significance and potential clinical relevance of candidate covariates is done in the covariate model building step, while the final model selection step might assess clinical relevance of covariates more in detail, resulting in the selection of the final poppk model. The inclusion of covariates into a model might have an impact on the identifiability of the covariance matrix for the random effects. This means that the modeler needs to decide in which order to execute the covariance model building and the covariate model building. The example in the supplementary material demonstrates that the order of these steps indeed matters. Model building: one step model sub-space evaluation Independently of the current model building step of consideration and of the approach (manual or workflow tool supported poppk modeling), the usual steps to be taken are to (1) select a model to test, (2) generate the model code, (3) run the estimation algorithm, and (4) process the results to produce goodness-of-fit assessments, visual predictive checks (VPCs), simulations, etc. Based on the obtained results, alternative models are selected and the whole process is repeated, eventually exploring the space of poppk models to find the model that is best supported by the available data and the prior knowledge about the compound. In comparison to the approach where all models are coded and executed manually, potentially requiring switching between different tools and computational environments to produce goodness of fit plots etc., a workflow tool based approach can considerably enhance this process, as outlined in Fig. 3. The modeler defines a model sub-space of interest to be assessed, using a workflow tool dependent syntax. Then the workflow tool automatically generates the relevant models, performs the parameter estimation, reads the results and generates goodness-of-fit plots and tables, allowing comparing the considered models all in one step. Such a workflow tool should at least cover all model features and options, listed in Table 2. A pseudo-code example for a one-step model sub-space evaluation for a base model:

10 In this example, the input to the workflow tool/function (PopPKModelSpace) has two parts. The first part (input1) defines general settings, where stepidentifier is an identifier that is used, e.g., as part of the path where the generated models and outputs are stored, convertinfo is the output of the modeling dataset generation step, data defines the path and filename of the modeling dataset, and options are a set of user defined options for the underlying estimation software. The second input (input2) defines the poppk model features to be tested. The argument parameters contains information about initial guesses for the fixed and random effects. Additionally, using the parameters argument, the modeler can define different scenarios, each defining which fixed and random effects to be estimated or kept fixed. Once this workflow tool is executed, it will generate a sub-space of poppk models that is defined by the combination of all model features with each other. This means that in the example above, 36 different poppk models multiplied by the number of parameters scenarios, the modeler has chosen, are generated. A good practice in the first base model iteration is to limit the number of scenarios to one, by assuming that all fixed and all random effects are estimated, or to select a reasonable setting through insight based on the available data. Refinement of this assumption can be done after the consideration of the estimation results. After the generation of all models the parameter estimations are performed. Parallel processing is used, if available. Subsequently, the most common goodness-of-fit (GoF) plots are generated for all evaluated models and stored in an ordered fashion. Additionally, tables for comparison are generated as text files, summarizing the results for the models (parameter estimates and some metrics of interest) in the user defined model space, allowing the modeler to get a quick overview of the performance of the different models and support decisions about the next modeling steps. Examples, covering such tables are shown and discussed more in detail in the supplementary material. Model building: technical requirements for the one step evaluation In order to be able to automatically generate models of interest, a model library is required. Since any model library is of finite size, the properties of the models in this library need to be defined upfront. An alternative to a model library is a general model that contains all the considered structural elements, representing the entirety of a model library by switching on and off certain features. A model structure, capturing all structural properties of the models defined by Table 2 is shown in Fig. 4. A dual approach has been taken in the implementation of the workflow tool, in order to improve the computational performance. Models that are linear and analytically solvable are taken from a model library, while non-linear models are implemented by using the general model in the form of ordinary differential equations (ODEs). PK parameters are always assumed to be log-normally distributed, except the bioavailability, which is assumed to be logit-normally distributed. In order to realistically pursue the path of such a workflow tool based exploration of poppk model subspaces, an important key assumption needs to be made: the availability of numerically stable estimation algorithms that, given a reasonably well defined estimation problem, are able to produce the output of interest in a reliable manner. The authors experience suggests that the implementations of the SAEM algorithm in both MONOLIX and NONMEM do meet this requirement. Other methods, such as FOCE in NONMEM do not generally qualify for the workflow based approach. Suitable initial parameter guesses and model parameterizations might need to be found, to allow for convergence of the estimation algorithm. In case of failure of the estimation algorithm, the error messages need to be assessed very carefully in order to determine needed changes in the model. This is a time consuming process which is counterproductive for efficiency and the outcome is often that the modeler needs to focus more on understanding the tool

11 Fig. 4 General structural poppk model Tlag Dosing Dosing 1 Fabs1 Fiv Dosing 2 compartment ka Peripheral compartment 1 (Vp1) Q1 Central compartment (Vc) Q2 Peripheral compartment 2 (Vp2) CL, VMAX, KM and possible workarounds, rather than on the interpretation of estimation results. Model building: hierarchical exploration of the model space An important factor in the workflow based approach is the selection of the model sub-space to be considered. Theoretically, it would be possible to consider the complete poppk model space, spanned by combination of all model features and options, shown in Table 2. However, this would lead to a tremendous combinatorial explosion of models to be evaluated and compared. Even though this might still be technically feasible, given enough computational resources and/or time, it is out of scope of this paper. Instead, the approach advocated here is to use a hierarchical/modular approach, following the steps in Fig. 2. After each step the modeler assesses the results and refines the model features to be considered in the analysis. In a first step, a base model is sought that adequately describes the data. A good guess about a possible model structural elements and to include selected covariates only if a good rational exists. An example for the first base model assessment would be to consider different numbers of compartments, selected error models, possible lag time on absorption, and linear and linear? saturable elimination (see previous pseudo-code example). All fixed and random effects could be estimated. The base model is then refined in a second step. Based on the results from the first step, the main structural elements for which there is support in the data are better understood (e.g., number of compartments, type of elimination) and the model space with respect to these elements can be reduced. Random effects for which the relative standard errors and/or the shrinkage are very large might be excluded from further consideration. Several different parameter scenarios could be considered in this second step, comparing the impact of different combinations of parameters for which to estimate fixed and random effects. This steps leads to the selection of the final base model. A pseudo-code example for a one-step model sub-space evaluation for a base model refinement: space for the base model should be already available based on the graphical exploration of the available data and/or from prior knowledge about the compound. It is useful to limit the number of model features and options to the main In this example, the same workflow function (PopPK- ModelSpace) is used as in the first base model assessment, but with a reduced selection of possible model features to be

12 tested. Using the parameters argument, three different scenarios are defined, testing different sets of random effects in the models (parameter names defined in Fig. 4). In a third step, the covariate model is addressed. Several possible methods are available in the literature and employed in day-to-day work. In the context of an efficient poppk workflow it is of less importance which methods are used, as long as they identify the clinically relevant covariates. What is more important is that this method is/ these methods are implemented and available as workflow tool(s) in a user-friendly and seamless way. The method used in the actual implementation of the workflow is the Full Model Estimation Approach [23]. A pseudo-code example for a one-step model sub-space evaluation for a covariate model: discussed here) to obtain an idea about the impact of the selected covariates on the PK parameters. Covariate coefficients can be estimated or fixed to initial guesses (default initial guesses are 0). Another important aspect is the transformation of covariates. Since lognormal distribution of individual PK parameters is assumed by the workflow tools, the covariates are automatically log transformed and centered by the median. The covariate step is discussed more in detail in the supplementary material. The covariance model is addressed in the fourth step of model building. As before, the best model is retained from the previous steps and the search space is expanded towards different structures of the random effect covariance matrix. A first guess about a suitable structure of the covariance matrix might be available, based on the information from the Again, the same workflow function (PopPKModel- Space) is used as in the previous examples. The model sub-space of consideration is spanned by the final base model and four different covariate models (covariate) to be assessed independently. The estimation results can be processed further with additional workflow tools (not previous models (i.e., correlation of individual parameter estimates). Additional scenarios should be considered, including a full covariance matrix. This step will lead to the selection of the final poppk model. A pseudo-code example for a one-step model sub-space evaluation for a covariance model:

13 The workflow tool/function in this step (PopPK- ModelSpace) is the same as in the previous example. The model sub-space of consideration is spanned by the final covariate model and 3 different covariance models (covariance) to be assessed independently. The total number of models, evaluated by the pseudocode examples above, is 46. The interpretation of the results by the modeler was needed in each step to select the model sub-space to be considered in the next step, resulting in a relatively small number of evaluated models. Alternatively, all the above tested model features could have been combined in a single call to the PopPKModelSpace workflow tool: Model building: additional analyses Additional analyses are of use and should be carried out at certain stages of model building. Often used ones are VPCs, both on data used for modeling and on other data for external validation, and bootstrapping to obtain a measure for uncertainty in the estimated parameters. For these analyses, user-friendly implementations should be available within the workflow supporting tools. For the purpose of model selection and rationale based decisions during the model building process, additional model based analyses are of interest and should be an integral part of the model building workflow. Two such analyses are: This single call generates 1,728 different candidate poppk models. On one hand this shows that such a workflow tool allows efficiently generating and evaluating a large number of models. On the other hand, it also shows that most of these models are not relevant and the modeler has the tremendous task of filtering them. It is important to note that the number 1,728 is still not the final, since some of the definitions above have been chosen based on the knowledge gained in previous steps. For example, the number of scenarios considered for the random effects might need to be increased in such a brute-force approach. Even though such an approach might have some academic value, the modeler guided step-wise model building approach is to be preferred. Robustness analysis Re-estimation of parameters for the same model but starting from randomized initial guesses. This supports the assessment of the actual identifiability of the estimated parameters. A model should be rejected if repeated estimation from different initial guesses leads to significant differences in the parameter estimates. In practical applications, this analysis has already given valuable information about the parameters that have most contribution to the over-parameterization by showing drifts. Such drifts are represented by correlations of two or more parameters with the objective function value, which can be explained by a shallow objective function surface in certain directions of the parameter space. The optimization algorithm might get stuck or shows a slower convergence in these directions.

14 In the actual implementation of the workflow tools, this analysis is available as option to the PopPKModel- Space function. Sensitivity analysis Here it is understood as using simulation based parametric sensitivity analysis to predict the expected information content in data that is not yet modelled (e.g., because a phase III study has not read out yet but planning for the poppk modeling needs to be done). Often, late phase studies have a very sparse PK sampling and re-estimation of all parameters might not be the best approach. Sensitivity analysis can be used as a scientifically valid motivation for fixing some parameters on results from previous analyses and only the parameters for which information in the new data is expected to be available might be re-estimated. Methods from the area of optimal design might be used for such an analysis, but with a focus on efficiency they would need to be made available as workflow tools, not requiring the recoding of the models into a different format. In the actual implementation of the workflow tools a simple parametric sensitivity analysis based on the population mean parameters has been chosen. Discussion and conclusion The landscape of available tools and the way on how population pharmacokinetic and pharmacodynamic (PKPD) modeling is often taught and performed supports the conclusion that the area of Pharmacometrics is very much tool-centric. Tremendous effort has been and is put into understanding and sharing tips-and-tricks on how to use certain tools rather than on modeling methodology and guidance. So far, the tool centricity might have limited the investigation of more streamlined and standard approaches to population modeling, associated workflows, and supporting these workflows by efficient integrated workflow tools. It is clear that a general approach to all possible pharmacometric population modeling activities is complex/ impossible to realize. However, when limiting the scope from all possible to very specific modeling activities, such as poppk, it can be shown that the type of models, model features and options are often quite similar. In this paper an attempt has been made to approach poppk model building in a workflow oriented way, supporting the workflow by efficient workflow tools. The center-piece of the presented approach is a workflow tool that, based on the definition of a poppk model sub-space of interest, generates all the corresponding models, performs the parameter estimation, and summarizes the results in a form that is easily accessible and interpretable by a modeler. One advantage of this approach is that all NLME parameter estimation tool specific issues can be hidden from the modeler, shifting the focus to defining the right model space to investigate and interpreting the results of the parameter estimation in order to select the next model features to test. Another key component of the presented work is the definition of a generalized dataset format. Such a format provides the basis for an efficient implementation of the poppk workflow. Not only can the most important graphical exploration plots of poppk data be generated automatically, but it also allows for more efficient dataset preparation, easier re-use of the data due to the additional annotation, and supports newcomers in the area of Pharmacometrics to get their dataset specification right the first time. The presented workflow has been used on a large number of poppk modeling activities, ranging from phase I to submission. In all these activities the workflow approach eliminated labor-intensive repetitive coding tasks, allowed a better focus on the modeling rather than the tools used, and led to a larger degree of trust in the final model and parameter estimates. The fact that the workflow tools ensure organized storage of models, model results, generated tables and graphics is a benefit not to be underestimated when collaborating with other modelers and model reviewers. To provide an approximate measure that can be linked to efficiency metrics, the example poppk (one single dose study and one multiple dose study, 86 subjects on active treatment, 1,331 observation and 1,166 dose records) in the supplementary material can be completed in the presented form in less than a day (without all the detailed documentation that was added for the purpose of this publication). In terms of quality metrics, the example poppk has been conducted, following the assumed underlying modeling guidance. Furthermore, the required code for the complete poppk is available as a single script that would be executable by every modeler without the need for interaction with the original author, ensuring reproducibility. All models and outputs are stored in an organized structure that allows finding the information of interest immediately. Although the presented work is limited to the poppk model features shown in Table 2, the approach is general and easily extensible. This means that, if needed, the approach can be extended to cover inter-occasion variability, zero-order and mixed absorption, nonlinear bioavailability, nonlinear inter-compartmental distribution, etc. For the proof-of-concept implementation and to gain experience with this approach a subset of model features had to be chosen. Additionally, when basing the workflow tools on a scripting language, it is possible to step out of the

Tips for the choice of initial estimates in NONMEM

Tips for the choice of initial estimates in NONMEM TCP 2016;24(3):119-123 http://dx.doi.org/10.12793/tcp.2016.24.3.119 Tips for the choice of initial estimates in NONMEM TUTORIAL Seunghoon Han 1,2 *, Sangil Jeon 3 and Dong-Seok Yim 1,2 1 Department of

More information

Predict Outcomes and Reveal Relationships in Categorical Data

Predict Outcomes and Reveal Relationships in Categorical Data PASW Categories 18 Specifications Predict Outcomes and Reveal Relationships in Categorical Data Unleash the full potential of your data through predictive analysis, statistical learning, perceptual mapping,

More information

ABSTRACT INTRODUCTION WHERE TO START? 1. DATA CHECK FOR CONSISTENCIES

ABSTRACT INTRODUCTION WHERE TO START? 1. DATA CHECK FOR CONSISTENCIES Developing Integrated Summary of Safety Database using CDISC Standards Rajkumar Sharma, Genentech Inc., A member of the Roche Group, South San Francisco, CA ABSTRACT Most individual trials are not powered

More information

3/8/2018. Agenda. NONMEM to NLME: 2-Compartment Oral Absorption Disease State Effect on V and CL FOCE. Comparison of Code Defining PK Model

3/8/2018. Agenda. NONMEM to NLME: 2-Compartment Oral Absorption Disease State Effect on V and CL FOCE. Comparison of Code Defining PK Model Agenda to : 2-Compartment Oral Absorption Disease State Effect on V and CL FOCE Model Description Exploratory Data Analysis Input Data Model in NM-TRAN and PML Code Demo Summary Q&A 2 : First Patient Structural

More information

How to write ADaM specifications like a ninja.

How to write ADaM specifications like a ninja. Poster PP06 How to write ADaM specifications like a ninja. Caroline Francis, Independent SAS & Standards Consultant, Torrevieja, Spain ABSTRACT To produce analysis datasets from CDISC Study Data Tabulation

More information

BioUML: plugin for population-based modeling

BioUML: plugin for population-based modeling BioUML: plugin for population-based modeling Kiselev I.N. 1,2, Kolpakov F.A. 1,2 1 Institute of Systems Biology, Ltd 2 Design Technological Institute of Digital Techniques SB RAS Abstract Motivation and

More information

Expectation-Maximization Methods in Population Analysis. Robert J. Bauer, Ph.D. ICON plc.

Expectation-Maximization Methods in Population Analysis. Robert J. Bauer, Ph.D. ICON plc. Expectation-Maximization Methods in Population Analysis Robert J. Bauer, Ph.D. ICON plc. 1 Objective The objective of this tutorial is to briefly describe the statistical basis of Expectation-Maximization

More information

OVERVIEW OF COMMENTS RECEIVED ON THE DRAFT GUIDELINE ON REPORTING THE RESULTS OF POPULATION PHARMACOKINETIC ANALYSES

OVERVIEW OF COMMENTS RECEIVED ON THE DRAFT GUIDELINE ON REPORTING THE RESULTS OF POPULATION PHARMACOKINETIC ANALYSES European Medicines Agency London, 21 June 2007 Doc. Ref. EMEA/CHMP/EWP/260830/2007 OVERVIEW OF COMMENTS RECEIVED ON THE DRAFT GUIDELINE ON REPORTING THE RESULTS OF POPULATION PHARMACOKINETIC ANALYSES Table

More information

Chapter 2 Overview of the Design Methodology

Chapter 2 Overview of the Design Methodology Chapter 2 Overview of the Design Methodology This chapter presents an overview of the design methodology which is developed in this thesis, by identifying global abstraction levels at which a distributed

More information

WELCOME! Lecture 3 Thommy Perlinger

WELCOME! Lecture 3 Thommy Perlinger Quantitative Methods II WELCOME! Lecture 3 Thommy Perlinger Program Lecture 3 Cleaning and transforming data Graphical examination of the data Missing Values Graphical examination of the data It is important

More information

Inclusion of Aleatory and Epistemic Uncertainty in Design Optimization

Inclusion of Aleatory and Epistemic Uncertainty in Design Optimization 10 th World Congress on Structural and Multidisciplinary Optimization May 19-24, 2013, Orlando, Florida, USA Inclusion of Aleatory and Epistemic Uncertainty in Design Optimization Sirisha Rangavajhala

More information

Correctly Compute Complex Samples Statistics

Correctly Compute Complex Samples Statistics SPSS Complex Samples 15.0 Specifications Correctly Compute Complex Samples Statistics When you conduct sample surveys, use a statistics package dedicated to producing correct estimates for complex sample

More information

Frequently Asked Questions Updated 2006 (TRIM version 3.51) PREPARING DATA & RUNNING TRIM

Frequently Asked Questions Updated 2006 (TRIM version 3.51) PREPARING DATA & RUNNING TRIM Frequently Asked Questions Updated 2006 (TRIM version 3.51) PREPARING DATA & RUNNING TRIM * Which directories are used for input files and output files? See menu-item "Options" and page 22 in the manual.

More information

Submission Guidelines

Submission Guidelines Submission Guidelines Clinical Trial Results invites the submission of phase I, II, and III clinical trials for publication in a brief print format, with full trials results online. We encourage the submission

More information

Pooling Clinical Data: Key points and Pitfalls

Pooling Clinical Data: Key points and Pitfalls Paper DH01 Pooling Clinical Data: Key points and Pitfalls Florence Buchheit, Novartis, Basel, Switzerland ABSTRACT Pooling of clinical data is used by all Pharmaceutical companies. Submission to Health

More information

Maximizing Statistical Interactions Part II: Database Issues Provided by: The Biostatistics Collaboration Center (BCC) at Northwestern University

Maximizing Statistical Interactions Part II: Database Issues Provided by: The Biostatistics Collaboration Center (BCC) at Northwestern University Maximizing Statistical Interactions Part II: Database Issues Provided by: The Biostatistics Collaboration Center (BCC) at Northwestern University While your data tables or spreadsheets may look good to

More information

Main challenges for a SAS programmer stepping in SAS developer s shoes

Main challenges for a SAS programmer stepping in SAS developer s shoes Paper AD15 Main challenges for a SAS programmer stepping in SAS developer s shoes Sebastien Jolivet, Novartis Pharma AG, Basel, Switzerland ABSTRACT Whether you work for a large pharma or a local CRO,

More information

Powering Knowledge Discovery. Insights from big data with Linguamatics I2E

Powering Knowledge Discovery. Insights from big data with Linguamatics I2E Powering Knowledge Discovery Insights from big data with Linguamatics I2E Gain actionable insights from unstructured data The world now generates an overwhelming amount of data, most of it written in natural

More information

IBM SPSS Categories. Predict outcomes and reveal relationships in categorical data. Highlights. With IBM SPSS Categories you can:

IBM SPSS Categories. Predict outcomes and reveal relationships in categorical data. Highlights. With IBM SPSS Categories you can: IBM Software IBM SPSS Statistics 19 IBM SPSS Categories Predict outcomes and reveal relationships in categorical data Highlights With IBM SPSS Categories you can: Visualize and explore complex categorical

More information

1 Topic. Image classification using Knime.

1 Topic. Image classification using Knime. 1 Topic Image classification using Knime. The aim of image mining is to extract valuable knowledge from image data. In the context of supervised image classification, we want to assign automatically a

More information

CFD Project Workflow Guide

CFD Project Workflow Guide CFD Project Workflow Guide Contents Select a problem with known results for proof-of-concept testing... 1 Set up and run a coarse test case... 2 Select and calibrate numerical methods... 3 Minimize & quantify

More information

ehepqual- HCV Quality of Care Performance Measure Program

ehepqual- HCV Quality of Care Performance Measure Program NEW YORK STATE DEPARTMENT OF HEALTH AIDS INSTITUTE ehepqual- HCV Quality of Care Performance Measure Program USERS GUIDE A GUIDE FOR PRIMARY CARE AND HEPATITIS C CARE PROVIDERS * * For use with ehepqual,

More information

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation Optimization Methods: Introduction and Basic concepts 1 Module 1 Lecture Notes 2 Optimization Problem and Model Formulation Introduction In the previous lecture we studied the evolution of optimization

More information

SELECTION OF A MULTIVARIATE CALIBRATION METHOD

SELECTION OF A MULTIVARIATE CALIBRATION METHOD SELECTION OF A MULTIVARIATE CALIBRATION METHOD 0. Aim of this document Different types of multivariate calibration methods are available. The aim of this document is to help the user select the proper

More information

PharmaSUG Paper SP08

PharmaSUG Paper SP08 PharmaSUG 2012 - Paper SP08 USING SAS TO CALCULATE NONCOMPARTMENTAL URINE PARAMETERS Vanessa Rubano, Boehringer Ingelheim Pharmaceuticals Inc., Ridgefield, CT Modesta Wiersema, Boehringer Ingelheim Pharma

More information

Tutorial #1: Using Latent GOLD choice to Estimate Discrete Choice Models

Tutorial #1: Using Latent GOLD choice to Estimate Discrete Choice Models Tutorial #1: Using Latent GOLD choice to Estimate Discrete Choice Models In this tutorial, we analyze data from a simple choice-based conjoint (CBC) experiment designed to estimate market shares (choice

More information

Latent Variable Indirect Response Modeling of Continuous and Categorical Clinical Endpoints

Latent Variable Indirect Response Modeling of Continuous and Categorical Clinical Endpoints Latent Variable Indirect Response Modeling of Continuous and Categorical Clinical Endpoints Scientific Director Model Based Drug Development Johnson & Johnson PAGE 2013 Overview Indirect response (IDR)

More information

Working with Composite Endpoints: Constructing Analysis Data Pushpa Saranadasa, Merck & Co., Inc., Upper Gwynedd, PA

Working with Composite Endpoints: Constructing Analysis Data Pushpa Saranadasa, Merck & Co., Inc., Upper Gwynedd, PA PharmaSug2016- Paper HA03 Working with Composite Endpoints: Constructing Analysis Data Pushpa Saranadasa, Merck & Co., Inc., Upper Gwynedd, PA ABSTRACT A composite endpoint in a Randomized Clinical Trial

More information

The IDN Variant TLD Program: Updated Program Plan 23 August 2012

The IDN Variant TLD Program: Updated Program Plan 23 August 2012 The IDN Variant TLD Program: Updated Program Plan 23 August 2012 Table of Contents Project Background... 2 The IDN Variant TLD Program... 2 Revised Program Plan, Projects and Timeline:... 3 Communication

More information

Keywords: DSM, DMM, Process simulation, product development, process planning

Keywords: DSM, DMM, Process simulation, product development, process planning 0 TH INTERNATIONAL DESIGN STRUCTURE MATRIX CONFERENCE, DSM 0 NOVEMBER 00, STOCKHOLM, SWEDEN SIMULATION OF PRODUCT CHANGE EFFECTS ON THE DURATION OF DEVELOPMENT PROCESSES BASED ON THE DSM Thomas Gärtner,

More information

186 Statistics, Data Analysis and Modeling. Proceedings of MWSUG '95

186 Statistics, Data Analysis and Modeling. Proceedings of MWSUG '95 A Statistical Analysis Macro Library in SAS Carl R. Haske, Ph.D., STATPROBE, nc., Ann Arbor, M Vivienne Ward, M.S., STATPROBE, nc., Ann Arbor, M ABSTRACT Statistical analysis plays a major role in pharmaceutical

More information

Data Curation Profile Human Genomics

Data Curation Profile Human Genomics Data Curation Profile Human Genomics Profile Author Profile Author Institution Name Contact J. Carlson N. Brown Purdue University J. Carlson, jrcarlso@purdue.edu Date of Creation October 27, 2009 Date

More information

An introduction to plotting data

An introduction to plotting data An introduction to plotting data Eric D. Black California Institute of Technology February 25, 2014 1 Introduction Plotting data is one of the essential skills every scientist must have. We use it on a

More information

Higher National Unit specification: general information. Graded Unit 2

Higher National Unit specification: general information. Graded Unit 2 Higher National Unit specification: general information This Graded Unit has been validated as part of the HND Computing: Software Development. Centres are required to develop the assessment instrument

More information

Pooling strategy of clinical data

Pooling strategy of clinical data Pooling strategy of clinical data Abraham Yeh, Xiaohong (Grace) Zhang, Shin-Ru Wang, Novartis Pharmaceuticals Corporation, East Hanover, NJ ABSTRACT Pooling of clinical data is used by all pharmaceutical

More information

How to review a CRF - A statistical programmer perspective

How to review a CRF - A statistical programmer perspective Paper DH07 How to review a CRF - A statistical programmer perspective Elsa Lozachmeur, Novartis Pharma AG, Basel, Switzerland ABSTRACT The design of the Case Report Form (CRF) is critical for the capture

More information

Control Invitation

Control Invitation Online Appendices Appendix A. Invitation Emails Control Invitation Email Subject: Reviewer Invitation from JPubE You are invited to review the above-mentioned manuscript for publication in the. The manuscript's

More information

Graph Theory for Modelling a Survey Questionnaire Pierpaolo Massoli, ISTAT via Adolfo Ravà 150, Roma, Italy

Graph Theory for Modelling a Survey Questionnaire Pierpaolo Massoli, ISTAT via Adolfo Ravà 150, Roma, Italy Graph Theory for Modelling a Survey Questionnaire Pierpaolo Massoli, ISTAT via Adolfo Ravà 150, 00142 Roma, Italy e-mail: pimassol@istat.it 1. Introduction Questions can be usually asked following specific

More information

The Analysis and Proposed Modifications to ISO/IEC Software Engineering Software Quality Requirements and Evaluation Quality Requirements

The Analysis and Proposed Modifications to ISO/IEC Software Engineering Software Quality Requirements and Evaluation Quality Requirements Journal of Software Engineering and Applications, 2016, 9, 112-127 Published Online April 2016 in SciRes. http://www.scirp.org/journal/jsea http://dx.doi.org/10.4236/jsea.2016.94010 The Analysis and Proposed

More information

Preparing for Data Analysis

Preparing for Data Analysis Preparing for Data Analysis Prof. Andrew Stokes March 21, 2017 Managing your data Entering the data into a database Reading the data into a statistical computing package Checking the data for errors and

More information

CDASH Standards and EDC CRF Library. Guang-liang Wang September 18, Q3 DCDISC Meeting

CDASH Standards and EDC CRF Library. Guang-liang Wang September 18, Q3 DCDISC Meeting CDASH Standards and EDC CRF Library Guang-liang Wang September 18, 2014 2014 Q3 DCDISC Meeting 1 Disclaimer The content of this presentation does not represent the views of my employer or any of its affiliates.

More information

Formal Verification for safety critical requirements From Unit-Test to HIL

Formal Verification for safety critical requirements From Unit-Test to HIL Formal Verification for safety critical requirements From Unit-Test to HIL Markus Gros Director Product Sales Europe & North America BTC Embedded Systems AG Berlin, Germany markus.gros@btc-es.de Hans Jürgen

More information

Module I: Clinical Trials a Practical Guide to Design, Analysis, and Reporting 1. Fundamentals of Trial Design

Module I: Clinical Trials a Practical Guide to Design, Analysis, and Reporting 1. Fundamentals of Trial Design Module I: Clinical Trials a Practical Guide to Design, Analysis, and Reporting 1. Fundamentals of Trial Design Randomized the Clinical Trails About the Uncontrolled Trails The protocol Development The

More information

Lecture 26: Missing data

Lecture 26: Missing data Lecture 26: Missing data Reading: ESL 9.6 STATS 202: Data mining and analysis December 1, 2017 1 / 10 Missing data is everywhere Survey data: nonresponse. 2 / 10 Missing data is everywhere Survey data:

More information

Traffic Analysis on Business-to-Business Websites. Masterarbeit

Traffic Analysis on Business-to-Business Websites. Masterarbeit Traffic Analysis on Business-to-Business Websites Masterarbeit zur Erlangung des akademischen Grades Master of Science (M. Sc.) im Studiengang Wirtschaftswissenschaft der Wirtschaftswissenschaftlichen

More information

Zambia Medicines Regulatory Authority APPLICATION FOR MARKETING AUTHORISATION OF A MEDICINE FOR HUMAN USE

Zambia Medicines Regulatory Authority APPLICATION FOR MARKETING AUTHORISATION OF A MEDICINE FOR HUMAN USE Zambia Medicines Regulatory Authority APPLICATION FOR MARKETING AUTHORISATION OF A MEDICINE FOR HUMAN USE COMMON TECHNICAL DOCUMENT FORMAT ZAMBIA Module 1 CTD-Modules 2-5 Version 03, April 2015 ZAMBIA

More information

Pooling Clinical Data: Key points and Pitfalls. October 16, 2012 Phuse 2012 conference, Budapest Florence Buchheit

Pooling Clinical Data: Key points and Pitfalls. October 16, 2012 Phuse 2012 conference, Budapest Florence Buchheit Pooling Clinical Data: Key points and Pitfalls October 16, 2012 Phuse 2012 conference, Budapest Florence Buchheit Introduction Are there any pre-defined rules to pool clinical data? Are there any pre-defined

More information

Qualitative Data Analysis: Coding. Qualitative Methods: Interviews Chickasaw Nation September 12, 2017

Qualitative Data Analysis: Coding. Qualitative Methods: Interviews Chickasaw Nation September 12, 2017 Qualitative Data Analysis: Coding Qualitative Methods: Interviews Chickasaw Nation September 12, 2017 Session Overview Memos to Codes Developing Codebook Coding Data Inter coder Agreement Thick Descriptions

More information

Creating an ADaM Data Set for Correlation Analyses

Creating an ADaM Data Set for Correlation Analyses PharmaSUG 2018 - Paper DS-17 ABSTRACT Creating an ADaM Data Set for Correlation Analyses Chad Melson, Experis Clinical, Cincinnati, OH The purpose of a correlation analysis is to evaluate relationships

More information

Automated Road Segment Creation Process

Automated Road Segment Creation Process David A. Noyce, PhD, PE Director and Chair Traffic Operations and Safety Laboratory Civil and Environmental Engineering Department Kelvin R. Santiago, MS, PE Assistant Researcher Traffic Operations and

More information

Computer Aided Draughting and Design: Graded Unit 1

Computer Aided Draughting and Design: Graded Unit 1 Higher National Graded Unit Specification General Information for Centres This Graded Unit has been validated as part of the HNC Computer Aided Draughting and Design (CADD) award. Centres are required

More information

Training/BestDose

Training/BestDose http://www.lapk.org/ Training/BestDose How can we optimize individual patient dosing? The parametric approach Population CL Data Posterior CL MAP-Bayesian Parametric model Maximum a posteriori Bayesian

More information

SAAM II Version 2.1 Basic Tutorials. Working with Parameters Basic

SAAM II Version 2.1 Basic Tutorials. Working with Parameters Basic SAAM II Version 2.1 Basic Tutorials Basic Introduction Parameters 2 Part 1a. Work with the Parameters dialog box Parameters 3 Part 1b. Hand-fit a model to data Parameters 12 Changing Parameters Manually

More information

Data Analysis and Solver Plugins for KSpread USER S MANUAL. Tomasz Maliszewski

Data Analysis and Solver Plugins for KSpread USER S MANUAL. Tomasz Maliszewski Data Analysis and Solver Plugins for KSpread USER S MANUAL Tomasz Maliszewski tmaliszewski@wp.pl Table of Content CHAPTER 1: INTRODUCTION... 3 1.1. ABOUT DATA ANALYSIS PLUGIN... 3 1.3. ABOUT SOLVER PLUGIN...

More information

Pmetrics Data and Model Files

Pmetrics Data and Model Files Pmetrics Data and Model Files General Workflow You!supply!these!files;!Pmetrics!does!the!rest! Data!.csv!file Model!.txt!file ITrun(),!NPrun() PreparaGon!program InstrucGon!file PMload() Engine!program

More information

A Finite Element Analysis Workflow in Biomechanics. Define the problem and identify FEA as the favorable modeling & simulation strategy

A Finite Element Analysis Workflow in Biomechanics. Define the problem and identify FEA as the favorable modeling & simulation strategy Before you start Understand the Premise of FEA Finite element analysis is a computational modeling & simulation strategy; specifically, a numerical procedure for solving field problems, e.g., stress, by

More information

8 th Grade Mathematics Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the

8 th Grade Mathematics Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 8 th Grade Mathematics Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 2012-13. This document is designed to help North Carolina educators

More information

parameters, network shape interpretations,

parameters, network shape interpretations, GIScience 20100 Short Paper Proceedings, Zurich, Switzerland, September. Formalizing Guidelines for Building Meaningful Self- Organizing Maps Jochen Wendel 1, Barbara. P. Buttenfield 1 1 Department of

More information

Genetic Analysis. Page 1

Genetic Analysis. Page 1 Genetic Analysis Page 1 Genetic Analysis Objectives: 1) Set up Case-Control Association analysis and the Basic Genetics Workflow 2) Use JMP tools to interact with and explore results 3) Learn advanced

More information

Final Project Report

Final Project Report 16.04.02 Final Project Report Document information Project Title HP Tool Repository of SESAR standard HP methods and tools Project Number 16.04.02 Project Manager DFS Deliverable Name 16.04.02 Final Project

More information

Tools to Facilitate the Creation of Pooled Clinical Trials Databases

Tools to Facilitate the Creation of Pooled Clinical Trials Databases Paper AD10 Tools to Facilitate the Creation of Pooled Clinical Trials Databases Patricia Majcher, Johnson & Johnson Pharmaceutical Research & Development, L.L.C., Raritan, NJ ABSTRACT Data collected from

More information

From ODM to SDTM: An End-to-End Approach Applied to Phase I Clinical Trials

From ODM to SDTM: An End-to-End Approach Applied to Phase I Clinical Trials PhUSE 2014 Paper PP05 From ODM to SDTM: An End-to-End Approach Applied to Phase I Clinical Trials Alexandre Mathis, Department of Clinical Pharmacology, Actelion Pharmaceuticals Ltd., Allschwil, Switzerland

More information

SECURITY AUTOMATION BEST PRACTICES. A Guide on Making Your Security Team Successful with Automation SECURITY AUTOMATION BEST PRACTICES - 1

SECURITY AUTOMATION BEST PRACTICES. A Guide on Making Your Security Team Successful with Automation SECURITY AUTOMATION BEST PRACTICES - 1 SECURITY AUTOMATION BEST PRACTICES A Guide on Making Your Security Team Successful with Automation SECURITY AUTOMATION BEST PRACTICES - 1 Introduction The best security postures are those that are built

More information

PharmaSUG Paper PO03

PharmaSUG Paper PO03 PharmaSUG 2012 - Paper PO03 Streamlining Regulatory Submission with CDISC/ADaM Standards for Nonstandard Pharmacokinetic/Pharmacodynamic Analysis Datasets Xiaopeng Li, Celerion Inc., Lincoln, NE (USA)

More information

Security Automation Best Practices

Security Automation Best Practices WHITEPAPER Security Automation Best Practices A guide to making your security team successful with automation TABLE OF CONTENTS Introduction 3 What Is Security Automation? 3 Security Automation: A Tough

More information

JMP Clinical. Release Notes. Version 5.0

JMP Clinical. Release Notes. Version 5.0 JMP Clinical Version 5.0 Release Notes Creativity involves breaking out of established patterns in order to look at things in a different way. Edward de Bono JMP, A Business Unit of SAS SAS Campus Drive

More information

DATA MINING AND WAREHOUSING

DATA MINING AND WAREHOUSING DATA MINING AND WAREHOUSING Qno Question Answer 1 Define data warehouse? Data warehouse is a subject oriented, integrated, time-variant, and nonvolatile collection of data that supports management's decision-making

More information

IBM SPSS Categories 23

IBM SPSS Categories 23 IBM SPSS Categories 23 Note Before using this information and the product it supports, read the information in Notices on page 55. Product Information This edition applies to version 23, release 0, modification

More information

BORN Ontario s Data Quality Framework

BORN Ontario s Data Quality Framework BORN Ontario s Data Quality Framework At BORN Ontario we make Data Privacy and Data Quality our highest priority. We recognize that the quality of the data directly impacts use of the data. With addition

More information

SECURITY AUTOMATION BEST PRACTICES. A Guide to Making Your Security Team Successful with Automation

SECURITY AUTOMATION BEST PRACTICES. A Guide to Making Your Security Team Successful with Automation SECURITY AUTOMATION BEST PRACTICES A Guide to Making Your Security Team Successful with Automation TABLE OF CONTENTS Introduction 3 What Is Security Automation? 3 Security Automation: A Tough Nut to Crack

More information

PopED Manual. Release version 2.0

PopED Manual. Release version 2.0 PopED Manual Release version 2.0 Introduction PopED (Population Experimental Design) is a software tool for computing optimal experimental designs. The software has been developed with an emphasis on drug

More information

Certification Authorities Software Team (CAST) Position Paper CAST-25

Certification Authorities Software Team (CAST) Position Paper CAST-25 Certification Authorities Software Team (CAST) Position Paper CAST-25 CONSIDERATIONS WHEN USING A QUALIFIABLE DEVELOPMENT ENVIRONMENT (QDE) IN CERTIFICATION PROJECTS COMPLETED SEPTEMBER 2005 (Rev 0) NOTE:

More information

CPSC 444 Project Milestone III: Prototyping & Experiment Design Feb 6, 2018

CPSC 444 Project Milestone III: Prototyping & Experiment Design Feb 6, 2018 CPSC 444 Project Milestone III: Prototyping & Experiment Design Feb 6, 2018 OVERVIEW... 2 SUMMARY OF MILESTONE III DELIVERABLES... 2 1. Blog Update #3 - Low-fidelity Prototyping & Cognitive Walkthrough,

More information

Higher National Unit specification: general information. Graded Unit title: Computing: Networking: Graded Unit 2

Higher National Unit specification: general information. Graded Unit title: Computing: Networking: Graded Unit 2 Higher National Unit specification: general information This Graded Unit has been validated as part of the HND Computing: Networking. Centres are required to develop the assessment instrument in accordance

More information

AMGET, an R-Based Postprocessing Tool for ADAPT 5

AMGET, an R-Based Postprocessing Tool for ADAPT 5 Original Article Citation: CPT: Pharmacometrics & Systems Pharmacology (203) 2, e6; doi:0.038/psp.203.36 203 ASCPT All rights reserved 263-8306/2 www.nature.com/psp AMGET, an R-Based Postprocessing Tool

More information

UX Research in the Product Lifecycle

UX Research in the Product Lifecycle UX Research in the Product Lifecycle I incorporate how users work into the product early, frequently and iteratively throughout the development lifecycle. This means selecting from a suite of methods and

More information

Package scarabee. August 22, 2014

Package scarabee. August 22, 2014 Package scarabee August 22, 2014 Type Package Title Optimization Toolkit for Pharmacokinetic-Pharmacodynamic Models Version 1.1-3 Date 2014-02-09 Author Sebastien Bihorel Maintainer Sebastien Bihorel

More information

Preparing for Data Analysis

Preparing for Data Analysis Preparing for Data Analysis Prof. Andrew Stokes March 27, 2018 Managing your data Entering the data into a database Reading the data into a statistical computing package Checking the data for errors and

More information

Optimization. Definition: Terms: Why Optimization is Necessary? 29/10/2016. Factor Levels Response Effects Interaction

Optimization. Definition: Terms: Why Optimization is Necessary? 29/10/2016. Factor Levels Response Effects Interaction Definition: Optimization The term Optimize is to make perfect. It is defined as follows: choosing the best element from some set of available alternatives. An art, process, or methodology of making something

More information

Document Name: Version Release Description for Kinetica 5.0 (r1) Document Number: 0847 Printed in the United States of America.

Document Name: Version Release Description for Kinetica 5.0 (r1) Document Number: 0847 Printed in the United States of America. Information in this document is subject to change without notice, is the property of Thermo Fisher Scientific, and is considered confidential. Reproduction, copying, and/or disclosure to others are strictly

More information

A Framework for Source Code metrics

A Framework for Source Code metrics A Framework for Source Code metrics Neli Maneva, Nikolay Grozev, Delyan Lilov Abstract: The paper presents our approach to the systematic and tool-supported source code measurement for quality analysis.

More information

Graded Unit title: Computing: Networking: Graded Unit 2

Graded Unit title: Computing: Networking: Graded Unit 2 SQA Advanced Unit specification: general information for centres This graded unit has been validated as part of the SQA Advanced Diploma in Computing: Networking. Centres are required to develop the assessment

More information

Step-by-Step Guide to Basic Genetic Analysis

Step-by-Step Guide to Basic Genetic Analysis Step-by-Step Guide to Basic Genetic Analysis Page 1 Introduction This document shows you how to clean up your genetic data, assess its statistical properties and perform simple analyses such as case-control

More information

Driven Cavity Example

Driven Cavity Example BMAppendixI.qxd 11/14/12 6:55 PM Page I-1 I CFD Driven Cavity Example I.1 Problem One of the classic benchmarks in CFD is the driven cavity problem. Consider steady, incompressible, viscous flow in a square

More information

Sandeep Kharidhi and WenSui Liu ChoicePoint Precision Marketing

Sandeep Kharidhi and WenSui Liu ChoicePoint Precision Marketing Generalized Additive Model and Applications in Direct Marketing Sandeep Kharidhi and WenSui Liu ChoicePoint Precision Marketing Abstract Logistic regression 1 has been widely used in direct marketing applications

More information

STATISTICS (STAT) Statistics (STAT) 1

STATISTICS (STAT) Statistics (STAT) 1 Statistics (STAT) 1 STATISTICS (STAT) STAT 2013 Elementary Statistics (A) Prerequisites: MATH 1483 or MATH 1513, each with a grade of "C" or better; or an acceptable placement score (see placement.okstate.edu).

More information

Experiences in Data Quality

Experiences in Data Quality Experiences in Data Quality MIT IQIS 2010 Annette Pence July 14-16, 2010 Approved for Public Release: 10-0686 Distribution Unlimited As a public interest company, MITRE works in partnership with the government

More information

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview Chapter 888 Introduction This procedure generates D-optimal designs for multi-factor experiments with both quantitative and qualitative factors. The factors can have a mixed number of levels. For example,

More information

Differentiation of Cognitive Abilities across the Lifespan. Online Supplement. Elliot M. Tucker-Drob

Differentiation of Cognitive Abilities across the Lifespan. Online Supplement. Elliot M. Tucker-Drob 1 Differentiation of Cognitive Abilities across the Lifespan Online Supplement Elliot M. Tucker-Drob This online supplement reports the results of an alternative set of analyses performed on a single sample

More information

Study Data Reviewer s Guide

Study Data Reviewer s Guide Revision History Date Study Data Reviewer s Guide Completion Guideline: Nonclinical (nnsdrg) Version Summary V1.1 03 March 2016 1.0 First Public Version: posted for Public Comment 1.1 Update from Public

More information

Pharmaceuticals, Health Care, and Life Sciences

Pharmaceuticals, Health Care, and Life Sciences Successful Lab Result Conversion for LAB Analysis Data with Minimum Effort Pushpa Saranadasa, Merck & Co., Inc. INTRODUCTION In the pharmaceutical industry, the statistical results of a clinical trial's

More information

In this chapter we explain the structure and use of BestDose, with examples.

In this chapter we explain the structure and use of BestDose, with examples. Using BestDose In this chapter we explain the structure and use of BestDose, with examples. General description The image below shows the view you get when you start BestDose. The actual view is called

More information

STEPHEN WOLFRAM MATHEMATICADO. Fourth Edition WOLFRAM MEDIA CAMBRIDGE UNIVERSITY PRESS

STEPHEN WOLFRAM MATHEMATICADO. Fourth Edition WOLFRAM MEDIA CAMBRIDGE UNIVERSITY PRESS STEPHEN WOLFRAM MATHEMATICADO OO Fourth Edition WOLFRAM MEDIA CAMBRIDGE UNIVERSITY PRESS Table of Contents XXI a section new for Version 3 a section new for Version 4 a section substantially modified for

More information

How to clean up dirty data in Patient reported outcomes

How to clean up dirty data in Patient reported outcomes Paper DH02 How to clean up dirty data in Patient reported outcomes Knut Mueller, UCB Schwarz Biosciences, Monheim, Germany ABSTRACT The current FDA Guidance for Industry - Patient Reported Outcome Measures

More information

Harmonization of usability measurements in ISO9126 software engineering standards

Harmonization of usability measurements in ISO9126 software engineering standards Harmonization of usability measurements in ISO9126 software engineering standards Laila Cheikhi, Alain Abran and Witold Suryn École de Technologie Supérieure, 1100 Notre-Dame Ouest, Montréal, Canada laila.cheikhi.1@ens.etsmtl.ca,

More information

Viewpoint Review & Analytics

Viewpoint Review & Analytics The Viewpoint all-in-one e-discovery platform enables law firms, corporations and service providers to manage every phase of the e-discovery lifecycle with the power of a single product. The Viewpoint

More information

Standard Safety Visualization Set-up Using Spotfire

Standard Safety Visualization Set-up Using Spotfire Paper SD08 Standard Safety Visualization Set-up Using Spotfire Michaela Mertes, F. Hoffmann-La Roche, Ltd., Basel, Switzerland ABSTRACT Stakeholders are requesting real-time access to clinical data to

More information

Guidance for the format and content of the final study report of non-interventional post-authorisation safety studies

Guidance for the format and content of the final study report of non-interventional post-authorisation safety studies 23 January 2013 EMA/48663/2013 Patient Health Protection Guidance for the format and content of the final study report of non-interventional post-authorisation safety studies Introduction From 10 January

More information

Exercise 10: Built-in Models in DME: Population PK/PD Model with PK Fixed

Exercise 10: Built-in Models in DME: Population PK/PD Model with PK Fixed Exercise 10: Built-in Models in DME: Population PK/PD Model with PK Fixed Background The dataset for this exercise contains simulated observations for 200 subjects. Objectives: Obtain initial estimates

More information

Correctly Compute Complex Samples Statistics

Correctly Compute Complex Samples Statistics PASW Complex Samples 17.0 Specifications Correctly Compute Complex Samples Statistics When you conduct sample surveys, use a statistics package dedicated to producing correct estimates for complex sample

More information