Screening Design Selection

Similar documents
Multivariate Capability Analysis

Moving Average (MA) Charts

Design of Experiments

If the active datasheet is empty when the StatWizard appears, a dialog box is displayed to assist in entering data.

Matrix Plot. Sample StatFolio: matrixplot.sgp

Biostatistics and Design of Experiments Prof. Mukesh Doble Department of Biotechnology Indian Institute of Technology, Madras

Multivariate Normal Random Numbers

CHAPTER 2 DESIGN DEFINITION

STATS PAD USER MANUAL

Two-Level Designs. Chapter 881. Introduction. Experimental Design. Experimental Design Definitions. Alias. Blocking

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview

Cpk: What is its Capability? By: Rick Haynes, Master Black Belt Smarter Solutions, Inc.

Lab #9: ANOVA and TUKEY tests

CHAPTER 3 AN OVERVIEW OF DESIGN OF EXPERIMENTS AND RESPONSE SURFACE METHODOLOGY

Minitab detailed

Research Methods for Business and Management. Session 8a- Analyzing Quantitative Data- using SPSS 16 Andre Samuel

General Program Description

FURTHER ORTHOGONAL ARRAYS

Modern Experimental Design

Descriptive Statistics, Standard Deviation and Standard Error

Box-Cox Transformation

Package blocksdesign

Data Analysis and Solver Plugins for KSpread USER S MANUAL. Tomasz Maliszewski

CHAPTER 4. OPTIMIZATION OF PROCESS PARAMETER OF TURNING Al-SiC p (10P) MMC USING TAGUCHI METHOD (SINGLE OBJECTIVE)

Dealing with Categorical Data Types in a Designed Experiment

Z-TEST / Z-STATISTIC: used to test hypotheses about. µ when the population standard deviation is unknown

Analysis of Two-Level Designs

Statistics I 2011/2012 Notes about the third Computer Class: Simulation of samples and goodness of fit; Central Limit Theorem; Confidence intervals.

And the benefits are immediate minimal changes to the interface allow you and your teams to access these

Binary Diagnostic Tests Clustered Samples

Harris: Quantitative Chemical Analysis, Eight Edition CHAPTER 05: QUALITY ASSURANCE AND CALIBRATION METHODS

Assignment 4/5 Statistics Due: Nov. 29

Box-Cox Transformation for Simple Linear Regression

STATGRAPHICS PLUS for WINDOWS

Categorical Data in a Designed Experiment Part 2: Sizing with a Binary Response

General Multilevel-Categoric Factorial Tutorial

Chemical Reaction dataset ( )

IQR = number. summary: largest. = 2. Upper half: Q3 =

Design of Experiments in a Transactional Environment

We have seen that as n increases, the length of our confidence interval decreases, the confidence interval will be more narrow.

Recall the expression for the minimum significant difference (w) used in the Tukey fixed-range method for means separation:

Getting Started with Minitab 17

Table Of Contents. Table Of Contents

Frequency Tables. Chapter 500. Introduction. Frequency Tables. Types of Categorical Variables. Data Structure. Missing Values

BIOMETRICS INFORMATION

Lab 5 - Risk Analysis, Robustness, and Power

MAT 110 WORKSHOP. Updated Fall 2018

Math 227 EXCEL / MEGASTAT Guide

DESIGN OF EXPERIMENTS IN THE MÖBIUS MODELING FRAMEWORK PATRICK GERALD WEBSTER

SPSS INSTRUCTION CHAPTER 9

Section 2.3: Simple Linear Regression: Predictions and Inference

Linear Methods for Regression and Shrinkage Methods

1. Determine the population mean of x denoted m x. Ans. 10 from bottom bell curve.

Algebra II Quadratic Functions

A HYBRID METHOD FOR SIMULATION FACTOR SCREENING. Hua Shen Hong Wan

Chapter 12: Statistics

Fly wing length data Sokal and Rohlf Box 10.1 Ch13.xls. on chalk board

QstatLab: software for statistical process control and robust engineering

Let s take a closer look at the standard deviation.

Psychology 282 Lecture #21 Outline Categorical IVs in MLR: Effects Coding and Contrast Coding

Exercise: Graphing and Least Squares Fitting in Quattro Pro

I n many cases, the SPRT will come to a decision with fewer samples than would have been required for a fixed size test.

Section 3.2: Multiple Linear Regression II. Jared S. Murray The University of Texas at Austin McCombs School of Business

Equivalence Tests for Two Means in a 2x2 Cross-Over Design using Differences

LP360 (Advanced) Planar Point Filter

Getting Started with Minitab 18

Minitab 17 commands Prepared by Jeffrey S. Simonoff

Metrics for Performance Evaluation How to evaluate the performance of a model? Methods for Performance Evaluation How to obtain reliable estimates?

Excel 2013 Intermediate

Pair-Wise Multiple Comparisons (Simulation)

End Behavior and Symmetry

Confidence Intervals: Estimators

THE L.L. THURSTONE PSYCHOMETRIC LABORATORY UNIVERSITY OF NORTH CAROLINA. Forrest W. Young & Carla M. Bann

BESTFIT, DISTRIBUTION FITTING SOFTWARE BY PALISADE CORPORATION

Package blocksdesign

DOWNLOAD PDF BIG IDEAS MATH VERTICAL SHRINK OF A PARABOLA

For Additional Information...

Example 5.25: (page 228) Screenshots from JMP. These examples assume post-hoc analysis using a Protected LSD or Protected Welch strategy.

Robust Linear Regression (Passing- Bablok Median-Slope)

E-Campus Inferential Statistics - Part 2

How Random is Random?

STAT 113: Lab 9. Colin Reimer Dawson. Last revised November 10, 2015

STAT 2607 REVIEW PROBLEMS Word problems must be answered in words of the problem.

COMPARING MODELS AND CURVES. Fitting models to biological data using linear and nonlinear regression. A practical guide to curve fitting.

Fifth Grade Math Rubric

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Section 4 General Factorial Tutorials

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

What's New in Rev. 3.0.

Software Tutorial Session Universal Kriging

Minitab 18 Feature List

DATA DEFINITION PHASE

Unit 5: Estimating with Confidence

THIS IS NOT REPRESNTATIVE OF CURRENT CLASS MATERIAL. STOR 455 Midterm 1 September 28, 2010

Things you ll know (or know better to watch out for!) when you leave in December: 1. What you can and cannot infer from graphs.

Performance Evaluation

Stat 528 (Autumn 2008) Density Curves and the Normal Distribution. Measures of center and spread. Features of the normal distribution

Measures of Dispersion

Machine Learning Techniques for Data Mining

ASSIGNMENT 6 Final_Tracts.shp Phil_Housing.mat lnmv %Vac %NW Final_Tracts.shp Philadelphia Housing Phil_Housing_ Using Matlab Eire

Transcription:

Screening Design Selection Summary... 1 Data Input... 2 Analysis Summary... 5 Power Curve... 7 Calculations... 7 Summary The STATGRAPHICS experimental design section can create a wide variety of designs intended to screen out the most important factors affecting a process. Included are two-level factorial designs, fractional factorials, mixed level fractions, irregular fractions, and Plackett-Burman designs. Each design may be replicated or augmented by the addition of centerpoints, depending on the magnitude of the effects that the experimenter wishes to detect. This procedure is designed to help select an appropriate design from that collection, including the number of replicates and centerpoints. Once the desired effect and maximum block size are specified, a list of the best designs of each type is generated. The power curves for each suggested design are also plotted. The desired design may then be created by selecting the Create Design procedure on the experimental design menu. Sample StatFolio: screendesign.sgp Sample Data: None. 2013 by StatPoint Technologies, Inc. Screening Design Selection p. 1

Data Input The first data input dialog box for this procedure is shown below. STATGRAPHICS Rev. 9/16/2013 Number of Factors: the number of factors k to be varied during the experiment, between 2 and 16. The primary goal of most screening experiments is to select the most important factors from a larger list of possible candidates. Designs to Consider: the type of screening designs to consider in the selection process. Designs on the list include: Full 2-level factorials designs consisting of all combination of 2 levels for each of the k factors in the experiment. Full factorials can estimate the main effects of each factor and all interactions amongst the factors. Fractional factorials designs containing a subset of the runs in the full 2-level factorial. These designs are classified by their resolution, which refers to the type of effects that can be estimated: Resolution V: the design can estimate all main effects and 2-factor interactions. These designs are preferable if they can be done in an acceptable number of runs. Resolution IV: the design can estimate all main effects, but the 2-factor interactions are confounded amongst themselves. These designs can sense the 2013 by StatPoint Technologies, Inc. Screening Design Selection p. 2

presence of interactions amongst the factors but cannot resolve which interactions are present. Resolution III: main effects are clear of each other but are confounded with the 2-factor interactions. These designs should be used with care since any interactions can distort the interpretation of the main effects of the factors. Irregular fractions designs involving a non-orthogonal subset of the runs in the full 2- level factorials. Although there is some confounding between the various effects, the designs included have confounding patterns that make them nearly the resolution indicated. Mixed level factions designs involving 3 levels of one factor and 2 levels of the other factors. Such designs can estimate curvature with respect to the factor that is run at 3 levels. Plackett-Burman designs subsets of the full 2-level factorial where, unlike the fractional factorials, the number of runs is not a power of 2. These designs are useful for screening large number of factors, although their confounding pattern is not as clean as the fractional factorials. Maximum runs per block: if the design is to be run in blocks, the maximum number of runs in a single block. Blocking is performed to eliminate the effect of an external factor that could impact the results, such as batch of raw material when all runs cannot be done with the same batch. This field may be left blank if the design will not be blocked. Minimum centerpoints per block: the smallest number of centerpoints per block. Centerpoints are runs done at a middle level of each of the factors, often replicated in order to estimate the magnitude of the experimental error. This setting insures that a minimum number of such runs will be generated. Experimental error sigma: the standard deviation of the experimental error. The experimental error is a measure of the extent to which the results of a single run might vary if the run was repeated at the same conditions of the experimental factors. The larger the experimental error, the more runs required to estimate an effect of a given magnitude. Normally, you would enter an estimate based on the results of previous experiments. If no estimate is available, you may leave this setting at 1.0. The effects displayed will then be expressed as a multiple of. The dialog box shown above requests the generation of experimental designs for studying k = 5 factors. Only designs that are of at least Resolution V (or nearly so) are to be considered. No blocking is to be done, nor is there any restriction on the minimum number of centerpoints. The experimental error is set to 1, so that the effects will be expressed in units of. The second data input dialog box elicits information about the goal of the experiment: 2013 by StatPoint Technologies, Inc. Screening Design Selection p. 3

Control: specifies the goal from among the following choices: 1. Absolute error: a confidence interval for each effect is to be constructed. That interval should not deviate from the point estimate in either direction by more than the absolute distance W indicated. The effect of a factor is defined as the magnitude of the change in the response over the range in which that factor is changed: effect = (response at high level of a factor) (response at low level of a factor) If = 1, then W = 1 implies a desired confidence interval for each effect of one sigma. 2. Relative error: a confidence interval for each effect is to be constructed. That interval should not deviate from the point estimate in either direction by more than the relative percentage P indicated. This is identical to the absolute error case with W set equal to P times the specified effect. 3. Power: a hypothesis test is to be performed to determine whether or not each effect is significantly different than zero. The power of the test 100(1-) % should equal the percentage specified when the true value of the effect equals = Effect to Detect. Power is defined as the probability of rejecting the null hypothesis when it is false, i.e., the probability that a factor will be judged to be statistically significant if the true effect equals. 4. Total Number of Runs: the maximum number of runs that can be performed. If this selection is chosen, then the procedure will determine the best design of each type in which the number of runs does not exceed this limit. Confidence Level: the level of confidence 100(1-)% used when constructing confidence intervals. The value is also used as the level of Type I error when testing hypotheses. A 2013 by StatPoint Technologies, Inc. Screening Design Selection p. 4

Type I error occurs when the null hypothesis is falsely rejected. In this case, it refers to the act of identifying an effect that does not actually exist. For example, the dialog box above indicates that the following tests will be performed after the experiment is run: Null hypothesis H 0 : effect = 0 Alt. hypothesis H A : effect 0 The probability of a Type I error (rejecting a true null hypothesis) is set to = 5%, while the probability of a Type II error (not rejecting a false null hypothesis) is set to = 10% when the true absolute effect equals 2 (sigma). In practice, this will be done by looking at the ANOVA table and identifying as statistically significant any effect that has a P-Value less than 0.05. The following table may be helpful in remembering how to set the error probabilities. H 0 is true (effect = 0) H 0 is false (effect 0) Conclude that effect is not significant Correct decision Type II error (false negative) risk = Conclude that effect is significant Type I error (false positive) risk = Correct decision power = 1- Analysis Summary The Analysis Summary shows the input parameters and the smallest available designs of each type: Screening Design Selection Input Number of Min. Centerpoints Max. Runs Exp. Error Desired Target Confidence Factors per Block per Block Sigma Power Effect Level 5 0 1.0 90.0% 2.0 95.0% Selected Designs Corner Center Error Design Runs Resol. Points Points D.F. Reps. Blocks Power (%) Factorial 2^5 32 V+ 32 0 16 1 1 99.863 Half fraction 2^5-1 22 V 16 6 6 1 1 91.4633 Mixed level fraction 3*2^4-1 24 ~V 16 0 7 1 1 92.7197 The output includes: Design: the type of design Runs: the total number of runs to be performed. Resol.: the resolution of the design. V+ indicates a design of resolution greater than V. ~V means a design of nearly resolution V. Corner Points: the number of runs N placed at the vertices of the experimental region, i.e., runs in which all factors are set either at their highest level or their lowest level. 2013 by StatPoint Technologies, Inc. Screening Design Selection p. 5

STATGRAPHICS Rev. 9/16/2013 Center Points: the number of runs C placed at the center of the experimental region, i.e., runs in which all factors are set midway between their low and high level. Error D.F.: the number of degrees of freedom available for estimating the experimental error. The higher the degrees of freedom, the better the estimate. In calculating this value, it is assumed that all estimable main effects and two-factor interactions will be estimated, including combinations of 2-factor interactions in the case of the fractional factorials. For the mixed level factorials, a single quadratic effect is also accounted for. Reps.: the number of replicates of the runs in the base design. Blocks: the number of blocks into which the runs have been divided. Power: the probability of identifying as statistically significant any effect of size equal to the Effect to detect specified on the data input dialog box. In the current example, three designs have been identified: 1. Factorial 2^5: this is a full factorial design containing all N = 32 combinations of the high and low level of the factors. Since the power is well in excess of 90%, no centerpoints were added. 2. Factorial 2^5-1: a fractional factorial design containing N = 16 of the 32 combinations of the runs in the full factorial. 6 centerpoints were added in order to achieve the desired power. 3. Mixed level fraction 3*2^4-1: a fractional factorial design involving 3 levels of one factor and 2 levels of the other four factors. There are N = 16 corner points, plus 8 additional points at a middle level of the factor that is run at 3 levels. This design is not perfectly orthogonal, due to the presence of the third level of the one factor, but it is close enough to be easily interpreted. The latter 2 designs are considerably smaller than the full factorial. If the desired power is 90%, one of these two designs should be selected. The choice between the two depends on whether or not the experimenter wishes to perform 2 extra runs in order to estimate possible curvature with respect to one selected factor. 2013 by StatPoint Technologies, Inc. Screening Design Selection p. 6

Power STATGRAPHICS Rev. 9/16/2013 Power Curve The Power Curve shows the probability that each of the selected designs will identify as statistically significant effects of a given magnitude. Power Curve Factorial 2^5 Half f raction 2^5-1 Mixed lev el f raction 3*2^4-1 100 80 60 40 20 0-2 -1 0 1 2 True Effect The half fraction and mixed level fraction have very similar power. Calculations Selection Algorithm The program begins with one replicate of the base design. It then adds points one at a time until the desired precision or power is achieved. Initially, centerpoints are added until enough additional points have been added such that a full replicate of the base design can be created instead. For blocked designs, the only designs considered are those with an equal number of centerpoints in all blocks. Standard Error of an Effect 2 4 effect (1) N Confidence Interval for an Effect estimate t ˆ (2) / 2, effect 2013 by StatPoint Technologies, Inc. Screening Design Selection p. 7