Presentation is loading. Please wait.

Presentation is loading. Please wait.

1 Chapter 5.8 What if We Have More Than Two Samples?

Similar presentations


Presentation on theme: "1 Chapter 5.8 What if We Have More Than Two Samples?"— Presentation transcript:

1 1 Chapter 5.8 What if We Have More Than Two Samples?

2 2 Example A manufacturer of paper used for making grocery bags is interested in improving the tensile strength of the product Product engineering thinks that tensile strength is a function of the hardwood concentration in the pulp and that the range of hardwood concentrations of practical interest is between 5 and 20% A team of engineers responsible for the study decides to investigate four levels of hardwood concentration: 5, 10, 15 and 20% They decide to make up six test specimens at each concentration level, using a pilot plant All 24 specimens are tested on a laboratory tensile tester, in random order The data from this experiment are as follows

3 3 Factor : Hardwood concentration Factor level or treatments: 4 (4 levels of the factor: 5, 10, 15, 20) 6 replicates for each level Randomization: all 4x6=24 runs are performed in a random order. (Completely Randomize Design) Single factor layout having 4 levels Terminology Single-Factor ANOVA Or One-way ANOVA (Analysis of Variance)

4 4

5 5 Notation a: different levels of a single factor y ij : jth observation taken under factor level i n i : number of replicates for factor level i If n 1 = n 2 =  = n a = n then “balanced design” otherwise, “unbalanced design”

6 6 run in random order…a completely randomized design (CRD) N = an total runs (For a balanced design) Objective Testing hypotheses about the equality of the a treatment means

7 7  Assumptions 1) Normality 2) Constant Variance 3) Independence Model  ij iid N(0,  2 ) Y ij : response variable

8 8 Hypothesis Formal statistical hypotheses are Alternatively,

9 9 Analysis of Variance The name “analysis of variance” stems from a partitioning of the total variability in the response variable (Y ij ) into components that are consistent with a model for the experiment. Total variability is measured by the total sum of squares:

10 10 A large value of SS Treatments reflects large differences in treatment means A small value of SS Treatments likely indicates no differences in treatment means While sums of squares cannot be directly compared to test the hypothesis of equal means, mean squares can be compared. A mean square is a sum of squares divided by its degrees of freedom:

11 11 We can show that Comparing MS treatments and MS E If MS treatments  MS E then we do not reject H 0 If MS treatments >> MS E then we reject H 0 Testing

12 12 ANOVA Table

13 13 Estimation of Model Parameters Point estimation Confidence interval on  i (the ith treatment mean)

14 14 Which Means Differ? : Post-ANOVA Comparison If the hypothesis is rejected, we don’t know which specific means are different Determining which specific means differ following an ANOVA is called the multiple comparisons problem: H 0 :  i =  j H 1 :  i ≠  j (i  j,) There are many quantitative methods for multiple comparisons (Fisher’s LSD method, Tukey’s test, Dunett’s method etc.) Fisher’s LSD method Least significant difference We conclude that the population means  i and  j differ if > LSD. Drawback: Overall type I error rate is inflated using this method. If  0 is used for H 0 :  1 =  2 H 0 :  1 =  3 H 0 :  2 =  3 Experimentwise error rate  =1-(1-  0 ) 3

15 Tukey’s Method  We conclude that the population means  i and  j differ if where q  ( a, f ) is given in some textbooks or we can do this test using minitab.  The overall experimentwise error rate is . 15

16 16 Model Checking Checking assumptions is important Normality Constant variance Independence Residual plots are very useful Residual Difference between an observation and its fitted (estimated) value from the model Checked by examining residuals

17 17 1. Normal Probability Plot of the Residuals Normality Assumption Check If the plotted points fall approximately along a straight line, conclude that the data follow the normal distribution.

18 18 2. Residuals vs. Time Plotting of residuals in time sequence is helpful in detecting correlation between the residuals. A tendency to have runs of positive and negative residuals indicates positive correlation. Changing error variance over time indicates nonconstant variance.

19 19 3. Residuals vs. Fitted Values Plotting of residuals versus fitted values should not reveal any obvious pattern. When a pattern appears in these plots, it suggests the need for a transformation Changing error variance over the magnitude of the fitted values may indicate the violation of constant variance.

20 20 Example In IC manufacturing steps, wafers are coated with a layer of material such as silicon dioxide or a metal. The unwanted material is then selectively removed by etching through a mask creating circuit patterns. A plasma etching process is widely used for this process. Figure 3-1. A single wafer plasma etching tool

21 21 The response variable is the etch rate for this tool. The experimenter wants to specify the RF power setting that will give a desired target etch rate. RF power can vary between 160 W and 220 W. The experimenter chooses 4 levels of RF power; 160W, 180W, 200W, 220W. The experiment is replicated 5 times. 20 runs are made in a random order.

22 Figure. Box plots and scatter diagram of the etch rate data. 22

23 23 1. Construct ANOVA table Minitab Result: One-way ANOVA: Etch Rate versus Power Source DF SS MS F P Power 3 66871 22290 66.80 0.000 Error 16 5339 334 Total 19 72210 2. Determine if the RF power setting affects the etch rate. Use a=0.05 Since p-value< 0.05, we reject the null hypothesis saying all treatment means are the same. That is, we conclude that the RF power setting affects the etch rate

24 24 3. Find the 95% confidence interval on the mean of 220W of RF power Individual 95% CIs For Mean Based on Pooled StDev Level N Mean StDev ---+---------+---------+---------+------ 160W 5 551.20 20.02 (--*---) 180W 5 587.40 16.74 (--*---) 200W 5 625.40 20.53 (--*---) 220W 5 707.00 15.25 (--*---) ---+---------+---------+---------+------ 550 600 650 700 Pooled StDev = 18.27

25 25 5. Check the model assumptions

26 The Randomized Complete Block Design (RCBD) Definition A nuisance factor is a design factor that probably has some effect on the response, but it’s of no interest to the experimenter. (…however, the variability it transmits to the response needs to be minimized.) ex) Batches of raw material, operators, pieces of test equipment, time (shifts, days, etc.), different experimental units. Dealing nuisance factors 1. Unknown, uncontrollable => randomization to guard against it 2. Known, uncontrollable => use Analysis of Covariance (see Chapter 14) 3. Known, Controllable => use “blocking” to eliminate its effect 26

27 EXAMPLE Consider a hardness testing machine that presses a rod with a pointed tip into a metal test coupon with a known force. By measuring the depth of the depression caused by the tip, the hardness of the coupon is determined. ■ We wish to determine whether or not four different tips produce different readings on a hardness testing machine. ■ Structure of a completely randomized single factor experiment: 4 ⅹ 4 = 16 runs, requiring 16 coupons.. ■ Nuisance factor: Test coupons (.If the metal coupons differ slightly in their hardness, they will contribute to the variability observed in the hardness data) Experimental error = Random error + Variability between coupons ■ Alternatively, the experimenter may want to test the tips across coupons of various hardness levels. => “Blocking” 27 We want to remove it, because we want to make experimental error as small as possible !!

28 Randomized Complete Block Design (RCBD)  To conduct this experiment as a RCBD, assign all 4 tips to each coupon.  Each coupon is called a “block”; that is, it’s a more homogenous experimental unit on which to test the tips.  Variability between blocks can be large, but variability within a block should be relatively small.  In general, a block is a specific level of the nuisance factor.  “Complete” means each block contains all the treatments (tip1,2,3,4)  A block represents a restriction on randomization.  All runs within a block are randomized.  RCBD improve the comparison accuracy among tips by eliminating the variability among coupons. 28

29 Notice the two-way structure of the experiment. Once again, we are interested in testing the equality of treatment means, but now we have to remove the variability associated with the nuisance factor (the blocks) 29

30 Extension of the ANOVA to the RCBD Suppose that there are a treatments (factor levels) and b blocks A statistical model (effects model) for the RCBD is  : overall mean  i : i th treatment effect  j : j th block effect  ij : random error ~ i.i.d. N(0,  2 ) 30

31 31 y 11 y 21 y 31. y a1 y 12 y 22 y 32. y a2 y 1b y 2b y 3b. y ab Block 1 Block 2 Block b Figure 4-1 The randomized complete block design

32 ANOVA partitioning of total variability: The degrees of freedom for the sums of squares above are as follows Ratios of sums of squares to their degrees of freedom result in mean squares. 32

33 33 The ratio of the mean square for treatments to the error mean square is an F statistic. If H o is true, the test statistic F o =MS Treatment /MS E  F a-1, (a-1)(b-1)

34 For manual computing, use 34


Download ppt "1 Chapter 5.8 What if We Have More Than Two Samples?"

Similar presentations


Ads by Google