One-Way Analysis of Variance Recapitulation Recapitulation 1. Comparing differences among three or more subsamples requires a different statistical test.

Slides:



Advertisements
Similar presentations
Multiple-choice question
Advertisements

ANALYSIS OF VARIANCE (ONE WAY)
ANOVA notes NR 245 Austin Troy
Conceptual Review Conceptual Formula, Sig Testing Calculating in SPSS
Lecture 10 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
Using Statistics in Research Psych 231: Research Methods in Psychology.
ANOVA Analysis of Variance: Why do these Sample Means differ as much as they do (Variance)? Standard Error of the Mean (“variance” of means) depends upon.
Experimental Design & Analysis
Analysis of Variance: Inferences about 2 or More Means
Comparing Means.
ANalysis Of VAriance (ANOVA) Comparing > 2 means Frequently applied to experimental data Why not do multiple t-tests? If you want to test H 0 : m 1 = m.
Chapter 3 Analysis of Variance
PSY 307 – Statistics for the Behavioral Sciences
Lecture 9: One Way ANOVA Between Subjects
Two Groups Too Many? Try Analysis of Variance (ANOVA)
8. ANALYSIS OF VARIANCE 8.1 Elements of a Designed Experiment
One-way Between Groups Analysis of Variance
Anthony J Greene1 ANOVA: Analysis of Variance 1-way ANOVA.
Statistical Methods in Computer Science Hypothesis Testing II: Single-Factor Experiments Ido Dagan.
1 Chapter 13: Introduction to Analysis of Variance.
Chap 10-1 Analysis of Variance. Chap 10-2 Overview Analysis of Variance (ANOVA) F-test Tukey- Kramer test One-Way ANOVA Two-Way ANOVA Interaction Effects.
Chapter 12: Analysis of Variance
Analysis of Variance (ANOVA) Quantitative Methods in HPELS 440:210.
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
The basic idea So far, we have been comparing two samples
COURSE: JUST 3900 INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE Instructor: Dr. John J. Kerbs, Associate Professor Joint Ph.D. in Social Work and Sociology.
Stats Lunch: Day 7 One-Way ANOVA. Basic Steps of Calculating an ANOVA M = 3 M = 6 M = 10 Remember, there are 2 ways to estimate pop. variance in ANOVA:
One-Way Analysis of Variance Comparing means of more than 2 independent samples 1.
Analysis of Covariance ANOVA is a class of statistics developed to evaluate controlled experiments. Experimental control, random selection of subjects,
Chapter 11 HYPOTHESIS TESTING USING THE ONE-WAY ANALYSIS OF VARIANCE.
 The idea of ANOVA  Comparing several means  The problem of multiple comparisons  The ANOVA F test 1.
Chapter 12: Introduction to Analysis of Variance
t(ea) for Two: Test between the Means of Different Groups When you want to know if there is a ‘difference’ between the two groups in the mean Use “t-test”.
PSY 307 – Statistics for the Behavioral Sciences Chapter 16 – One-Factor Analysis of Variance (ANOVA)
Statistics 11 Confidence Interval Suppose you have a sample from a population You know the sample mean is an unbiased estimate of population mean Question:
ANOVA (Analysis of Variance) by Aziza Munir
Testing Multiple Means and the Analysis of Variance (§8.1, 8.2, 8.6) Situations where comparing more than two means is important. The approach to testing.
5-5 Inference on the Ratio of Variances of Two Normal Populations The F Distribution We wish to test the hypotheses: The development of a test procedure.
T- and Z-Tests for Hypotheses about the Difference between Two Subsamples.
Testing Hypotheses about Differences among Several Means.
Chapter 14 – 1 Chapter 14: Analysis of Variance Understanding Analysis of Variance The Structure of Hypothesis Testing with ANOVA Decomposition of SST.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
I. Statistical Tests: A Repetive Review A.Why do we use them? Namely: we need to make inferences from incomplete information or uncertainty þBut we want.
The Completely Randomized Design (§8.3)
Analysis of Variance 1 Dr. Mohammed Alahmed Ph.D. in BioStatistics (011)
Chapter 14 – 1 Chapter 14: Analysis of Variance Understanding Analysis of Variance The Structure of Hypothesis Testing with ANOVA Decomposition of SST.
Chapter 13 - ANOVA. ANOVA Be able to explain in general terms and using an example what a one-way ANOVA is (370). Know the purpose of the one-way ANOVA.
One-way ANOVA: - Comparing the means IPS chapter 12.2 © 2006 W.H. Freeman and Company.
Chapter Seventeen. Figure 17.1 Relationship of Hypothesis Testing Related to Differences to the Previous Chapter and the Marketing Research Process Focus.
Copyright © Cengage Learning. All rights reserved. 12 Analysis of Variance.
Chapter 12 Introduction to Analysis of Variance PowerPoint Lecture Slides Essentials of Statistics for the Behavioral Sciences Eighth Edition by Frederick.
Experimental Statistics - week 3
Psych 230 Psychological Measurement and Statistics Pedro Wolf November 18, 2009.
Hypothesis test flow chart frequency data Measurement scale number of variables 1 basic χ 2 test (19.5) Table I χ 2 test for independence (19.9) Table.
McGraw-Hill, Bluman, 7th ed., Chapter 12
IE241: Introduction to Design of Experiments. Last term we talked about testing the difference between two independent means. For means from a normal.
Other Types of t-tests Recapitulation Recapitulation 1. Still dealing with random samples. 2. However, they are partitioned into two subsamples. 3. Interest.
Research Methods and Data Analysis in Psychology Spring 2015 Kyle Stephenson.
Introduction to ANOVA Research Designs for ANOVAs Type I Error and Multiple Hypothesis Tests The Logic of ANOVA ANOVA vocabulary, notation, and formulas.
Oneway/Randomized Block Designs Q560: Experimental Methods in Cognitive Science Lecture 8.
The p-value approach to Hypothesis Testing
EDUC 200C Section 9 ANOVA November 30, Goals One-way ANOVA Least Significant Difference (LSD) Practice Problem Questions?
Significance Tests for Regression Analysis. A. Testing the Significance of Regression Models The first important significance test is for the regression.
Formula for Linear Regression y = bx + a Y variable plotted on vertical axis. X variable plotted on horizontal axis. Slope or the change in y for every.
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 4 Investigating the Difference in Scores.
Inferential Statistics Psych 231: Research Methods in Psychology.
Chapter 12 Introduction to Analysis of Variance
The 2 nd to last topic this year!!.  ANOVA Testing is similar to a “two sample t- test except” that it compares more than two samples to one another.
i) Two way ANOVA without replication
Chapter 13 Group Differences
Presentation transcript:

One-Way Analysis of Variance

Recapitulation Recapitulation 1. Comparing differences among three or more subsamples requires a different statistical test than either z-tests or t-tests. 2. The solution is to perform an analysis of variance (ANOVA). 3. ANOVA involves the comparison of two estimates for the population variance. 4. One variance estimate captures only the random differences among sampled units, the other these random differences plus the effects of being in the different subsamples. 5. The ratio between the two estimated variances is evaluated using the F-statistic sampling distributions.

Recapitulation (continued) Recapitulation (continued) 6. ANOVA is based on the general linear model. 7. The general linear model is: Y ij =  +  j X ij +  ij where X ij is the subgroup difference and  j is a constant estimating its effect on Y ij. 8. When subgroup differences do not exist,  j = The null hypothesis is: H 0 :  1 =  2 =  3 =...  j

As an example, consider an experiment on worker productivity in an introductory psychology class. Thirty students were randomly selected for the experiment from PSYCH 100 and randomly assigned to one of three subgroups. The productivity measure (Y ij ) was the number of puzzles that these students solved in a fixed period of time. The three experimental conditions (treatments, X ij ) were: left alone to solve puzzles; solving puzzles in the presence of the other nine group members (so that each subject could observe her or his own rate of puzzle solving); and solving puzzles in the presence of other subjects AND in the presence of a monitor, meant to simulate a supervisor. The results look like this:

————————————————————————————————————— Not Monitored Alone Not Monitored Together Monitored Subject Y i,1 Subject Y i,2 Subject Y i,3 ————————————————————————————————————— ————————————————————————————————————— N 1 = 10 N 2 = 10 N 3 = 10  1 = 118  2 = 100  3 = 82 _ __ Y 1 = 11.8Y 2 = 10.0Y 3 = 8.2 _ Y = 10.0

Our hypothesis (H 1 ) is that working conditions affect worker performance in ways that we do not fully understand: H 1 :  1   2   3 Our null hypothesis (H 0 ) is that worker performance is unaffected by working conditions: H 0 :  1 =  2 =  3 Since a comparison of THREE subgroup means is required, t-tests are inappropriate. The approach known generically as the analysis of variance must be used.

————————————————————————————————————— Not Monitored Alone Not Monitored Together Monitored Subject Y i,1 Subject Y i,2 Subject Y i,3 ————————————————————————————————————— ————————————————————————————————————— N 1 = 10 N 2 = 10 N 3 = 10  1 = 118  2 = 100  3 = 82 _ __ Y 1 = 11.8Y 2 = 10.0Y 3 = 8.2 _ Y = 10.0

First we calculate the total sum of squares: We begin with the first score in the first group and continue through the 30th score in the third group, as follows: SS Total = ( ) 2 + ( ) ( ) 2 + ( ) ( ) 2 + ( ) ( ) 2 = 116

Next we calculate the sum of squares between, as follows : For the first of the three subgroups, we find the difference between the group mean and the grand mean, square that difference then multiply it by the size of the subgroup, then do the same for the other two subgroups. Then we sum these three products, as follows: SS Between = 10( ) ( ) ( ) 2 = 64.8

————————————————————————————————————— Not Monitored Alone Not Monitored Together Monitored Subject Y i,1 Subject Y i,2 Subject Y i,3 ————————————————————————————————————— ————————————————————————————————————— N 1 = 10 N 2 = 10 N 3 = 10  1 = 118  2 = 100  3 = 82 _ __ Y 1 = 11.8Y 2 = 10.0Y 3 = 8.2 _ Y = 10.0

Finally, we calculate the sum of squares within: This means that we find the squared difference between each of the ten scores in the first group and the mean for the first group, then the squared difference between the ten scores in the second group and the mean for the SECOND group, then the squared difference between the ten scores in the third group and the mean for the THIRD group, and finally add all 30 squared differences together: SS Within = ( ) 2 + ( ) ( ) 2 + ( ) ( ) 2 + ( ) ( ) 2 = 51.2

To check our calculations, remember the identity Total SS = Between SS + Within SS 116 = Next, we need the degrees of freedom. Total degrees of freedom is simply number of cases less one, N - 1. Here, there are 30 cases, so there are 29 total degrees of freedom. For degrees of freedom between, the three subgroup means are treated as scores, so there are J - 1 across subgroups, here 3 - 1, giving us 2 degrees of freedom between. Finally, we lose a degree of freedom by partitioning into subgroups, i.e., N - J. Here we have three subgroups, so we lose a degree of freedom for each giving us or 27 degrees of freedom within.

Analysis of variance results by convention are reported in what is called an "ANOVA summary table": ————————————————————————————————————— SourceSS dfMean Square F ————————————————————————————————————— Between Groups Within Groups Total —————————————————————————————————————

We perform a significance test in the usual way, first by selecting alpha, then locating the appropriate sampling distribution, finding the critical value, and comparing this value to the value of the F-statistic. With alpha = 0.05, we find Appendix 3, p In this example we have 2 and 27 degrees of freedom. The table of critical values has degrees of freedom between as COLUMN headings (n 1 ) and degrees of freedom within as ROW headings (n 2 ). In column 2, row 27 we find the critical value to be Since our F-value is 17.05, GREATER than 3.35, we know that it lies well inside the region of rejection, hence we REJECT the null hypothesis at the 0.05 level. Substantively, this means that we infer that the conditions under which one performs a task DO have an effect on performance.

The F-test is a significance test, an inferential statistic. It tells us only whether or not exposure to the treatment variable has measurable consequences that are different from chance. It does NOT tell us about the strength of association between the treatment (X ij ) and the dependent variable, Y ij. For this we need a measure of association. The sum of squares BETWEEN represents the variance attributable to the treatment variable, X ij. The TOTAL sum of squares expresses the total amount of variance in the dependent variable, Y ij, that is, the total variance "to be explained" statistically. A ratio of the two is a straightforward description of the percentage of variance in Y ij accounted for by its association with X ij. Statistically this is called R-square.

From the example above, the sum of squares between is and the total sum of squares is Thus, R-square is : The F-test tells us that treatment categories (working conditions) differ in ways that cannot be explained as chance. R-square tells us that 56 percent of the variation in task performance is associated with differences in working conditions.

Knowing that the treatment variable has a statistically significant effect does not tell us WHICH specific treatment category or categories have greater impact than others. In our example, we know only that AT LEAST ONE of the puzzle-solving conditions differs from one (or both) of the remaining two, but we do not know which. In other words, we do not know which of the following alternative hypotheses is (are) supported:  1   2 =  3  1 =  2   3  1   3 =  2  1   2   3 We need a way to statistically compare the subgroups.

There are two strategies: comparisons explicitly planned in advance are called a priori tests; those performed after an initial ANOVA are called post hoc comparison tests. Of the latter, we will use only the method known as the Scheffé test. The Scheffé method creates a “threshold” for comparing subgroup means (once an ANOVA null hypothesis has been rejected) called the “minimum significant difference.” Differences between two subgroup means that exceed this minimum significant difference are statistically significant; that is, their difference appears to be “real” rather than due to chance. The algorithm is in Sirkin (1999), p. 333.

where _ _ Y j and Y j+1 are the subsample means being compared df Between is degrees of freedom between in the ANOVA F  is the critical value of F at the significance level (  ) chosen for the comparison MS Within is the ANOVA mean square within and n j and n j+1 are the sizes two subsamples being compared

In the puzzle-solving example, _ _ _ Y 1 = 11.8, Y 2 = 10.0, and Y 3 = 8.2 df Between = 2 F  = 2.51 (  =.10, df = 2, 27) MS Within = 1.90 and n 1 = n 2 = n 2 = 10 Hence,

The value is the “minimum significant difference,” the “threshold” we use to compare subsample means with  set at [Sirkin (1999) contains no 0.10 F table.] Here is how Sirkin would organize our comparison tests: _ _ H 0 |Y j – Y j+1 | = Critical Value Conclusion ——————————————————————————  1 =  2 |11.8 – 10.0| = 1.80 > Reject H 0  2 =  3 |10.0 – 8.2| = 1.80 > Reject H 0  1 =  3 |11.8 – 8.2| = 3.60 > Reject H 0

Sample SAS Program Puzzle-Solving Example libname old 'a:\'; libname library 'a:\'; options nodate nonumber ps=66; proc glm data=old.example; class setting; model puzzles = setting; means setting / scheffe alpha = 0.1; contrast 'Alone vs. Together' setting ; contrast 'Alone vs. Monitor' setting ; contrast 'Together vs. Monitor' setting ; contrast 'Alone vs. Others' setting ; contrast 'Together vs. Others' setting ; title1 'ANOVA With Comparison Tests'; run;

ANOVA With Comparison Tests General Linear Models Procedure Class Level Information Class Levels Values SETTING 3 (1) alone (2) monitor (3) together Number of observations in data set = 30

ANOVA With Comparison Tests General Linear Models Procedure Dependent Variable: PUZZLES Sum of Mean Source DF Squares Square F Value Pr > F Model Error Corrected Total R-Square C.V. Root MSE PUZZLES Mean Source DF Type I SS Mean Square F Value Pr > F SETTING Source DF Type III SS Mean Square F Value Pr > F SETTING

ANOVA With Comparison Tests General Linear Models Procedure Scheffe's test for variable: PUZZLES NOTE: This test controls the type I experimentwise error rate but generally has a higher type II error rate than REGWF for all pairwise comparisons Alpha = 0.1 df = 27 MSE = Critical Value of F = Minimum Significant Difference = 1.38 Means with the same letter are not significantly different. Scheffe Grouping Mean N SETTING A alone B together C monitor

ANOVA With Comparison Tests General Linear Models Procedure Scheffe's test for variable: PUZZLES NOTE: This test controls the type I experimentwise error rate but generally has a higher type II error rate than REGWF for all pairwise comparisons Alpha= 0.1 df= 27 MSE= Critical Value of F= Minimum Significant Difference= 1.38 Means with the same letter are not significantly different. Scheffe Grouping Mean N SETTING A alone B together B B monitor

ANOVA With Comparison Tests General Linear Models Procedure Scheffe's test for variable: PUZZLES NOTE: This test controls the type I experimentwise error rate but generally has a higher type II error rate than REGWF for all pairwise comparisons Alpha= 0.1 df= 27 MSE= Critical Value of F= Minimum Significant Difference= 1.38 Means with the same letter are not significantly different. Scheffe Grouping Mean N SETTING A alone A A together A A monitor

ANOVA With Comparison Tests General Linear Models Procedure Dependent Variable: PUZZLES Contrast DF Contrast SS Mean Square F Value Pr > F Alone vs. Together Alone vs. Monitor Together vs. Monitor Alone vs. Others Together vs. Others

One-Way Analysis of Variance Exercise Four groups of randomly selected and randomly assigned students were taught a basic course in statistics by four different methods. A standardized test was given at the end of the semester to all four groups. Evaluate the differences in teaching approaches using the Analysis of Variance. Assume that  = 0.05, and use the F distribution (Appendix 3, p. 544). Group 1Group 2Group 3Group Expressed symbolically, what is the null hypothesis? ______________ 2. What is the value of the sum of squares between? ______________ 3. What is the value of the sum of squares within? ______________ 4. How many degrees of freedom between? ______________ 5. How many degrees of freedom within? ______________ 6. What is the value of the mean square between? ______________ 7. What is the value of the mean square within? ______________ 8. What is the value of the F-ratio? ______________ 9. What is the critical value of F? ______________ 10. Do you reject the null hypothesis? ______________

One-Way Analysis of Variance Exercise Answers Four groups of randomly selected and randomly assigned students were taught a basic course in statistics by four different methods. A standardized test was given at the end of the semester to all four groups. Evaluate the differences in teaching approaches using the Analysis of Variance. Assume that  = 0.05, and use the F distribution (Appendix 3, p. 544). Group 1Group 2Group 3Group Expressed symbolically, what is the null hypothesis?  1 =  2 =  3 =  4 2. What is the value of the sum of squares between? What is the value of the sum of squares within? How many degrees of freedom between? 3 5. How many degrees of freedom within? What is the value of the mean square between? What is the value of the mean square within? What is the value of the F-ratio? What is the critical value of F? Do you reject the null hypothesis? Yes, Reject