Presentation is loading. Please wait.

Presentation is loading. Please wait.

Chapter 12 Introduction to Analysis of Variance

Similar presentations


Presentation on theme: "Chapter 12 Introduction to Analysis of Variance"— Presentation transcript:

1 Chapter 12 Introduction to Analysis of Variance
PowerPoint Lecture Slides Essentials of Statistics for the Behavioral Sciences Eighth Edition by Frederick J. Gravetter and Larry B. Wallnau

2 Chapter 12 Learning Outcomes
Explain purpose and logic of Analysis of Variance 2 Perform Analysis of Variance on data from single-factor study 3 Know when and why to use post hoc tests (posttests) 4 Compute Tukey’s HSD and Scheffé test post hoc tests 5 Compute η2 to measure effect size

3 Tools You Will Need Variability (Chapter 4)
Sum of squares Sample variance Degrees of freedom Introduction to hypothesis testing (Chapter 8) The logic of hypothesis testing Independent-measures t statistic (Chapter 10)

4 12.1 Introduction to Analysis of Variance
Used to evaluate mean differences between two or more treatments Uses sample data as basis for drawing general conclusions about populations Clear advantage over a t test: it can be used to compare more than two treatments at the same time

5 Figure 12.1 Typical Situation for Using ANOVA
FIGURE A typical situation in which ANOVA would be used. Three separate samples are obtained to evaluate the mean differences among three populations (or treatments) with unknown means.

6 Terminology Factor Levels Factorial design
The independent (or quasi-independent) variable that designates the groups being compared Levels Individual conditions or values that make up a factor Factorial design A study that combines two or more factors

7 Figure 12.2 Two-Factor Research Design
FIGURE A research design with two factors. The research study uses two factors: One factor uses two levels of therapy technique (I versus II), and the second factor uses three levels of time (before, after and 6 months after). Also notice that the therapy factor uses two separate groups (independent measures) and the time factor uses the same group for all three levels (repeated measures).

8 Statistical Hypotheses for ANOVA
Null hypothesis: the level or value on the factor does not affect the dependent variable In the population, this is equivalent to saying that the means of the groups do not differ from each other

9 Alternate Hypothesis for ANOVA
H1: There is at least one mean difference among the populations (Acceptable shorthand is “Not H0”) Issue: how many ways can H0 be wrong? All means are different from every other mean Some means are not different from some others, but other means do differ from some means Some instructors may want to spend some time listing some of the ways that the null hypothesis cold be wrong that do NOT include μ1 ≠ μ2 ≠ μ3 because many students make the logic error of assuming that if null is wrong then the ONLY way it can be wrong is if all the groups are significantly different from each other.

10 Test statistic for ANOVA
F-ratio is based on variance instead of sample mean differences

11 Test statistic for ANOVA
Not possible to compute a sample mean difference between more than two samples F-ratio based on variance instead of sample mean difference Variance used to define and measure the size of differences among sample means (numerator) Variance in the denominator measures the mean differences that would be expected if there is no treatment effect

12 Type I Errors and Multiple-Hypothesis tests
Why ANOVA (if t can compare two means)? Experiments often require multiple hypothesis tests—each with Type I error (testwise alpha) Type I error for a set of tests accumulates testwise alpha  experimentwise alpha > testwise alpha ANOVA evaluates all mean differences simultaneously with one test—regardless of the number of means—and thereby avoids the problem of inflated experimentwise alpha Testwise Type I error rate and Experimentwise Type I error rate are sufficiently complex concepts that some instructors may want to expand the lecture materials on this topic.

13 12.2 Analysis of Variance Logic
Between-treatments variance Variability results from general differences between the treatment conditions Variance between treatments measures differences among sample means Within-treatments variance Variability within each sample Individual scores are not the same within each sample

14 Sources of Variability Between Treatments
Systematic differences caused by treatments Random, unsystematic differences Individual differences Experimental (measurement) error

15 Sources of Variability Within Treatments
No systematic differences related to treatment groups occur within each group Random, unsystematic differences Individual differences Experimental (measurement) error

16 Figure 12.3 Total Variability Partitioned into Two Components
FIGURE The independent-measures ANOVA partitions, or analyzes, the total variability into two components: variance between treatments and variance within treatments.

17 F-ratio If H0 is true: If H1 is true:
Size of treatment effect is near zero F is near 1.00 If H1 is true: Size of treatment effect is more than 0. F is noticeably larger than 1.00 Denominator of the F-ratio is called the error term

18 Learning Check Decide if each of the following statements is True or False T/F ANOVA allows researchers to compare several treatment conditions without conducting several hypothesis tests If the null hypothesis is true, the F-ratio for ANOVA is expected (on average) to have a value of 0

19 Learning Check - Answers
True Several conditions can be compared in one test False If the null hypothesis is true, the F-ratio will have a value near 1.00

20 Table (p. 344) Hypothetical data from an experiment examining learning performance under three temperature conditions. Table (p. 344) Essentials of Statistics for Behavioral Science, 6th Edition by Frederick Gravetter and Larry Wallnau Copyright  2008 Wadsworth Publishing, a division of Thomson Learning. All rights reserved.

21 12.3 ANOVA Notation and Formulas
Number of treatment conditions: k Number of scores in each treatment: n1, n2… Total number of scores: N When all samples are same size, N = kn Sum of scores (ΣX) for each treatment: T Grand total of all scores in study: G = ΣT No universally accepted notation for ANOVA; Other sources may use other symbols

22 Figure 12.4 ANOVA Calculation Structure and Sequence
FIGURE The structure and sequence of calculations for the ANOVA.

23 Figure 12.5 Partitioning SS for Independent-measures ANOVA
FIGURE Partitioning the sum of squares (SS) for the independent-measures ANOVA.

24 ANOVA equations

25 Degrees of Freedom Analysis
Total degrees of freedom dftotal= N – 1 Within-treatments degrees of freedom dfwithin= N – k Between-treatments degrees of freedom dfbetween= k – 1

26 Figure 12.6 Partitioning Degrees of Freedom
FIGURE Partitioning degrees of freedom (df) for the independent-measures ANOVA.

27 Figure 12.11 Formulas for ANOVA

28 Mean Squares and F-ratio

29 ANOVA Summary Table Concise method for presenting ANOVA results
Helps organize and direct the analysis process Convenient for checking computations “Standard” statistical analysis program output Source SS df MS F Between Treatments 40 2 20 10 Within Treatments Total 60 12 Some instructors may want to begin the discussion of the computational procedures with this slide. Students sometimes get lost in all the formulas, but if they can see that the summary table is the “goal” of the process, they can break it down into the small steps required to fill in each part of the table.

30 Table (p. 344) Hypothetical data from an experiment examining learning performance under three temperature conditions. Table (p. 344) Essentials of Statistics for Behavioral Science, 6th Edition by Frederick Gravetter and Larry Wallnau Copyright  2008 Wadsworth Publishing, a division of Thomson Learning. All rights reserved.

31 Table (p. 354) A portion of the F distribution table. Entries in roman type are critical values for the .05 level of significance, and bold type values are for the .01 level of significance. The critical values for df = 2, 12 have been highlighted (see text). Table (p. 354) Essentials of Statistics for Behavioral Science, 6th Edition by Frederick Gravetter and Larry Wallnau Copyright  2008 Wadsworth Publishing, a division of Thomson Learning. All rights reserved.

32 Table (p. 354) The effect of drug treatment on the amount of time (in seconds) a stimulus is endured. Table (p. 354) Essentials of Statistics for Behavioral Science, 6th Edition by Frederick Gravetter and Larry Wallnau Copyright  2008 Wadsworth Publishing, a division of Thomson Learning. All rights reserved.

33 Figure (p. 365) Hypothetical results from a research study comparing three treatment conditions. Summary statistics are presented for each treatment along with the outcome from the analysis of variance. Figure (p. 365) Essentials of Statistics for Behavioral Science, 6th Edition by Frederick Gravetter and Larry Wallnau Copyright  2008 Wadsworth Publishing, a division of Thomson Learning. All rights reserved.

34 Table (p. 365) Hypothetical results from a research study comparing three treatment conditions. Summary statistics are presented for each treatment along with the outcome from the analysis of variance. Table (p. 365) Essentials of Statistics for Behavioral Science, 6th Edition by Frederick Gravetter and Larry Wallnau Copyright  2008 Wadsworth Publishing, a division of Thomson Learning. All rights reserved.

35 Figure 3.10 (p. 278) Figure 3.10 (p. 398) Formulas for ANOVA.
Essentials of Statistics for Behavioral Science, 6th Edition by Frederick Gravetter and Larry Wallnau Copyright  2008 Wadsworth Publishing, a division of Thomson Learning. All rights reserved.

36 Learning Check An analysis of variance produces SStotal = 80 and SSwithin = 30. For this analysis, what is SSbetween? A 50 B 110 C 2400 D More information is needed

37 Learning Check - Answer
An analysis of variance produces SStotal = 80 and SSwithin = 30. For this analysis, what is SSbetween? A 50 B 110 C 2400 D More information is needed

38 12.4 Distribution of F-ratios
If the null hypothesis is true, the value of F will be around 1.00 Because F-ratios are computed from two variances, they are always positive numbers Table of F values is organized by two df df numerator (between) shown in table columns df denominator (within) shown in table rows

39 Figure 12.7 Distribution of F-ratios
FIGURE The distribution of F-ratios with df = 2, 12. Of all the values in the distribution, only 5% are larger than F = 3.88, and only 1% are larger than F = 6.93.

40 12.5 Examples of Hypothesis Testing and Effect Size
Hypothesis tests use the same four steps that have been used in earlier hypothesis tests. Computation of the test statistic F is done in stages Compute SStotal, SSbetween, SSwithin Compute MStotal, MSbetween, MSwithin Compute F

41 Figure 12.8 Critical region for α=.01 in Distribution of F-ratios
FIGURE The distribution of F-ratios with df = 3, 20. The critical values for α = .01 is 4.94.

42 Measuring Effect size for ANOVA
Compute percentage of variance accounted for by the treatment conditions In published reports of ANOVA, effect size is usually called η2 (“eta squared”) r2 concept (proportion of variance explained)

43 In the Literature Treatment means and standard deviations are presented in text, table or graph Results of ANOVA are summarized, including F and df p-value η2 E.g., F(3,20) = 6.45, p<.01, η2 = 0.492

44 Figure 12.9 Visual Representation of Between & Within Variability
A visual representation of the between-treatments variability and the within0-treatments variability that form the numerator and denominator, respectively, of the F-ratio. In (a), the difference between treatments is relatively large and easy to see. In (b), the same 4-point difference between treatments is relatively small and is overwhelmed by the within-treatments variability.

45 MSwithin and Pooled Variance
In the t-statistic and in the F-ratio, the variances from the separate samples are pooled together to create one average value for the sample variance Numerator of F-ratio measures how much difference exists between treatment means. Denominator measures the variance of the scores inside each treatment

46 post hoc Tests ANOVA compares all individual mean differences simultaneously, in one test A significant F-ratio indicates that at least one difference in means is statistically significant Does not indicate which means differ significantly from each other! post hoc tests are follow up tests done to determine exactly which mean differences are significant, and which are not

47 Experimentwise Alpha post hoc tests compare two individual means at a time (pairwise comparison) Each comparison includes risk of a Type I error Risk of Type I error accumulates and is called the experimentwise alpha level. Increasing the number of hypothesis tests increases the total probability of a Type I error post hoc (“posttests”) use special methods to try to control experimentwise Type I error rate

48 Tukey’s Honestly Significant Difference
A single value that determines the minimum difference between treatment means that is necessary to claim statistical significance–a difference large enough that p < αexperimentwise Honestly Significant Difference (HSD)

49 The Scheffé Test The Scheffé test is one of the safest of all possible post hoc tests Uses an F-ratio to evaluate significance of the difference between two treatment conditions

50 Learning Check Which combination of factors is most likely to produce a large value for the F-ratio? A large mean differences and large sample variances B large mean differences and small sample variances C small mean differences and large sample variances D small mean differences and small sample variances

51 Learning Check - Answer
Which combination of factors is most likely to produce a large value for the F-ratio? A large mean differences and large sample variances B large mean differences and small sample variances C small mean differences and large sample variances D small mean differences and small sample variances

52 Learning Check Decide if each of the following statements is True or False T/F Post tests are needed if the decision from an analysis of variance is “fail to reject the null hypothesis” A report shows ANOVA results: F(2, 27) = 5.36, p < .05. You can conclude that the study used a total of 30 participants

53 Learning Check - Answers
False post hoc tests are needed only if you reject H0 (indicating at least one mean difference is significant) True Because dftotal = N-1 and Because dftotal = dfbetween + dfwithin

54 12.7 Relationship between ANOVA and t tests
For two independent samples, either t or F can be used Always result in same decision F = t2 For any value of α, (tcritical)2 = Fcritical

55 Figure 12.10 Distribution of t and F statistics
FIGURE The distribution of t statistics with df = 18 and the corresponding distribution of F-ratios with df = 1, 18. Notice that the critical values for α = .05 are t = ±2.101 and F = = 4.41.

56 Independent Measures ANOVA Assumptions
The observations within each sample must be independent The population from which the samples are selected must be normal The populations from which the samples are selected must have equal variances (homogeneity of variance) Violating the assumption of homogeneity of variance risks invalid test results

57 Figure 12.12 Distribution of t and F statistics
FIGURE SPSS Output of the ANOVA for the studying strategy experiment in Example 12.1.

58 Any Questions? Concepts? Equations?


Download ppt "Chapter 12 Introduction to Analysis of Variance"

Similar presentations


Ads by Google