Analysis of Variance: A Difference of Means Tests for Two or More Levels of an IV An analysis of variance looks for the causal impact of a nominal level.

Slides:



Advertisements
Similar presentations
Lecture 2 ANALYSIS OF VARIANCE: AN INTRODUCTION
Advertisements

ANALYSIS OF VARIANCE (ONE WAY)
1-Way Analysis of Variance
Independent t -test Features: One Independent Variable Two Groups, or Levels of the Independent Variable Independent Samples (Between-Groups): the two.
PTP 560 Research Methods Week 9 Thomas Ruediger, PT.
Two-Way ANOVA. Two-way Analysis of Variance  Two-way ANOVA is applied to a situation in which you have two independent nominal-level variables and one.
Analysis of variance (ANOVA)-the General Linear Model (GLM)
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 12: Analysis of Variance: Differences among Means of Three or More Groups.
Comparing k Populations Means – One way Analysis of Variance (ANOVA)
ANOVA: Analysis of Variance
ANOVA notes NR 245 Austin Troy
Conceptual Review Conceptual Formula, Sig Testing Calculating in SPSS
Lecture 10 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
Practice for the Mid-Term
ANOVA Analysis of Variance: Why do these Sample Means differ as much as they do (Variance)? Standard Error of the Mean (“variance” of means) depends upon.
Chi-square Test of Independence
Lecture 9: One Way ANOVA Between Subjects
Two Groups Too Many? Try Analysis of Variance (ANOVA)
One-way Between Groups Analysis of Variance
Anthony J Greene1 ANOVA: Analysis of Variance 1-way ANOVA.
Inferences About Process Quality
Analysis of Variance & Multivariate Analysis of Variance
Introduction to Analysis of Variance (ANOVA)
Repeated Measures ANOVA Used when the research design contains one factor on which participants are measured more than twice (dependent, or within- groups.
Two-Way Analysis of Variance STAT E-150 Statistical Methods.
Inferential Statistics
Chapter 12: Analysis of Variance
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
Inferential Statistics: SPSS
ANOVA Greg C Elvers.
1 of 46 MGMT 6970 PSYCHOMETRICS © 2014, Michael Kalsher Michael J. Kalsher Department of Cognitive Science Inferential Statistics IV: Factorial ANOVA.
Stats Lunch: Day 7 One-Way ANOVA. Basic Steps of Calculating an ANOVA M = 3 M = 6 M = 10 Remember, there are 2 ways to estimate pop. variance in ANOVA:
Chapter 11 HYPOTHESIS TESTING USING THE ONE-WAY ANALYSIS OF VARIANCE.
ANOVA (Analysis of Variance) by Aziza Munir
Psychology 301 Chapters & Differences Between Two Means Introduction to Analysis of Variance Multiple Comparisons.
Stats Lunch: Day 4 Intro to the General Linear Model and Its Many, Many Wonders, Including: T-Tests.
Testing Hypotheses about Differences among Several Means.
ANOVA Conceptual Review Conceptual Formula, Sig Testing Calculating in SPSS.
Inferential Statistics
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
One-Way Analysis of Variance
6/2/2016Slide 1 To extend the comparison of population means beyond the two groups tested by the independent samples t-test, we use a one-way analysis.
Analysis of Variance 1 Dr. Mohammed Alahmed Ph.D. in BioStatistics (011)
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
ANOVA: Analysis of Variance.
Chapter 13 - ANOVA. ANOVA Be able to explain in general terms and using an example what a one-way ANOVA is (370). Know the purpose of the one-way ANOVA.
Chapter 14 Repeated Measures and Two Factor Analysis of Variance
Analysis of Variance (One Factor). ANOVA Analysis of Variance Tests whether differences exist among population means categorized by only one factor or.
Chi-square Test of Independence
1 ANALYSIS OF VARIANCE (ANOVA) Heibatollah Baghi, and Mastee Badii.
Chapter Seventeen. Figure 17.1 Relationship of Hypothesis Testing Related to Differences to the Previous Chapter and the Marketing Research Process Focus.
Chapter 12 Introduction to Analysis of Variance PowerPoint Lecture Slides Essentials of Statistics for the Behavioral Sciences Eighth Edition by Frederick.
Chapter 13 Repeated-Measures and Two-Factor Analysis of Variance
ONE-WAY BETWEEN-GROUPS ANOVA Psyc 301-SPSS Spring 2014.
Hypothesis test flow chart frequency data Measurement scale number of variables 1 basic χ 2 test (19.5) Table I χ 2 test for independence (19.9) Table.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Lecture Slides Elementary Statistics Eleventh Edition and the Triola.
Social Science Research Design and Statistics, 2/e Alfred P. Rovai, Jason D. Baker, and Michael K. Ponton Between Subjects Analysis of Variance PowerPoint.
One-Way Analysis of Variance Recapitulation Recapitulation 1. Comparing differences among three or more subsamples requires a different statistical test.
Introduction to ANOVA Research Designs for ANOVAs Type I Error and Multiple Hypothesis Tests The Logic of ANOVA ANOVA vocabulary, notation, and formulas.
Handout Six: Sample Size, Effect Size, Power, and Assumptions of ANOVA EPSE 592 Experimental Designs and Analysis in Educational Research Instructor: Dr.
Analysis of Variance STAT E-150 Statistical Methods.
Analysis of variance Tron Anders Moger
Formula for Linear Regression y = bx + a Y variable plotted on vertical axis. X variable plotted on horizontal axis. Slope or the change in y for every.
HYPOTHESIS TESTING FOR DIFFERENCES BETWEEN MEANS AND BETWEEN PROPORTIONS.
 List the characteristics of the F distribution.  Conduct a test of hypothesis to determine whether the variances of two populations are equal.  Discuss.
Chapter 12 Introduction to Analysis of Variance
Comparing Three or More Means
Statistics for the Social Sciences
Chapter 13 Group Differences
One way ANOVA One way Analysis of Variance (ANOVA) is used to test the significance difference of mean of one dependent variable across more than two.
Presentation transcript:

Analysis of Variance: A Difference of Means Tests for Two or More Levels of an IV An analysis of variance looks for the causal impact of a nominal level independent variable (factor) on an interval or better level dependent variable The basic question you seek to answer with an difference of means test is whether or not there is a difference in scores on the dependent variable attributable to membership in one or the other category of the independent variable

Types of Difference of Means Tests Varieties of test for difference of means where there is a single independent variable or factor t-test: two levels of the independent variable What is the impact of gender (M, F) on annual salary? Analysis of Variance (ANOVA): Two or more levels or conditions of the independent variable What is the impact of ethnicity (Hispanic, African-American, Asian- Pacific Islander, Caucasian, etc) on annual salary? ANOVA models can be fixed or random Fixed model overwhelmingly used Effects obtained in the fixed model only generalizable to other identical levels of the factor studied (e.g., only to treatments A, B, C such as online vs. classroom instruction) Effects obtained in the random model generalizable to a wider range of values of the IV than just the three levels Time of day could be a random factor and you randomly decide to compare classes taught at 8 am, noon, and 3 pm but these values are “replaceable” by other randomly drawn values or you could add more time periods Subject matter or teacher could be another random factor

Repeated Measures and Analysis of Covariance In a repeated measures ANOVA design, the same Ss are tested across different levels of the factor (example, time 1, time 2, time 3, …time n) In an analysis of covariance, we statistically control for the effects of pre- existing differences among subjects on the DV of interest (e.g. controlling for the effects of an individual’s computer experience in evaluating impact of presence or absence of narrative on enjoyment of computer game play)

More on Tests of Difference of Means: Analysis of Variance with Two Independent Variables (Factors) Two-way ANOVA: two or more levels of two IVs or factors What is the impact of diet type and educational attainment on pounds lost in six months, and how do they interact? This data suggests two significant factors that behave the same way regardless of the level of the other factor (Diet C is always better, post grad always better); don’t interact Diet ADiet BDiet C High School 6810 College Post Graduate Average pounds lost as a function of educational attainment and diet type

When Factors Interact Diet ADiet BDiet C High School College 10 Post Graduate In this data set there seems to be an interaction between diet type and educational attainment, such that Diet C is more effective for people with lower educational attainment, Diet A works better for people with high attainment, and Diet B works equally well regardless of educational attainment. Impact of one factor depends on the level of the second factor Average pounds lost as a function of educational attainment and diet type

Single-factor ANOVA Example (one Independent Variable) Suppose you believed that interviewer status (a manipulated variable in which you systematically varied the dress of the same interviewer across the conditions, high medium, and low would have an effect on interviewee self- disclosure, such that the amount of disclosure of negative personal information would vary across conditions. (The null hypothesis would be that the interviewees all came from the same population of interviewers) Let’s say you conducted your study and got the data on the right, where higher scores equal more self-disclosure Interviewer Status 1. High Status2. Medium Status3. Low Status X 11 3X 12 3X 13 4 X 21 2X 22 4X 23 5 X 31 1X 32 2X 33 3 X 41 2X 42 3X 43 4 Sums Means 234 N 444 Self-disclosure scores for 12 subjects; 4 subjects in each of three interviewer conditions

Some Typical Data for ANOVA The sum over all rows and columns, denoted as ∑∑X ij, = 36 i j (That’s ) The grand mean, denoted X ij, is 3 (That’s divided by 3) The overall N is 12 (That’s 4 subjects in each of three conditions) Interviewer Status High StatusMedium StatusLow Status X 11 3X 12 3X 13 4 X 21 2X 22 4X 23 5 X 31 1X 32 2X 33 3 X 41 2X 42 3X 43 4 Sums Means 234 N 444

Partitioning the Variance for ANOVA: Within and Between Estimates: how to obtain the test statistic, F, for the difference of means To obtain the F statistic, we are going to make two estimates of the common population variance, σ 2 The first is called the “within” estimate, which will be a weighted average of the variances within each of the three samples. This is an unbiased estimate of σ 2 and is an estimate of how much of the variance in self-disclosure scores is attributable to more or less random individual differences The second estimate of the common variance σ 2 is called the “between” (or “among”) estimate and it involves the variance of the sample means about the grand mean. This is an estimate of how much of the variation in self-disclosure scores is attributable to the levels of the factor (interviewer status). The “between” refers to between-levels variation If our factor has a meaningful effect the “between estimate” should be large relative to the “within estimate”; that is, there should be more variation between the levels of interviewer status than within them

Meaning of the F Statistic, the Statistic used in ANOVA The sampling distribution of the F ratio will be used to determine how probable it is that our obtained value of F was due to sampling error The null hypothesis would be that the population means for the three treatment levels would not differ If the null hypothesis is false, and the population means are not equal, then the F ratio will be greater than unity (one). Whether or not the means are significantly different will depend on how large this ratio is There is a sampling distribution for F (see p. 479 in Kendrick) called the “Distribution of the Critical Values of F”; note that there are separate tables for the.05 and.01 confidence levels). (see also the next slide) The columns refer to n 1, the DF of the between groups estimate (K-1, where K is the number of conditions or treatments of the independent variable) and the rows refer to n 2, the DF of the within groups estimate (N (total) – K) For our example n 1, the between DF, would be 2 and n 2, the within DF, would be 9

Critical values of F

Partitioning the Variation in ANOVA The twelve self-disclosure scores we have obtained vary quite a bit from the grand mean of all the scores, which was 3 The total variation is the sum of the squared deviations from the grand (overall) mean. This quantity is also called the “total sum of squares” or the total SS. Its DF is equal to N-1, where N is the total over all the cases. The total variation has two components The within sum of squares: the sum of the squared deviations of the individual scores from their own category (group) mean. We divide this by the df (N-K) to obtain the within estimate. This represents the variability among individuals within the sample The between (among) sum of squares: this is based on the squared deviations of the means of the IV levels from the grand mean, and is a measure of the variability between the conditions. We want this quantity to be big! We divide the betwenn SS by the df K-1 to get the between estimate The within and between estimates are also called the between and within “mean squares”

A Hand Calculation of ANOVA: Obtaining the Between and Within Estimates To get the between estimate, the first thing we calculate is the between sum of squares: We find the difference between each group mean and the grand mean (3), square this deviation, multiply by the number of scores in the group, and sum these quantities Interviewer Status High StatusMedium StatusLow Status X 11 3X 12 3X 13 4 X 21 2X 22 4X 23 5 X 31 1X 32 2X 33 3 X 41 2X 42 3X 43 4 Sums Means 234 N 444

Between Estimate Calculations So we have High Status: 2-3 squared X 4 = 4 Medium Status: 3-3 squared X 4 = 0 Low Status: 4-3 squared X 4 = 4 So the between sum of squares = = 8 And the between estimate is obtained by dividing the between SS by the between degrees of freedom, K-1 Thus the between estimate is 8/2 or 4 Interviewer Status High StatusMedium StatusLow Status X 11 3X 12 3X 13 4 X 21 2X 22 4X 23 5 X 31 1X 32 2X 33 3 X 41 2X 42 3X 43 4 Sums∑ Means 234 N 444 ∑∑=36

Calculating the Total Sum of Squares The next thing we calculate is the total sum of squares. This figure is obtained by summing the squared deviations of each of the individual scores from the grand mean of 3. So the total sum of squares is 3-3 squared plus 2-3 squared plus 1- 3 squared plus 2-3 squared plus 3-3 squared plus 4-3 squared…. plus 4-3 squared = 14 Interviewer Status High StatusMedium StatusLow Status X 11 3X 12 3X 13 4 X 21 2X 22 4X 23 5 X 31 1X 32 2X 33 3 X 41 2X 42 3X 43 4 Sums∑ Means 234 N 444

Calculating the Within Estimate Finally, we calculate the within sum of squares. We obtain that by subtracting the between SS (8) from the total SS (14). So the within SS = 6. And the within estimate is obtained by dividing the within SS by its DF, so the within estimate or within mean square is 6/(N-k) or 6/9 or.667 Recall that for the null hypothesis, that the population means for the three conditions are equal, to be true, the between estimate should equal the within estimate, yet our between estimate is very large in relation to the within estimate. This is good; it means that the variance “explained” by the status manipulation is much greater than what individual differences alone can explain See the table on the next page which shows the estimates for the different sources of variation

Basic Output of an ANOVA Source of Variation Sums of Squares (SS) DFEstimatesF Total14N-1(11)6 Between8K-1(2)4 Within6N – K (9).667 The between and within estimates are obtained by dividing the between and within SS by their respective DFs. The F statistic is obtained by dividing the between estimate by the within estimate (4/.667 = 6) The obtained value of F tells us that the variation between the conditions is much greater than the variation within each condition. We look up the F statistic in the table with 2 DF (conditions minus 1) in the numerator and 9 DF (total N minus number of conditions) in the denominator and we find that we need a F of 4.26 to reject the null hypothesis at p <.05. (see next slide) Called “mean squares”

Looking up the F Value in the Table of Critical Values of F With our obtained F of 6 we can reject the null hypothesis

ANOVA in SPSS Now let’s try that in SPSS. Go here to download the data file disclosure.sav and open it in SPSSdisclosure.sav In Data Editor go to Analyze/Compare Means/One- Way Anova Move the Interviewer Status variable into the Factor window and move the Self-Disclosure variable into the Dependent List window Under Options select Descriptive, then press Continue and then OK Compare the results in your Output Window to the hand calculations and to the next slide

SPSS Output, One-Way ANOVA The results of this analysis suggest that interviewer status has a significant impact on interviewee self-disclosure, F (2,9) = 6, p <.05 ( or p =.022)

Planned Comparisons vs. Post-hoc Comparison of Means Even if we have obtained a significant value of F and the overall difference of means is significant, the F statistic isn’t telling us anything about how the mean scores varied among the levels of the IV. Fortunately, we know that this will be the case in advance, and so we can plan some comparisons between the pairwise group means that we will specify in advance. These are called planned comparisons. Alternatively, we can compare the means of the groups on a pair- wise basis after the fact Doing comparison-of-means tests after the fact, when we have had time to check out the means and see what direction they’re tending (for example, we can look and see that there was more disclosure to the low-status interviewer than to the high-status interviewer), it’s not really the done thing to allow a low confidence level like.10 when we know the direction of the results. We should use a more conservative alpha region in order to reduce the risk of Type I error (rejecting a true null hypothesis)

Post-hoc Tests in SPSS In SPSS data editor, make sure you have the disclosure.sav data file open Go to Analyze/Compare Means/One-Way Anova Move Interviewer Status into the Factor box (this is where the IVs go) Move Self-disclosure into the Dependent List box Under Options, select Descriptive, Homogenity of Variance test, and Means Plot, and click Continue Under Post Hoc, click Sheffé and Tukey and set the confidence interval to.05, then click Continue and OK Compare your output to the next slide

Output for Post-Hoc Comparisons Variances are equal Both Tukey and Sheffé tests show significant differences between high and low status condition but not between medium status and other two conditions. Tukey can only be used with groups of equal size. Sheffé critical value (test statistic that must be exceeded) = k-1 times the critical value of F needed for the one-way anova at a particular alpha level. If variances are unequal by Levene, use the Tamhane’s T2 test for post-hoc comparisons Important!

Writing up Your Result To test the hypothesis that interviewer status would have a significant effect on interviewee self-disclosure, a one-way analysis of variance was performed. Levene’s test for the equality of variances indicated that the variances did not differ significantly across levels of the independent variable (Levene statistic = 000, df = 2, 9, p=1.00). Interviewer status had a significant main effect on interviewee self-disclosure (F (2,9) = 6, p =.022). Sheffe’ post-hoc tests indicated that there were significant differences between mean levels of disclosure for subjects in the high status (M = 2) and low status (M = 4) conditions (p =.022), suggesting an inverse relationship between interviewer status and interviewee disclosure. Subjects disclosed more to the low-status interviewer. Mean Interviewee Self-Disclosure as a Function of Level of Interviewee Status High Status2 b * 3 ab 4 a Medium Status Low Status *Higher scores indicate greater disclosure; means with common subscripts are not significantly different from one another at p =.022

More SPSS ANOVA Using the general social survey data, let’s test the hypothesis that one’s father’s highest earned degree has a significant impact on one’s current socio-economic status Download the socialsurvey.sav file and open it in Data Editorsocialsurvey.sav Go to Analyze/Compare Means/One-Way Anova Move Father’s Highest Degree into the Factor box and move Respondent Socioeconomic Index into the Dependent List box Under Options, select Descriptive and Homogeneity of Variance test and click Continue Under Post Hoc select Sheffé and set the significance level to.05, select Continue and then OK Compare your output to the next slides

What Will Your Results Section Say?

Using the General Linear Model in SPSS Now we are going to redo the same analysis but with a few more bells and whistles. This time, for example, we are going to get measures of the effect size (impact of the IV, father’s highest degree) on the DV, respondent’s SES, and we will also get a power estimate In the Data Editor, make sure your socialsurvey.sav file is open Go to Analyze/General Linear Model/Univariate (in the case of ANOVA, univariate means you only analyze one DV at a time) Put Father’s Highest Degree into the Fixed Factor box and Respondent’s SES into the Dependent Variable box Under Post Hoc, move padeg (shorthand for Father’s Highest Degree) into the Post Hoc Tests for box and under Equal Variances assumed select Sheffé (we can do this because we already know that the variances are not significantly different from our previous analysis) and click Continue Click on Options and move padeg into the Display Means for box Under Display, click on Descriptive Statistics, Estimates of Effect Size, and Observed Power, and set the significance level to.05. Click continue and then OK. Compare your result to the next slide

SPSS GLM Output, Univariate Analysis Note partial eta squared which is the ratio of the between-groups SS to the sum of the between groups SS and the error SS. It describes the amount of variation in the dependent variable explained by the independent variable (Father’s highest degree). In this case the amount of variation accounted for, about 7%, is not very impressive despite a significant result Note that we have all the power required to detect an effect power corrected means that the variance accounted for by the intercept has been removed = Independent variable

SPSS GLM Output, Univariate Analysis, cont’d Note confidence intervals around the mean difference estimates. These intervals should not contain zero (recall that the null hypothesis is of no differences on the dependent variable between levels of the IV) Note also above that some of the confidence levels around the category means themselves contain the mean of the other category. So this sort of data should be studied as well as significance tests