Comparing Means.

Slides:



Advertisements
Similar presentations
Intro to ANOVA.
Advertisements

Dr. AJIT SAHAI Director – Professor Biometrics JIPMER, Pondicherry
Ch 14 實習(2).
Lesson #24 Multiple Comparisons. When doing ANOVA, suppose we reject H 0 :  1 =  2 =  3 = … =  k Next, we want to know which means differ. This does.
Analysis of Variance (ANOVA) ANOVA methods are widely used for comparing 2 or more population means from populations that are approximately normal in distribution.
Analysis of Variance (ANOVA) Statistics for the Social Sciences Psychology 340 Spring 2010.
Lecture 10 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
Part I – MULTIVARIATE ANALYSIS
ANOVA Determining Which Means Differ in Single Factor Models Determining Which Means Differ in Single Factor Models.
Analysis of Variance: Inferences about 2 or More Means
Lecture 13 Multiple comparisons for one-way ANOVA (Chapter 15.7)
Analysis of Variance Chapter 15 - continued Two-Factor Analysis of Variance - Example 15.3 –Suppose in Example 15.1, two factors are to be examined:
POST HOC COMPARISONS What is the Purpose?
Lecture 9: One Way ANOVA Between Subjects
Two Groups Too Many? Try Analysis of Variance (ANOVA)
Analyses of K-Group Designs : Omnibus F & Pairwise Comparisons ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation.
Statistics for the Social Sciences Psychology 340 Spring 2005 Analysis of Variance (ANOVA)
One-way Between Groups Analysis of Variance
Lecture 12 One-way Analysis of Variance (Chapter 15.2)
K-group ANOVA & Pairwise Comparisons ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation & Correction LSD & HSD procedures.
Comparing Means.
Statistical Methods in Computer Science Hypothesis Testing II: Single-Factor Experiments Ido Dagan.
Introduction to Analysis of Variance (ANOVA)
Linear Contrasts and Multiple Comparisons (Chapter 9)
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
Intermediate Applied Statistics STAT 460
1 Multiple Comparison Procedures Once we reject H 0 :   =   =...  c in favor of H 1 : NOT all  ’s are equal, we don’t yet know the way in which.
When we think only of sincerely helping all others, not ourselves,
Comparing Means. Anova F-test can be used to determine whether the expected responses at the t levels of an experimental factor differ from each other.
1 1 Slide © 2005 Thomson/South-Western Chapter 13, Part A Analysis of Variance and Experimental Design n Introduction to Analysis of Variance n Analysis.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 13 Experimental Design and Analysis of Variance nIntroduction to Experimental Design.
Analysis of Variance Chapter 12 Introduction Analysis of variance compares two or more populations of interval data. Specifically, we are interested.
Analysis of Variance ST 511 Introduction n Analysis of variance compares two or more populations of quantitative data. n Specifically, we are interested.
Chapter 11 HYPOTHESIS TESTING USING THE ONE-WAY ANALYSIS OF VARIANCE.
 The idea of ANOVA  Comparing several means  The problem of multiple comparisons  The ANOVA F test 1.
Economics 173 Business Statistics Lectures 9 & 10 Summer, 2001 Professor J. Petry.
January 31 and February 3,  Some formulae are presented in this lecture to provide the general mathematical background to the topic or to demonstrate.
t(ea) for Two: Test between the Means of Different Groups When you want to know if there is a ‘difference’ between the two groups in the mean Use “t-test”.
Introduction to Biostatistics, Harvard Extension School, Fall, 2005 © Scott Evans, Ph.D.1 Analysis of Variance (ANOVA)
Statistics 11 Confidence Interval Suppose you have a sample from a population You know the sample mean is an unbiased estimate of population mean Question:
ANOVA (Analysis of Variance) by Aziza Munir
Orthogonal Linear Contrasts This is a technique for partitioning ANOVA sum of squares into individual degrees of freedom.
Everyday is a new beginning in life. Every moment is a time for self vigilance.
Orthogonal Linear Contrasts
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
I. Statistical Tests: A Repetive Review A.Why do we use them? Namely: we need to make inferences from incomplete information or uncertainty þBut we want.
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, October 15, 2013 Analysis of Variance (ANOVA)
STA MCP1 Multiple Comparisons: Example Study Objective: Test the effect of six varieties of wheat to a particular race of stem rust. Treatment:
1 Analysis of Variance Chapter 14 2 Introduction Analysis of variance helps compare two or more populations of quantitative data. Specifically, we are.
Orthogonal Linear Contrasts This is a technique for partitioning ANOVA sum of squares into individual degrees of freedom.
Analysis of Variance (One Factor). ANOVA Analysis of Variance Tests whether differences exist among population means categorized by only one factor or.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 13: One-way ANOVA Marshall University Genomics Core.
- We have samples for each of two conditions. We provide an answer for “Are the two sample means significantly different from each other, or could both.
Statistics for the Social Sciences Psychology 340 Spring 2009 Analysis of Variance (ANOVA)
Introduction to ANOVA Research Designs for ANOVAs Type I Error and Multiple Hypothesis Tests The Logic of ANOVA ANOVA vocabulary, notation, and formulas.
Chapters Way Analysis of Variance - Completely Randomized Design.
Multiple comparisons 郭士逢 輔大生科系 2008 Advanced Biostatistics.
ANOVA: Why analyzing variance to compare means?.
1/54 Statistics Analysis of Variance. 2/54 Statistics in practice Introduction to Analysis of Variance Analysis of Variance: Testing for the Equality.
ANalysis Of VAriance (ANOVA) Used for continuous outcomes with a nominal exposure with three or more categories (groups) Result of test is F statistic.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Posthoc Comparisons finding the differences. Statistical Significance What does a statistically significant F statistic, in a Oneway ANOVA, tell us? What.
Statistics Analysis of Variance.
Multiple Comparisons Q560: Experimental Methods in Cognitive Science Lecture 10.
Post Hoc Tests on One-Way ANOVA
Linear Contrasts and Multiple Comparisons (§ 8.6)
Multiple Comparisons: Example
I. Statistical Tests: Why do we use them? What do they involve?
Comparing Means.
1-Way Analysis of Variance - Completely Randomized Design
Presentation transcript:

Comparing Means

Anova F-test can be used to determine whether the expected responses at the t levels of an experimental factor differ from each other When the null hypothesis is rejected, it may be desirable to find which mean(s) is (are) different, and at what ranking order. In practice, it is actually not primary interest to test the null hypothesis, instead the investigators want to make specific comparisons of the means and to estimate pooled error

Means comparison Three categories: 1. Pair-wise comparisons 2. Comparison specified prior to performing the experiment (Planned comparison) 3. Comparison specified after observing the outcome of the experiment (Post-Hoc Comparison) Statistical inference procedures of pair-wise comparisons: Fisher’s least significant difference (LSD) method Duncan’s Multiple Range Test (DMRT) Student Newman Keul Test (SNK) Tukey’s HSD (“Honestly Significantly Different”) Procedure

Suppose there are t means Pair Comparison Suppose there are t means An F-test has revealed that there are significant differences amongst the t means Performing an analysis to determine precisely where the differences exist.

Pair Comparison Two means are considered different if the difference between the corresponding sample means is larger than a critical number. Then, the larger sample mean is believed to be associated with a larger population mean. Conditions common to all the methods: The ANOVA model is the one way analysis of variance The conditions required to perform the ANOVA are satisfied. The experiment is fixed-effect.

Comparing Pair-comparison methods With the exception of the F-LSD test, there is no good theoretical argument that favors one pair-comparison method over the others. Professional statisticians often disagree on which method is appropriate. In terms of Power and the probability of making a Type I error, the tests discussed can be ordered as follows: MORE Power HIGHER P[Type I Error] Tukey HSD Test Student-Newman-Keuls Test Duncan Multiple Range Test Fisher LSD Test Pairwise comparisons are traditionally considered as “post hoc” and not “a priori”, if one needs to categorize all comparisons into one of the two groups

Fisher Least Significant Different (LSD) Method This method builds on the equal variances t-test of the difference between two means. The test statistic is improved by using MSE rather than sp2. It is concluded that mi and mj differ (at a% significance level if |mi - mj| > LSD, where

Experimentwise Type I error rate (aE) (the effective Type I error) The Fisher’s method may result in an increased probability of committing a type I error. The experimentwise Type I error rate is the probability of committing at least one Type I error at significance level of a. It is calculated by aE = 1-(1 – a)C where C is the number of pairwise comparisons (all: C = t(t-1)/2 The Bonferroni adjustment determines the required Type I error probability per pairwise comparison (a) , to secure a pre-determined overall aE.

Bonferroni Adjustment The procedure: Compute the number of pairwise comparisons (C) [all: C=t(t-1)/2], where t is the number of populations. Set a = aE/C, where aE is the true probability of making at least one Type I error (called experimentwise Type I error). It is concluded that mi and mj differ at a/C% significance level if

Duncan’s Multiple Range Test The Duncan Multiple Range test uses different Significant Difference values for means next to each other along the real number line, and those with 1, 2, … , a means in between the two means being compared. The Significant Difference or the range value: where ra,p,n is the Duncan’s Significant Range Value with parameters p (= range-value) and n (= MSE degree-of-freedom), and experiment-wise alpha level a (= ajoint).

Duncan’s Multiple Range Test MSE is the mean square error from the ANOVA table and n is the number of observations used to calculate the means being compared. The range-value is: 2 if the two means being compared are adjacent 3 if one mean separates the two means being compared 4 if two means separate the two means being compared …

Significant Ranges for Duncan’s Multiple Range Test

Student-Newman-Keuls Test Similar to the Duncan Multiple Range test, the Student-Newman-Keuls Test uses different Significant Difference values for means next to each other, and those with 1, 2, … , a means in between the two means being compared. The Significant Difference or the range value for this test is where qa,a,n is the Studentized Range Statistic with parameters p (= range-value) and n (= MSE degree-of-freedom), and experiment-wise alpha level a (= ajoint).

Student-Newman-Keuls Test MSE is the mean square error from the ANOVA table and n is the number of observations used to calculate the means being compared. The range-value is: 2 if the two means being compared are adjacent 3 if one mean separates the two means being compared 4 if two means separate the two means being compared …

Studentized Range Statistic

Tukey HSD Procedure Tukey HSD Procedure: The Tukey Procedure is used to compare all individual pairs of means after a significant ANOVA test has been conducted. HSD = “honestly significantly different”

Tukey HSD Procedure The test procedure: Assumes equal number of observation per populations. Find a critical number w as follows: dft = treatment degrees of freedom n =degrees of freedom = dfe ng = number of observations per population a = significance level qa(dft,n) = a critical value obtained from the studentized range table

Studentized Range Statistic

Tukey Multiple Comparisons Select a pair of means. Calculate the difference between the larger and the smaller mean. If there is sufficient evidence to conclude that mmax > mmin . Repeat this procedure for each pair of samples. Rank the means if possible. If the sample sizes are not extremely different, it can be used the above procedure with ng calculated as the harmonic mean of the sample sizes.

Planned Comparisons or Contrasts In some cases, an experimenter may know ahead of time that it is of interest to compare two different means, or groups of means. An effective way to do this is to use contrasts or planned comparisons. These represent specific hypotheses in terms of the treatment means such as:

Planned Comparisons or Contrasts Each contrast can be specified as: and it is required: A sum-of-squares can be calculated for a contrast as

Planned Comparisons or Contrasts Each contrast has 1 degree-of-freedom, and a contrast can be tested by comparing it to the MSE for the ANOVA: If more than 1 contrast is tested, it is important that the contrasts all be orthogonal, that is Note that It can be tested at most t-1 orthogonal contrasts.

Contrast Examples Design Characteristics 1 3 colors, with cartoons 2 3 colors, without cartoons 3 5 colors, with cartoons 4 5 colors, without cartoons Comparison of… The mean sales for the two 3-color designs m1 = m2 The mean sales for the 3-color and 5-color designs (m1+m2)/2 = (m3+m4)/2 The mean sales for designs with and without cartoons (m1+m3)/2 = (m2+m4)/2 The mean sales for design 1 with average sales for all four designs m1 = (m1+m2+m3+m4)/4

Post Hoc comparisons 1. F test for contrast →Scheffe test 2. To test all linear combinations at once. 3. Very conservative; not to be used for pairwise comparisons 4. A Priori comparisons

Scheffe's Critical Differences (for Linear contrasts) A linear contrast is declared significant if it exceeds this amount. = the tabled value for F distribution (p -1 = df for comparing p means, n = df for Error)

Scheffe's Critical Differences (for comparing two means) Two means are declared significant if they differ by more than this amount.