Comparing Means. Anova F-test can be used to determine whether the expected responses at the t levels of an experimental factor differ from each other.

Slides:



Advertisements
Similar presentations
Intro to ANOVA.
Advertisements

Week 2 – PART III POST-HOC TESTS. POST HOC TESTS When we get a significant F test result in an ANOVA test for a main effect of a factor with more than.
Dr. AJIT SAHAI Director – Professor Biometrics JIPMER, Pondicherry
Lesson #24 Multiple Comparisons. When doing ANOVA, suppose we reject H 0 :  1 =  2 =  3 = … =  k Next, we want to know which means differ. This does.
Analysis of Variance (ANOVA) ANOVA methods are widely used for comparing 2 or more population means from populations that are approximately normal in distribution.
Hypothesis Testing Steps in Hypothesis Testing:
Inference for Regression
Analysis of variance (ANOVA)-the General Linear Model (GLM)
Model Adequacy Checking in the ANOVA Text reference, Section 3-4, pg
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
Analysis of Variance (ANOVA) Statistics for the Social Sciences Psychology 340 Spring 2010.
Comparing Means.
DOCTORAL SEMINAR, SPRING SEMESTER 2007 Experimental Design & Analysis Contrasts, Trend Analysis and Effects Sizes February 6, 2007.
Lecture 13 Multiple comparisons for one-way ANOVA (Chapter 15.7)
Analysis of Variance Chapter 15 - continued Two-Factor Analysis of Variance - Example 15.3 –Suppose in Example 15.1, two factors are to be examined:
Lecture 9: One Way ANOVA Between Subjects
Two Groups Too Many? Try Analysis of Variance (ANOVA)
Statistics for the Social Sciences Psychology 340 Spring 2005 Analysis of Variance (ANOVA)
One-way Between Groups Analysis of Variance
Lecture 12 One-way Analysis of Variance (Chapter 15.2)
Comparing Means.
Today Concepts underlying inferential statistics
Statistical Methods in Computer Science Hypothesis Testing II: Single-Factor Experiments Ido Dagan.
Chapter 14 Inferential Data Analysis
Linear Contrasts and Multiple Comparisons (Chapter 9)
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
1 Multiple Comparison Procedures Once we reject H 0 :   =   =...  c in favor of H 1 : NOT all  ’s are equal, we don’t yet know the way in which.
When we think only of sincerely helping all others, not ourselves,
1 1 Slide © 2005 Thomson/South-Western Chapter 13, Part A Analysis of Variance and Experimental Design n Introduction to Analysis of Variance n Analysis.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 13 Experimental Design and Analysis of Variance nIntroduction to Experimental Design.
Lecture 8 Analysis of Variance and Covariance Effect of Coupons, In-Store Promotion and Affluence of the Clientele on Sales.
Chapter 11 HYPOTHESIS TESTING USING THE ONE-WAY ANALYSIS OF VARIANCE.
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
 The idea of ANOVA  Comparing several means  The problem of multiple comparisons  The ANOVA F test 1.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
January 31 and February 3,  Some formulae are presented in this lecture to provide the general mathematical background to the topic or to demonstrate.
t(ea) for Two: Test between the Means of Different Groups When you want to know if there is a ‘difference’ between the two groups in the mean Use “t-test”.
Statistics 11 Confidence Interval Suppose you have a sample from a population You know the sample mean is an unbiased estimate of population mean Question:
ANOVA (Analysis of Variance) by Aziza Munir
Orthogonal Linear Contrasts This is a technique for partitioning ANOVA sum of squares into individual degrees of freedom.
Everyday is a new beginning in life. Every moment is a time for self vigilance.
Orthogonal Linear Contrasts
I. Statistical Tests: A Repetive Review A.Why do we use them? Namely: we need to make inferences from incomplete information or uncertainty þBut we want.
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, October 15, 2013 Analysis of Variance (ANOVA)
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
STA MCP1 Multiple Comparisons: Example Study Objective: Test the effect of six varieties of wheat to a particular race of stem rust. Treatment:
1 Analysis of Variance Chapter 14 2 Introduction Analysis of variance helps compare two or more populations of quantitative data. Specifically, we are.
Orthogonal Linear Contrasts This is a technique for partitioning ANOVA sum of squares into individual degrees of freedom.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Analysis of Variance and Covariance Effect of Coupons, In-Store Promotion and Affluence of the Clientele on Sales.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 13: One-way ANOVA Marshall University Genomics Core.
ANOVA P OST ANOVA TEST 541 PHL By… Asma Al-Oneazi Supervised by… Dr. Amal Fatani King Saud University Pharmacy College Pharmacology Department.
Statistics for the Social Sciences Psychology 340 Spring 2009 Analysis of Variance (ANOVA)
Research Methods and Data Analysis in Psychology Spring 2015 Kyle Stephenson.
Introduction to ANOVA Research Designs for ANOVAs Type I Error and Multiple Hypothesis Tests The Logic of ANOVA ANOVA vocabulary, notation, and formulas.
Chapters Way Analysis of Variance - Completely Randomized Design.
Multiple comparisons 郭士逢 輔大生科系 2008 Advanced Biostatistics.
1/54 Statistics Analysis of Variance. 2/54 Statistics in practice Introduction to Analysis of Variance Analysis of Variance: Testing for the Equality.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Posthoc Comparisons finding the differences. Statistical Significance What does a statistically significant F statistic, in a Oneway ANOVA, tell us? What.
MEASURES OF CENTRAL TENDENCY Central tendency means average performance, while dispersion of a data is how it spreads from a central tendency. He measures.
Week 2 – PART III POST-HOC TESTS.
Factorial Experiments
ANOVA Econ201 HSTS212.
Post Hoc Tests on One-Way ANOVA
Statistics for Business and Economics (13e)
Differences Among Group Means: One-Way Analysis of Variance
I. Statistical Tests: Why do we use them? What do they involve?
Comparing Means.
Presentation transcript:

Comparing Means

Anova F-test can be used to determine whether the expected responses at the t levels of an experimental factor differ from each other F-test can be used to determine whether the expected responses at the t levels of an experimental factor differ from each other When the null hypothesis is rejected, it may be desirable to find which mean(s) is (are) different, and at what ranking order. When the null hypothesis is rejected, it may be desirable to find which mean(s) is (are) different, and at what ranking order. In practice, it is actually not primary interest to test the null hypothesis, instead the investigators want to make specific comparisons of the means and to estimate pooled error In practice, it is actually not primary interest to test the null hypothesis, instead the investigators want to make specific comparisons of the means and to estimate pooled error

Means comparison Three categories: Three categories: 1. Pair-wise comparisons (Post-Hoc Comparison) 2. Comparison specified prior to performing the experiment (Planned comparison) (Planned comparison) 3. Comparison specified after observing the outcome of the experiment (Un-planned comparison) experiment (Un-planned comparison) Statistical inference procedures of pair-wise comparisons: Statistical inference procedures of pair-wise comparisons: Fisher’s least significant difference (LSD) method Fisher’s least significant difference (LSD) method Duncan’s Multiple Range Test (DMRT) Duncan’s Multiple Range Test (DMRT) Student Newman Keul Test (SNK) Student Newman Keul Test (SNK) Tukey’s Tukey’s HSD (“Honestly Significantly Different”) Procedure

Suppose are t means Suppose there are t means An F-test has revealed that there are significant differences amongst the t means Performing an analysis to determine precisely where the differences exist. Pair Comparison

Two means are considered different if the difference between the corresponding sample means is larger than a critical number. Then, the larger sample mean is believed to be associated with a larger population mean. Two means are considered different if the difference between the corresponding sample means is larger than a critical number. Then, the larger sample mean is believed to be associated with a larger population mean. Conditions common to all the methods: Conditions common to all the methods: The ANOVA model is the one way analysis of variance The ANOVA model is the one way analysis of variance The conditions required to perform the ANOVA are satisfied. The conditions required to perform the ANOVA are satisfied. The experiment is fixed-effect. The experiment is fixed-effect. Pair Comparison

Comparing Pair-comparison methods With the exception of the F-LSD test, there is no good theoretical argument that favors one pair-comparison method over the others. Professional statisticians often disagree on which method is appropriate. With the exception of the F-LSD test, there is no good theoretical argument that favors one pair-comparison method over the others. Professional statisticians often disagree on which method is appropriate. In terms of Power and the probability of making a Type I error, the tests discussed can be ordered as follows: In terms of Power and the probability of making a Type I error, the tests discussed can be ordered as follows: Tukey HSD Test Student-Newman-Keuls Test Duncan Multiple Range Test Fisher LSD Test MORE Power HIGHER P[Type I Error] Pairwise comparisons are traditionally considered as “post hoc” and not “a priori”, if one needs to categorize all comparisons into one of the two groups

Fisher Least Significant Different (LSD) Method This method builds on the equal variances t-test of the difference between two means. This method builds on the equal variances t-test of the difference between two means. The test statistic is improved by using MSE rather than s p 2. The test statistic is improved by using MSE rather than s p 2. It is concluded that  i and  j differ (at  % significance level if |  i -  j | > LSD, where It is concluded that  i and  j differ (at  % significance level if |  i -  j | > LSD, where

Critical t for a test about equality = t  (2),

Example: Cassava yields (ton/ha) Source of variation Degrees of Freedom Sum of Square Mena Square F calculated Treatment313645,33334 Block34013,33310 Error9121,33 Total1518 F-table: 3,86

Duncan’s Multiple Range Test The Duncan Multiple Range test uses different Significant Difference values for means next to each other along the real number line, and those with 1, 2, …, a means in between the two means being compared. The Duncan Multiple Range test uses different Significant Difference values for means next to each other along the real number line, and those with 1, 2, …, a means in between the two means being compared. The Significant Difference or the range value: The Significant Difference or the range value: where r ,p, is the Duncan’s Significant Range Value with parameters p (= range-value) and (= MSE degree-of-freedom), and experiment-wise alpha level  (=  joint ).

Duncan’s Multiple Range Test MSE is the mean square error from the ANOVA table and n is the number of observations used to calculate the means being compared. The range-value is: 2 if the two means being compared are adjacent 3 if one mean separates the two means being compared 4 if two means separate the two means being compared …

Significant Ranges for Duncan’s Multiple Range Test

Student-Newman-Keuls Test Similar to the Duncan Multiple Range test, the Student- Newman-Keuls Test uses different Significant Difference values for means next to each other, and those with 1, 2, …, a means in between the two means being compared. Similar to the Duncan Multiple Range test, the Student- Newman-Keuls Test uses different Significant Difference values for means next to each other, and those with 1, 2, …, a means in between the two means being compared. The Significant Difference or the range value for this test is The Significant Difference or the range value for this test is where q ,a, is the Studentized Range Statistic with parameters p (= range-value) and (= MSE degree-of-freedom), and experiment-wise alpha level  (=  joint ). where q ,a, is the Studentized Range Statistic with parameters p (= range-value) and (= MSE degree-of-freedom), and experiment-wise alpha level  (=  joint ).

Student-Newman-Keuls Test MSE is the mean square error from the ANOVA table and n is the number of observations used to calculate the means being compared. The range-value is: 2 if the two means being compared are adjacent 3 if one mean separates the two means being compared 4 if two means separate the two means being compared …

Studentized Range Statistic

The test procedure: Assumes equal number of observation per populations. Assumes equal number of observation per populations. Find a critical number  as follows: Find a critical number  as follows: dft = treatment degrees of freedom =degrees of freedom = dfe n g = number of observations per population  = significance level q  (dft, ) = a critical value obtained from the studentized range table Tukey HSD Procedure

Studentized Range Statistic

There are many multiple (post hoc) comparison procedures Considerable controversy: “I have not included the multiple comparison methods of Duncan because I have been unable to understand their justification” Scheffe

Planned Comparisons or Contrasts In some cases, an experimenter may know ahead of time that it is of interest to compare two different means, or groups of means. In some cases, an experimenter may know ahead of time that it is of interest to compare two different means, or groups of means. An effective way to do this is to use contrasts or planned comparisons. These represent specific hypotheses in terms of the treatment means such as: An effective way to do this is to use contrasts or planned comparisons. These represent specific hypotheses in terms of the treatment means such as:

Planned Comparisons or Contrasts Each contrast can be specified as: and it is required: Each contrast can be specified as: and it is required: A sum-of-squares can be calculated for a contrast as A sum-of-squares can be calculated for a contrast as

Planned Comparisons Each contrast has 1 degree-of-freedom, and a contrast can be tested by comparing it to the MSE for the ANOVA: Each contrast has 1 degree-of-freedom, and a contrast can be tested by comparing it to the MSE for the ANOVA:

Un-planned Comparisons or Contrasts If more than 1 contrast is tested, it is important that the contrasts all be orthogonal, that is If more than 1 contrast is tested, it is important that the contrasts all be orthogonal, that is Note that It can be tested at most t-1 orthogonal contrasts.

Contrast orthogonal examples TreatmentYields (ton/ha) Adira-419 GH-625 GH-718 Local18 The mean effect of local and high yielding varieties The mean effect of high yielding and promising lines

Orthogonal Polynomials Restrictive assumptions: Require quantitative factors Equal spacing of factor levels (d) Equal numbers of observations at each cell (r j ) Usually, only the linear and quadratic contrasts are of interest Special sets of coefficients that test for bends but manage to remain uncorrelated with one another. Sometimes orthogonal polynomials can be used to analyze experimental data to test for curves.

Polynomial Examples Treatment (Urea dosage) kg/ha Yields (ton/ha)

26 Orthogonal Polynomial The linear regression model y = X  +  is a general model for fitting any relationship that is linear in the unknown parameter . Polynomial regression model:

Polynomial Models in One Variable A second-order model (quadratic model): A second-order model (quadratic model):

A second-order model (quadratic model)

Polynomial models are useful in situations where the analyst knows that curvilinear effects are present in the true response function. Polynomial models are also useful as approximating functions to unknown and possible very complex nonlinear relationship. Polynomial model is the Taylor series expansion of the unknown function. Polynomial Models

30 Theoretical background Theoretical background Scatter diagram Scatter diagram Orthogonal polynomial test Orthogonal polynomial test Choosing order of the model

Theoretical background Can be searched from previous research or literature Can be searched from previous research or literature Examples: Examples: The relationship between dosages of nitrogen application and yield (The law of diminishing return) The relationship between pesticide application and pest mortality (Linear model/probit analysis) The relationship between population density and yield (Exponential model/Cob-Douglass Model)

Scatter Diagram

Orthogonal Linear Contrasts for Polynomial Regression