T-tests continued.

Slides:



Advertisements
Similar presentations
Introduction to Hypothesis Testing
Advertisements

Overview of Lecture Partitioning Evaluating the Null Hypothesis ANOVA
Lecture 2 ANALYSIS OF VARIANCE: AN INTRODUCTION
Chapter 10: The t Test For Two Independent Samples
Comparing Two Population Parameters
Chapter 13 Comparing Two Populations: Independent Samples.
Tests of Significance for Regression & Correlation b* will equal the population parameter of the slope rather thanbecause beta has another meaning with.
1 COMM 301: Empirical Research in Communication Lecture 15 – Hypothesis Testing Kwan M Lee.
Statistics Review – Part II Topics: – Hypothesis Testing – Paired Tests – Tests of variability 1.
Chapter 10 Estimation and Hypothesis Testing II: Independent and Paired Sample T-Test.
PTP 560 Research Methods Week 9 Thomas Ruediger, PT.
Chapter 14 Comparing two groups Dr Richard Bußmann.
PSY 307 – Statistics for the Behavioral Sciences
Inferences About Means of Two Independent Samples Chapter 11 Homework: 1, 2, 3, 4, 6, 7.
The t-test:. Answers the question: is the difference between the two conditions in my experiment "real" or due to chance? Two versions: (a) “Dependent-means.
Lecture 8 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
BCOR 1020 Business Statistics
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 9: Hypothesis Tests for Means: One Sample.
Overview of Lecture Parametric Analysis is used for
Don’t spam class lists!!!. Farshad has prepared a suggested format for you final project. It will be on the web
Tuesday, October 22 Interval estimation. Independent samples t-test for the difference between two means. Matched samples t-test.
Inferences About Means of Two Independent Samples Chapter 11 Homework: 1, 2, 4, 6, 7.
Business Statistics: A Decision-Making Approach, 7e © 2008 Prentice-Hall, Inc. Chap 10-1 Business Statistics: A Decision-Making Approach 7 th Edition Chapter.
One-way Between Groups Analysis of Variance
Chapter 11: Inference for Distributions
Major Points Formal Tests of Mean Differences Review of Concepts: Means, Standard Deviations, Standard Errors, Type I errors New Concepts: One and Two.
Getting Started with Hypothesis Testing The Single Sample.
Bootstrapping applied to t-tests
Nonparametric or Distribution-free Tests
Hypothesis Testing and T-Tests. Hypothesis Tests Related to Differences Copyright © 2009 Pearson Education, Inc. Chapter Tests of Differences One.
AP Statistics Section 13.1 A. Which of two popular drugs, Lipitor or Pravachol, helps lower bad cholesterol more? 4000 people with heart disease were.
Analysis of Variance. ANOVA Probably the most popular analysis in psychology Why? Ease of implementation Allows for analysis of several groups at once.
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
Comparing Means From Two Sets of Data
T-test Mechanics. Z-score If we know the population mean and standard deviation, for any value of X we can compute a z-score Z-score tells us how far.
Education 793 Class Notes T-tests 29 October 2003.
The paired sample experiment The paired t test. Frequently one is interested in comparing the effects of two treatments (drugs, etc…) on a response variable.
Dependent Samples: Hypothesis Test For Hypothesis tests for dependent samples, we 1.list the pairs of data in 2 columns (or rows), 2.take the difference.
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
Comparing Two Proportions
One-sample In the previous cases we had one sample and were comparing its mean to a hypothesized population mean However in many situations we will use.
Psychology 301 Chapters & Differences Between Two Means Introduction to Analysis of Variance Multiple Comparisons.
Chap 9-1 Two-Sample Tests. Chap 9-2 Two Sample Tests Population Means, Independent Samples Means, Related Samples Population Variances Group 1 vs. independent.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
Essential Question:  How do scientists use statistical analyses to draw meaningful conclusions from experimental results?
Copyright © 2010 Pearson Education, Inc. Chapter 22 Comparing Two Proportions.
© Copyright McGraw-Hill 2000
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
T tests comparing two means t tests comparing two means.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 1 – Slide 1 of 26 Chapter 11 Section 1 Inference about Two Means: Dependent Samples.
Modern Approaches The Bootstrap with Inferential Example.
HYPOTHESIS TESTING FOR DIFFERENCES BETWEEN MEANS AND BETWEEN PROPORTIONS.
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 4 Investigating the Difference in Scores.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Chapter 10: The t Test For Two Independent Samples.
When the means of two groups are to be compared (where each group consists of subjects that are not related) then the excel two-sample t-test procedure.
Dependent-Samples t-Test
Estimation & Hypothesis Testing for Two Population Parameters
Math 4030 – 10a Tests for Population Mean(s)
Psychology 202a Advanced Psychological Statistics
12 Inferential Analysis.
Reasoning in Psychology Using Statistics
Comparing Populations
Hypothesis Testing.
Reasoning in Psychology Using Statistics
12 Inferential Analysis.
What are their purposes? What kinds?
Reasoning in Psychology Using Statistics
Reasoning in Psychology Using Statistics
Introduction to the t Test
Presentation transcript:

T-tests continued

Outline Review the One-sample case Independent samples Paired samples Basics Unequal N Example Issues Paired samples Examples Comparing formulas Conclusion

Two samples Previously compared sample mean to a known population mean Now want to compare two samples Null hypothesis: the mean of the population of scores from which one set of data is drawn is equal to the mean of the population of the second data set H0: m1=m2 or m1 - m2 = 0

Independent samples Consider the original case Now want to consider not just 1 mean but the difference between 2 means The ‘nil’ hypothesis, as before, states there will be no difference H0: m1 - m2 = 0

Which leads to... Now statistic of interest is not a single mean, but the difference between means: Mean of the ‘sampling distribution of the differences between means’ is:

Variability Standard error of the difference between means Since there are two independent variables, variance of the difference between means equals sum of their variances

Same problem, same solution Usually we do not know population variance (standard deviation) Again use sample to estimate it Result is distributed as t (rather than z)

Formula All of which leads to:

But... If we are dealing with a ‘nil’ null hypothesis: So the formula reduces to: Across the 2 samples we have (n1-1) and (n2-1) degrees of freedom1 df = (n1-1) + (n2-1) = n1 + n2 - 2

Unequal sample sizes Assumption: independent samples t test requires samples come from populations with equal variances Two estimates of variance (one from each sample) Generate an overall estimate that reflects the fact that bigger samples offer better estimates Oftentimes the sample sizes will be unequal

Weighted average Which gives us: Final result is:

Example New drug MemoPlus is supposed to aid memory. 5 people take the drug and we compare them to a control group of 10 people who take a placebo. Look at their free recall of 10 items on a word list. Control: mean = 4.5, SD = 2.0 Drug: mean = 6.0, SD = 3.0

Start with variance Using Get: [note pooled variance nearer to group’s with larger n]

Calculate t Enter the values: Critical value approach Specific p

Conclusion? Our result does not give us a strong case for rejecting the null. MemoPlus, like Ginko Biloba1, does nothing special for memory. 1. Solomon et al. (2002) JAMA.

More issues with the t-test In the two-sample case we have an additional assumption (along with normal distribution of sample scores and independent observations) We assume that there are equal variances in the groups Recall our homoscedasticity discussion, it is the exact same assumption1 Often this assumption is untenable, and the results, like other violations result in using calculated probabilities that are inaccurate Can use a correction, e.g. Welch’s t 1. Note also that our normality assumption still technically deals with the residuals in terms of the general linear model, more on that later.

More issues with the t-test It is one thing to say that they are unequal, but what might that mean? Consider a control and treatment group, treatment group variance is significantly greater While we can do a correction, the unequal variances may suggest that those in the treatment group vary widely in how they respond to the treatment (e.g. half benefit, ¼ unchanged, ¼ got worse) Another reason for heterogeneity of variance may be related to an unreliable measure being used No version of the t-test takes either into consideration Other techniques, assuming enough information has been gathered, may be more appropriate (e.g. a hierarchical approach), and more reliable measures may be attainable 1. Note that, if those in the treatment are truly more variable, a more reliable measure would actually detect this more so (i.e. more reliability would lead to a less powerful test). We will consider this more later.

T-test for dependent samples Paired sample t-test Comparing two distributions that are likely to be correlated in some way Gist: some sort of meaningful relationship between our two samples. When are two distributions likely to be correlated? Natural Pairs: Comparing the scores of two subjects who are related naturally E.g. twins. Matched Pairs: Pairs two subjects on some characteristic that is likely to be important to the variable that you are studying. Members of the pair are randomly assigned to the two conditions. E.g. married couples Repeated Measures: Same person at two different times.

Difference scores While we technically have two samples of data, the test regards the single sample of difference scores from one pair to another Transform the paired scores into a single set of scores Find the difference for each case Now are we have a single set of scores and a known population mean, a familiar situation Use the one sample t test concept If the null hypothesis is true then on average there should be no difference Thus mean difference is 0 ( mD = 0 )

Paired t test Where = Mean of the difference scores sD = Standard deviation of the difference scores n = Number of difference scores (# pairs) df = n-1

Example We want to know if dog owners affection for their dog changes after they take a course in dog psychology We measure their responses on the the Dog Affection Inventory both before and after the course

Data t (9)= 3.71, p = .005 2 Conclusion ? Note that which sample is taken from which is arbitrary, just as in the independent samples t-test. Common practice is to report positive t statistics, with tabled or in text reports very clear on which mean is greater and which lesser. t.test(before,after, paired=T)

ts and Zs

The good and the bad regarding t-tests If assumptions are met, t-test is fine When assumptions aren’t met, t-test may still be robust with regard to type I error in some situations With equal n and normal populations HoV violations won’t increase type I much With non-normal distributions with equal variances, type I error rate is maintained also The bad Even small departures from the assumptions result in power taking a noticeable hit (type II error is not maintained) t-statistic, CIs will be biased