Objectives (BPS chapter 18) Inference about a Population Mean  Conditions for inference  The t distribution  The one-sample t confidence interval 

Slides:



Advertisements
Similar presentations
Chapter 18: Inference about One Population Mean STAT 1450.
Advertisements

Lesson Tests about a Population Mean. Knowledge Objectives Define the one-sample t statistic.
AP STATISTICS LESSON 11 – 1 (DAY 3) Matched Pairs t Procedures.
CHAPTER 9 Testing a Claim
Inference for a population mean BPS chapter 18 © 2006 W. H. Freeman and Company.
Inference for distributions: - for the mean of a population IPS chapter 7.1 © 2006 W.H Freeman and Company.
Conditions with σ Unknown Note: the same as what we saw before.
Chapter 9 Tests of Significance Target Goal: I can perform a significance test to support the alternative hypothesis. I can interpret P values in context.
Objectives (BPS chapter 15)
Introduction Comparing Two Means
© 2013 Pearson Education, Inc. Active Learning Lecture Slides For use with Classroom Response Systems Introductory Statistics: Exploring the World through.
Inference for Distributions - for the Mean of a Population IPS Chapter 7.1 © 2009 W.H Freeman and Company.
Inference for Distributions - for the Mean of a Population
BPS - 5th Ed. Chapter 171 Inference about a Population Mean.
Chapter 11: Inference for Distributions
CHAPTER 19: Two-Sample Problems
Lecture Unit 5 Section 5.7 Testing Hypotheses about Means 1.
7.1 Lecture 10/29.
Tests of significance: The basics BPS chapter 15 © 2006 W.H. Freeman and Company.
Chapter 8 Testing Hypotheses about Means 1. Sweetness in cola soft drinks Cola manufacturers want to test how much the sweetness of cola drinks is affected.
AP Statistics Section 13.1 A. Which of two popular drugs, Lipitor or Pravachol, helps lower bad cholesterol more? 4000 people with heart disease were.
Hypothesis Testing – Examples and Case Studies
Inferences Based on Two Samples
Tests of significance: The basics BPS chapter 15 © 2006 W.H. Freeman and Company.
14. Introduction to inference
BPS - 3rd Ed. Chapter 141 Tests of Significance: The Basics.
Confidence Intervals and Hypothesis Tests for the Difference between Two Population Means µ 1 - µ 2 : Independent Samples Inference for  1  2 1.
Confidence Intervals and Hypothesis Tests for the Difference between Two Population Means µ 1 - µ 2 Inference for  1  Independent Samples.
IPS Chapter 7 © 2012 W.H. Freeman and Company  7.1: Inference for the Mean of a Population  7.2: Comparing Two Means  7.3: Optional Topics in Comparing.
BPS - 5th Ed. Chapter 171 Inference about a Population Mean.
CHAPTER 18: Inference about a Population Mean
1 Happiness comes not from material wealth but less desire.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 10 Comparing Two Populations or Groups 10.2.
Inference for a population mean BPS chapter 18 © 2006 W.H. Freeman and Company.
Essential Statistics Chapter 131 Introduction to Inference.
INTRODUCTION TO INFERENCE BPS - 5th Ed. Chapter 14 1.
CHAPTER 14 Introduction to Inference BPS - 5TH ED.CHAPTER 14 1.
AP Statistics Section 13.1 A. Which of two popular drugs, Lipitor or Pravachol, helps lower bad cholesterol more? 4000 people with heart disease were.
Inference for a population mean BPS chapter 16 © 2006 W.H. Freeman and Company.
Essential Statistics Chapter 141 Thinking about Inference.
Statistics 101 Chapter 10 Section 2. How to run a significance test Step 1: Identify the population of interest and the parameter you want to draw conclusions.
MATH 2400 Ch. 15 Notes.
BPS - 3rd Ed. Chapter 161 Inference about a Population Mean.
Chapter 20 Confidence Intervals and Hypothesis Tests for a Population Mean  ; t distributions t distributions confidence intervals for a population mean.
Section 10.1 Estimating with Confidence AP Statistics February 11 th, 2011.
Essential Statistics Chapter 161 Inference about a Population Mean.
BPS - 3rd Ed. Chapter 141 Tests of significance: the basics.
Week111 The t distribution Suppose that a SRS of size n is drawn from a N(μ, σ) population. Then the one sample t statistic has a t distribution with n.
Inference for Distributions - for the Mean of a Population IPS Chapter 7.1 © 2009 W.H Freeman and Company.
11.1 Inference for the Mean of a Population.  Perform a one sample t-test for the mean of a population  Perform a matched pairs t-test.
Lecture PowerPoint Slides Basic Practice of Statistics 7 th Edition.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 1 – Slide 1 of 26 Chapter 11 Section 1 Inference about Two Means: Dependent Samples.
Comparing Means: Confidence Intervals and Hypotheses Tests for the Difference between Two Population Means µ 1 - µ 2 Chapter 24 Independent Samples Chapter.
+ Unit 6: Comparing Two Populations or Groups Section 10.2 Comparing Two Means.
Learning Objectives After this section, you should be able to: The Practice of Statistics, 5 th Edition1 DESCRIBE the shape, center, and spread of the.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 9 Testing a Claim 9.3 Tests About a Population.
Essential Statistics Chapter 171 Two-Sample Problems.
Inference for a population mean BPS chapter 16 © 2006 W.H. Freeman and Company.
When  is unknown  The sample standard deviation s provides an estimate of the population standard deviation .  Larger samples give more reliable estimates.
16/23/2016Inference about µ1 Chapter 17 Inference about a Population Mean.
10/31/ Comparing Two Means. Does smoking damage the lungs of children exposed to parental smoking? Forced vital capacity (FVC) is the volume (in.
Inference for Distributions Inference for the Mean of a Population PBS Chapter 7.1 © 2009 W.H Freeman and Company.
Statistics for Business and Economics Module 1:Probability Theory and Statistical Inference Spring 2010 Lecture 7: Tests of significance and confidence.
Inference for distributions: - for the mean of a population IPS chapter 7.1 © 2006 W.H Freeman and Company.
Inference for Distributions
Chapter 23 CI HT for m1 - m2: Paired Samples
The Practice of Statistics in the Life Sciences Fourth Edition
Basic Practice of Statistics - 3rd Edition
Objectives 7.1 Inference for the mean of a population
Essential Statistics Inference about a Population Mean
Presentation transcript:

Objectives (BPS chapter 18) Inference about a Population Mean  Conditions for inference  The t distribution  The one-sample t confidence interval  Using technology  Matched pairs t procedures  Robustness of t procedures

Conditions for inference about a mean  We can regard our data as a simple random sample (SRS) from the population. This condition is very important.  Observations from the population have a Normal distribution with mean  and standard deviation . In practice, it is enough that the distribution be symmetric and single-peaked unless the sample is very small. Both  and standard deviation  are unknown.

Sweetening colas Cola manufacturers want to test how much the sweetness of a new cola drink is affected by storage. The sweetness loss due to storage was evaluated by 10 professional tasters (by comparing the sweetness before and after storage): Taster Sweetness loss      5 −0.4   7 −1.3    Obviously, we want to test if storage results in a loss of sweetness, thus H 0 :  = 0 versus H a :  > 0 This looks familiar. However, here we do not know the population parameter .  The population of all cola drinkers is too large.  Since this is a new cola recipe, we have no population data. This situation is very common with real data.

The t distributions We test a null and alternative hypotheses with one sample of size n from a normal population N(µ,σ):  When  is known, the sampling distribution is normal N(  /√n).  When  is estimated from the sample standard deviation s, then the sampling distribution follows a t distribution t( ,s/√n) with degrees of freedom n − 1. The value (s/√n) is the standard error of the mean or SEM.

When n is very large, s is a very good estimate of  and the corresponding t distributions are very close to the normal distribution. The t distributions become wider for smaller sample sizes, reflecting the lack of precision in estimating  from s.

Confidence intervals Reminder: The confidence interval is a range of values with a confidence level C representing the probability that the interval contains the true population parameter. We have a set of data from a population with both  and  unknown. We use to estimate , and s to estimate  using a t distribution (df n − 1). C t*t*−t*−t* m Practical use of t: t*  C is the area under the t (df: n−1) curve between −t* and t*.  We find t* in the line of Table C for df = n−1 and confidence level C.  The margin of error m is:

Red wine, in moderation Drinking red wine in moderation may protect against heart attacks. The polyphenols it contains act on blood cholesterol and thus are a likely cause. To test the hypothesis that moderate red wine consumption increases the average blood level of polyphenols, a group of nine randomly selected healthy men were assigned to drink half a bottle of red wine daily for 2 weeks. Their blood polyphenol levels were assessed before and after the study and the percent change is presented here: Firstly: Are the data approximately normal? Normal? When the sample size is small, histograms can be difficult to interpret. There is a low value, but overall the data can be considered reasonably normal.

Red wine, in moderation (continued) What is the 95% confidence interval for the average percent change in blood polyphenols? Sample average = 5.5; s = 2.517; df = n − 1 = 8 (…) The sampling distribution is a t distribution with n − 1 degrees of freedom. For df = 8 and C = 95%, t* = The margin of error m is : m = t*s/√n = 2.306*2.517/√9 ≈ The 95% confidence interval is therefore 5.5 ± With 95% confidence, the average percent increase in polyphenol blood levels of healthy men drinking half a bottle of red wine daily is between 3.6% and 7.6%. Important: The confidence interval shows how large the increase is, but not if it can have an impact on men’s health.

One-sided (one-tailed) Two-sided (two-tailed) Review: test of significance The P-value is the probability, if H 0 is true, of randomly drawing a sample like the one obtained, or more extreme, in the direction of H a. The P-value is calculated as the corresponding area under the curve, one-tailed or two-tailed depending on H a :

Table C How to: For a one-sided H a, this is the P-value (between 0.01 and 0.02); for a two-sided H a, the P-value is doubled (between 0.02 and 0.04). The calculated value of t is 2.7. We find the two closest t values < t = 2.7 < thus 0.02 > upper tail p > 0.01

Sweetening colas (continued) Is there evidence that storage results in sweetness loss for the new cola recipe at the 0.05 level of significance (  = 5%)? H 0 :  = 0 versus H a :  > 0 (one-sided test)  the critical value t  = t > t  thus the result is significant.  p > 0.01 p <  thus the result is significant. The t-test has a significant p-value. We reject H 0. There is a significant loss of sweetness, on average, following storage. Taster Sweetness loss ___________________________ Average 1.02 Standard deviation 1.196

Red wine, in moderation (continued) Does moderate red wine consumption increase the average blood level of polyphenols in healthy men? Sample average = 5.5; s = 2.517; t = (5.5 − 0)/(2.517/√9) ≈ H 0 :  = 0 versus H a :  > 0 (one-sided test) From Table C, df = 8: t > and therefore p < The P-value is very small (well below 1%), and thus the result is very significant. Moderate red wine consumption significantly increases the average polyphenol blood levels of healthy men. Important: This test does not say how large the increase is, or what the impact on men’s health is. Test statistic would be off the chart to the right

Matched pairs t procedures Sometimes we want to compare treatments or conditions at the individual level. These situations produce two samples that are not independent — they are related to each other. The members of one sample are identical to, or matched (paired) with, the members of the other sample.  Example: Pre-test and post-test studies look at data collected on the same sample elements before and after some experiment is performed.  Example: Twin studies often try to sort out the influence of genetic factors by comparing a variable between sets of twins.  Example: Using people matched for age, sex, and education in social studies allows us to cancel out the effect of these potential lurking variables.

In these cases, we use the paired data to test the difference in the two population means. The variable studied becomes X = x 1 − x 2, and H 0 : µ difference =0; H a : µ difference >0 (or <0, or ≠0) Conceptually, this does not differ from tests on one population.

Sweetening colas (revisited) The sweetness loss due to storage was evaluated by 10 professional tasters (comparing the sweetness before and after storage): Taster Sweetness loss      5 −0.4   7 −1.3    We want to test if storage results in a loss of sweetness, thus H 0 :  = 0 versus H a :  > 0 Although the text did not mention it explicitly, this is a pre-/post-test design, and the variable is the difference in cola sweetness before and after storage. A matched pairs test of significance is indeed just like a one-sample test.

Does lack of caffeine increase depression? Individuals diagnosed as caffeine-dependent are deprived of all caffeine-rich foods and assigned to receive daily pills. At some time, the pills contain caffeine and at another time they contain a placebo. Depression was assessed.  There are two data points for each subject, but we will only look at the difference.  The sample distribution appears appropriate for a t-test. 11 "difference" data points Placebo - Caffeine

Does lack of caffeine increase depression? For each individual in the sample, we have calculated a difference in depression score (placebo minus caffeine). There were 11 “difference” points, thus df = n − 1 = 10. We calculate that = 7.36; s = 6.92 H 0 :  difference = 0 ; H 0 :  difference > 0 For df = 10, p > Caffeine deprivation causes a significant increase in depression. Placebo - Caffeine

Robustness The t procedures are exactly correct when the population is distributed exactly normally. However, most real data are not exactly normal. The t procedures are robust to small deviations from normality. This means that the results will not be affected too much. Factors that do strongly matter are:  Random sampling. The sample must be an SRS from the population.  Outliers and skewness. They strongly influence the mean and therefore the t procedures. However, their impact diminishes as the sample size gets larger because of the Central Limit Theorem. Specifically:  When n < 15, the data must be close to normal and without outliers.  When 15 > n > 40, mild skewness is acceptable, but not outliers.  When n > 40, the t statistic will be valid even with strong skewness.