Presentation is loading. Please wait.

Presentation is loading. Please wait.

The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 9 Testing a Claim 9.1 Significance Tests:

Similar presentations


Presentation on theme: "The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 9 Testing a Claim 9.1 Significance Tests:"— Presentation transcript:

1 The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 9 Testing a Claim 9.1 Significance Tests: The Basics

2 Learning Objectives After this section, you should be able to: The Practice of Statistics, 5 th Edition2 STATE the null and alternative hypotheses for a significance test about a population parameter. INTERPRET a P-value in context. DETERMINE whether the results of a study are statistically significant and MAKE an appropriate conclusion using a significance level. INTERPRET a Type I and a Type II error in context and GIVE a consequence of each. Significance Tests: The Basics

3 The Practice of Statistics, 5 th Edition3 Introduction Confidence intervals are one of the two most common types of statistical inference. Use a confidence interval when your goal is to estimate a population parameter. The second common type of inference, called significance tests, has a different goal: to assess the evidence provided by data about some claim concerning a population. A significance test is a formal procedure for comparing observed data with a claim (also called a hypothesis) whose truth we want to assess. The claim is a statement about a parameter, like the population proportion p or the population mean µ. We express the results of a significance test in terms of a probability that measures how well the data and the claim agree.

4 The Practice of Statistics, 5 th Edition4 Activity: I’m a Great Free-Throw Shooter! A basketball player claims to make 80% of the free throws that he attempts. We think he might be exaggerating. To test this claim, we’ll ask him to shoot some free throws—virtually— using The Reasoning of a Statistical Test applet at the book’s Web site. 1. Launch the applet. 2. Set the applet to take 25 shots. Click “Shoot.” Record how many of the 25 shots the player makes. 3. Click “Shoot” again for 25 more shots. Repeat until you are convinced either that the player makes less than 80% of his shots or that the player’s claim is true. 4. Click “Show true probability.” Were you correct?

5 The Practice of Statistics, 5 th Edition5 Stating Hypotheses A significance test starts with a careful statement of the claims we want to compare. The claim we weigh evidence against in a statistical test is called the null hypothesis (H 0 ). Often the null hypothesis is a statement of “no difference.” The claim about the population that we are trying to find evidence for is the alternative hypothesis (H a ). The claim we weigh evidence against in a statistical test is called the null hypothesis (H 0 ). Often the null hypothesis is a statement of “no difference.” The claim about the population that we are trying to find evidence for is the alternative hypothesis (H a ). In the free-throw shooter example, our hypotheses are H 0 : p = 0.80 H a : p < 0.80 where p is the long-run proportion of made free throws.

6 The Practice of Statistics, 5 th Edition6 Stating Hypotheses In any significance test, the null hypothesis has the form H 0 : parameter = value The alternative hypothesis has one of the forms H a : parameter < value H a : parameter > value H a : parameter ≠ value To determine the correct form of H a, read the problem carefully. The alternative hypothesis is one-sided if it states that a parameter is larger than the null hypothesis value or if it states that the parameter is smaller than the null value. It is two-sided if it states that the parameter is different from the null hypothesis value (it could be either larger or smaller). The alternative hypothesis is one-sided if it states that a parameter is larger than the null hypothesis value or if it states that the parameter is smaller than the null value. It is two-sided if it states that the parameter is different from the null hypothesis value (it could be either larger or smaller).

7 The Practice of Statistics, 5 th Edition7 Stating Hypotheses The hypotheses should express the hopes or suspicions we have before we see the data. It is cheating to look at the data first and then frame hypotheses to fit what the data show. Hypotheses always refer to a population, not to a sample. Be sure to state H 0 and H a in terms of population parameters. It is never correct to write a hypothesis about a sample statistic, such as

8 The Practice of Statistics, 5 th Edition8 Watch “Juicy Pineapples”, p. 540

9 The Practice of Statistics, 5 th Edition9 The Reasoning of Significance Tests We can use software to simulate 400 sets of 50 shots assuming that the player is really an 80% shooter. You can say how strong the evidence against the player’s claim is by giving the probability that he would make as few as 32 out of 50 free throws if he really makes 80% in the long run. Suppose a basketball player claimed to be an 80% free-throw shooter. To test this claim, we have him attempt 50 free-throws. He makes 32 of them. His sample proportion of made shots is 32/50 = 0.64. What can we conclude about the claim based on this sample data? Based on the simulation, our estimate of this probability is 3/400 = 0.0075.

10 The Practice of Statistics, 5 th Edition10 The Reasoning of Significance Tests The observed statistic is so unlikely if the actual parameter value is p = 0.80 that it gives convincing evidence that the players claim is not true. There are two possible explanations for the fact that he made only 64% of his free throws. 1) The null hypothesis is correct. The player’s claim is correct (p = 0.8), and just by chance, a very unlikely outcome occurred. 2) The alternative hypothesis is correct. The population proportion is actually less than 0.8, so the sample result is not an unlikely outcome. An outcome that would rarely happen if the null hypothesis were true is good evidence that the null hypothesis is not true. Basic Idea

11 The Practice of Statistics, 5 th Edition11 Answer Concept 1 of 9.1 in your class work. Watch video “I’m a Great Free Throw Shooter”, p. 543 to get an idea of what a P value is.

12 The Practice of Statistics, 5 th Edition12 Interpreting P-Values The null hypothesis H 0 states the claim that we are seeking evidence against. The probability that measures the strength of the evidence against a null hypothesis is called a P-value. The probability, computed assuming H 0 is true, that the statistic would take a value as extreme as or more extreme than the one actually observed is called the P-value of the test. Small P-values are evidence against H 0 because they say that the observed result is unlikely to occur when H 0 is true. Large P-values fail to give convincing evidence against H 0 because they say that the observed result is likely to occur by chance when H 0 is true.

13 The Practice of Statistics, 5 th Edition13 Now watch “Healthy Bones”, p. 544 But is that significant?

14 The Practice of Statistics, 5 th Edition14 Statistical Significance The final step in performing a significance test is to draw a conclusion about the competing claims you were testing. We make one of two decisions based on the strength of the evidence against the null hypothesis (and in favor of the alternative hypothesis): reject H 0 or fail to reject H 0. Note: A fail-to-reject H 0 decision in a significance test doesn’t mean that H 0 is true. For that reason, you should never “accept H 0 ” or use language implying that you believe H 0 is true. In a nutshell, our conclusion in a significance test comes down to P-value small → reject H 0 → convincing evidence for H a P-value large → fail to reject H 0 → not convincing evidence for H a In a nutshell, our conclusion in a significance test comes down to P-value small → reject H 0 → convincing evidence for H a P-value large → fail to reject H 0 → not convincing evidence for H a

15 The Practice of Statistics, 5 th Edition15 Statistical Significance There is no rule for how small a P-value we should require in order to reject H 0. But we can compare the P-value with a fixed value that we regard as decisive, called the significance level. We write it as , the Greek letter alpha. When we use a fixed level of significance to draw a conclusion in a significance test, P-value < α → reject H 0 → convincing evidence for H a P-value ≥ α → fail to reject H 0 → not convincing evidence for H a When we use a fixed level of significance to draw a conclusion in a significance test, P-value < α → reject H 0 → convincing evidence for H a P-value ≥ α → fail to reject H 0 → not convincing evidence for H a If the P-value is smaller than alpha, we say that the data are statistically significant at level α. In that case, we reject the null hypothesis H 0 and conclude that there is convincing evidence in favor of the alternative hypothesis H a.

16 The Practice of Statistics, 5 th Edition16 Watch “Better Batteries”, p. 546 Answer the questions in Concept 2 of 9.1 in your classwork.

17 The Practice of Statistics, 5 th Edition17 Type I and Type II Errors When we draw a conclusion from a significance test, we hope our conclusion will be correct. But sometimes it will be wrong. There are two types of mistakes we can make. If we reject H 0 when H 0 is true, we have committed a Type I error. If we fail to reject H 0 when H a is true, we have committed a Type II error. If we reject H 0 when H 0 is true, we have committed a Type I error. If we fail to reject H 0 when H a is true, we have committed a Type II error. Truth about the population H 0 true H 0 false (H a true) Conclusion based on sample Reject H 0 Type I error Correct conclusion Fail to reject H 0 Correct conclusion Type II error

18 The Practice of Statistics, 5 th Edition18 Type I and Type II Errors The probability of a Type I error is the probability of rejecting H 0 when it is really true…this is exactly the significance level of the test. Significance and Type I Error The significance level α of any fixed-level test is the probability of a Type I error. That is, α is the probability that the test will reject the null hypothesis H 0 when H 0 is actually true. Consider the consequences of a Type I error before choosing a significance level.

19 The Practice of Statistics, 5 th Edition19 Watch the video “Perfect Potatoes”, p. 548 Answer the questions for Concept 3 of 9.1 in your classwork. Memorize the Type 1 and Type II Error chart. It will make everything much easier!

20 Section Summary In this section, we learned how to… The Practice of Statistics, 5 th Edition20 STATE the null and alternative hypotheses for a significance test about a population parameter. INTERPRET a P-value in context. DETERMINE whether the results of a study are statistically significant and MAKE an appropriate conclusion using a significance level. INTERPRET a Type I and a Type II error in context and GIVE a consequence of each. Finish your classwork for Section 9.1 NOW if needed. Complete the homework for Section 9.1 tonight. Quiz tomorrow! Significance Tests: The Basics


Download ppt "The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 9 Testing a Claim 9.1 Significance Tests:"

Similar presentations


Ads by Google