Presentation is loading. Please wait.

Presentation is loading. Please wait.

+ Chapter 9 Testing a Claim 9.1Significance Tests: The Basics 9.2Tests about a Population Proportion 9.3Tests about a Population Mean.

Similar presentations


Presentation on theme: "+ Chapter 9 Testing a Claim 9.1Significance Tests: The Basics 9.2Tests about a Population Proportion 9.3Tests about a Population Mean."— Presentation transcript:

1 + Chapter 9 Testing a Claim 9.1Significance Tests: The Basics 9.2Tests about a Population Proportion 9.3Tests about a Population Mean

2 + Section 9.1 Significance Tests: The Basics After this section, you should be able to… STATE correct hypotheses for a significance test about a population proportion or mean. INTERPRET P-values in context. INTERPRET a Type I error and a Type II error in context, and give the consequences of each. DESCRIBE the relationship between the significance level of a test, P(Type II error), and power. Learning Objectives

3 + CONFIDENCE INTERVALS are one of the two most common types of statisticalinference. Use a confidence interval when your goal isto estimate a population parameter. The second common type of inference,called SIGNIFICANCE TESTS, has a different goal To assess the evidence provided by dataabout some claim concerning a population.

4 + Activity: I am a great free shooter!!! A basketball player claims to make 80% of the free throws that he attempts. We might think he might be exaggerating. To test this claim, we will ask him to shoot some free throws- virtually. www.whfreeman.com/tps4e

5 + In the free throw activity, Parameter of interest: proportion p of free throws that the player will make if he shoots forever Player’s claim: p= 0.80 To test the claim, through the applet, we simulate 25 shots. If the player makes only 10 of 25 throws ( p-hat = 0.4) which is strong evidence that he does not actually shoot 80%. What if he makes 19 of 25 throws (p-hat= 0.76)? This provides some evidence that his true long term percent may be less than 80%, but is not nearly as convincing as p-hat=0.40. Statistical tests weigh the evidence against a claim like p= 0.8.

6 + A SIGNIFICANCE TEST is a formal procedure for comparing observed data with a claim (also called a hypothesis) whose truth we want to assess. The claim is a statement about a parameter, like the population proportion p or the population mean µ. We express the results of a significance test in terms of a probability that measures how well the data and the claim agree. In this chapter, we’ll learn the underlying logic of statistical tests, how to perform tests about population proportions and population means, and how tests are connected to confidence intervals.

7 + The Reasoning of Significance Tests Statistical tests deal with claims about a population. Tests ask if sample data give good evidence against a claim. A test might say, “If we took many random samples and the claim were true, we would rarely get aresult like this.” To get a numerical measure of how strong the sampleevidence is, replace the vague term “rarely” by a probability. We can use software to simulate 400 sets of 50 shots assuming that the player is really an 80% shooter. You can say how strong the evidence against the player’s claim is by giving the probability that he would make as few as 32 out of 50 free throws if he really makes 80% in the long run. Suppose a basketball player claimed to be an 80% free-throw shooter. To test this claim, we have him attempt 50 free-throws. He makes 32 of them. His sample proportion of made shots is 32/50 = 0.64. What can we conclude about the claim based on this sample data?

8 + The Reasoning of Significance Tests Based on the evidence, we might conclude the player’s claim is incorrect. In reality, there are two possible explanations for the fact that he made only 64% of his free throws. Significance Tests: The Basics 1) The player’s claim is correct (p = 0.8), and by bad luck, a very unlikely outcome occurred. 2) The population proportion is actually less than 0.8, so the sample result is not an unlikely outcome. An outcome that would rarely happen if a claim were true is good evidence that the claim is not true. Basic Idea

9 + Stating Hypotheses A significance test starts with a careful statement of the claims we want to compare. The first claim is called the null hypothesis. Usually, the null hypothesis is a statement of “no difference.” The claim we hope orsuspect to be true instead of the null hypothesis is called the alternative hypothesis. Significance Tests: The Basics Definition: The claim tested by a statistical test is called the null hypothesis (H 0 ). The test is designed to assess the strength of the evidence against the null hypothesis. Often the null hypothesis is a statement of “NO DIFFERENCE.” The claim about the population that we are trying to find evidence for is the alternative hypothesis (H a ). In the free-throw shooter example, our hypotheses are H 0 : p = 0.80 H a : p < 0.80 where p is the long-run proportion of made free throws.

10 + The NULL HYPOTHESIS ( H 0 ) is assumed to be true to begin with. It is a claim about the population parameter that is assumed to be true until it is declared false. ------------------------------------------------------- ALTERNATIVE HYPOTHESIS (H a ) It is a claim about a population parameter that will be true if the null hypothesis is false

11 The null hypothesis Stating the null hypothesis is the next step in the test The null hypothesis says nothing has changed or there is no effect If it is true then the variation is due to chance It is written as Or when you are comparing two situations, It can be written as It is read “H-nought”

12 The alternative hypothesis The alternative hypothesis is that something has changed There is some variation in the statement because it can be >, < or ≠ It is written as It is pronounced “H-a” In this case it is μ > 0

13 + Stating Hypotheses Definition: The alternative hypothesis is one-sided if it states that a parameter is larger than the null hypothesis value or if it states that the parameter is smaller than the null value. It is two-sided if it states that the parameter is different from the null hypothesis value (it could be either larger or smaller). In any significance test, the null hypothesis has the form H 0 : parameter = value The alternative hypothesis has one of the forms H a : parameter < value H a : parameter > value H a : parameter ≠ value ( different) To determine the correct form of H a, read the problem carefully.

14 + IMPORTANT TIPS:-------------------- Hypotheses always refer to a population, not to a sample. Be sure to state H 0 and H a in terms of POPULATION PARAMETERS. It is never correct to write a hypothesis about a sample statistic, such as

15 + Example: At the Hawaii Pineapple Company, managers are interested in the size of the pineapples grown in the company’s fields. Last year, the mean weight of the pineapples harvested from one large field was 31 ounces. A different irrigation system was installed in this field after the growing season. Managers wonder how this change will affect the mean weight of pineapples grown in the field this year. a)Describe the parameter of interest in this setting. b) State appropriate hypotheses for performing a significance test. The parameter of interest is the mean weight µ of all the pineapples grown this year. Because managers wonder whether the mean weight of this year’s pineapples will differ from last year’s mean weight of 31 ounces, the alt. hypotheses is two sided : that is either µ 31. For simplicity, we write this as µ ≠ 31. That is, H 0 : µ = 31 H a : µ ≠ 31

16 + Try: Mike is an avid golfer who would like to improve his play. A friend suggests getting new clubs and lets Mike try out his 7-iron. Based on years of experience, Mike has established that the mean distance that balls travel when hit with his old 7-iron is µ = 175 yards with a standard deviation of σ = 15 yards. He is hoping that this new club will make his shots with a 7-iron more consistent (less variable), so he goes to the driving range and hits 50 shots with the new 7-iron. Problem: (a) Describe the parameter of interest in this setting. (b) State appropriate hypotheses for performing a significance test. Solution:(a) Since Mike is interested in being more consistent, the parameter of interest is the standard deviation σ of the distance he hits the ball when using the new 7-iron. (b) Because Mike wants to be more consistent, he wants the standard deviation of the distance he hits the ball to be smaller than 15 yards.,

17 + Do CYU : P- 532. 1.(a) Parameter of interest : p= proportion of students at Jannie’s high school who get less than 8 hours of sleep at night. (b) The hypotheses are: 2. (a) Parameter of interest : µ = mean amount of time that it takes to complete the census form. (b) The hypotheses are:

18 + The ides of getting null hypothesis that we want to find evidence against seems odd at first. THINK OF A CRIMINAL TRIAL.  The defendant is “ innocent until proven guilty”. That is the Null Hypothesis is Innocence and the prosecution must try to provide convincing evidence against this hypothesis. That’ s exactly how statistical tests work ( although in Stat, we deal with evidence provided by data and use a probability to say how strong the evidence is.)

19 + Interpreting P -Values The null hypothesis H 0 states the claim that we are seeking evidence against. The probability that measures the strength of the evidence against a null hypothesis is called a P -value. Definition: The probability, computed assuming H 0 is true, that the statistic would take a value as extreme as or more extreme than the one actually observed is called the P-value of the test. Smaller the P-value, the stronger the evidence against H 0 provided by the data. Small P-values are evidence against H 0 because they say that the observed result is unlikely to occur when H 0 is true. Large P-values FAIL to give convincing evidence against H 0 because they say that the observed result is likely to occur by chance when H 0 is true.

20 + Example: Studying Job Satisfaction a)Explain what it means for the null hypothesis to be true in this setting. b) Interpret the P-value in context. In this setting, H 0 : µ = 0 says that the mean difference in satisfaction scores (self-paced - machine-paced) for the entire population of assembly-line workers at the company is 0. If H 0 is true, then the workers don’t favor one work environment over the other, on average. For the job satisfaction study, the hypotheses are H 0 : µ = 0 H a : µ ≠ 0

21 + Statistical Significance: The final step in performing a significance test is to draw a conclusion about the competing claims youwere testing. We will make one of two decisions based on the strength of theevidence against the null hypothesis (and in favor of the alternativehypothesis) -- reject H 0 or fail to reject H 0. If our sample result is too unlikely to have happened by chanceassuming H 0 is true, then we’ll reject H 0. Otherwise, we will fail to reject H 0. Significance Tests: The Basics Note: A fail-to-reject H 0 decision in a significance test doesn’t mean that H 0 is true. For that reason, you should never “accept H 0 ” or use language implying that you believe H 0 is true.

22 + Statistical Significance There is no rule for how small a P -value we should require in order to reject H 0 — it’s a matter of judgment and depends on the specific circumstances. But we can compare the P -value with a fixed value that we regard as decisive, called the significance level. We write it as α, the Greek letter alpha. When our P -value is less than the chosen α, we say that the result is statistically significant. Definition: If the P-value is smaller than alpha, we say that the data are statistically significant at level α. In that case, we reject the null hypothesis H 0 and conclude that there is convincing evidence in favor of the alternative hypothesis H a. When we use a fixed level of significance to draw a conclusion in a significance test, P-value < α → reject H 0 → conclude H a (in context) P-value ≥ α → fail to reject H 0 → cannot conclude H a (in context)

23 Recap: When to Reject: P < -- Statistically Significant at level -- Reject the Null Hypothesis P ≥ --Not Sta. Significant at level -- Fail to Reject the Null. ( NEVER SAY ACCEPT the Null).

24 + Rejecting or Failing to Reject the Null Hypothesis If the p-value is less than the significance level, reject the null hypothesis. For example, if α = 0.05 and the p-value is 0.03, reject the null hypothesis because we expect to see the observed outcome only 3% of the time if the null hypothesis is true. So the observed outcome isn’t very likely.

25 + If the p-value is greater than the significance level, fail to reject the null hypothesis. For example, if α = 0.05 and the p-value is 0.15, fail to reject the null hypothesis. The observed outcome is expected 15% of the time if the null hypothesis is true. This may not seem very likely, but it is more likely than 5% so the conclusion is to fail to reject the null hypothesis, and we say, “there is not sufficient evidence to reject the null hypothesis.”

26 + Why shouldn’t the conclusion be, “there is sufficient evidence to accept the null hypothesis”? It is a convention based on the fact that in mathematics, statements are not proved with examples. A claim can be disproven with one example, but even one million examples in favor of the claim can’t prove it. To borrow a common phrase, “Absence of evidence is not evidence of absence”. However, hypothesis tests don’t actually prove anything anyways. They are just a method of judging the evidence for or against a hypothesis. Yet, the tradition is strong enough, that a conclusion should never be, “there is sufficient evidence to accept the null hypothesis”.

27 + Analogy : Read it later Until the 17th century Europeans thought every swan was white because for centuries, every swan they saw was white. Then, a black swan was discovered in Australia, instantly disproving the hypothesis that all swans are white.

28 + Analogy Suppose a person thinks that there might have been a skunk in his yard the previous night. A null hypothesis is that there was no skunk in the yard (status quo). The alternative hypothesis would then be that there was a skunk in the yard. H 0 : There was no skunk in the yard. H A : There was a skunk in the yard. He could go outside the next day and look for evidence that there was a skunk. If he finds skunk fur or smells a skunk, then he would have evidence to reject the null hypothesis in favor of the alternative hypothesis (that there was a skunk).

29 + On the other hand, if he doesn’t find evidence that a skunk was there, that does not mean that the null hypothesis is true. A skunk could have been there without leaving evidence. That is why he shouldn’t say he accepts the null hypothesis. He doesn’t know for sure that there wasn’t a skunk. He just doesn’t have evidence to support the claim that there was a skunk. So he says there is not sufficient evidence to reject the null hypothesis or that he fails to reject the null hypothesis. If he rejects the null hypothesis, he could technically say that he accepts the alternative hypothesis. However, tradition dictates that conclusions are always stated as rejecting or failing to reject hypothesis rather than accepting hypothesis. If he finds skunk fur in the yard, he would reject the null hypothesis. Yet he still hasn’t proved that there was a skunk. The dog could have brought the fur into the yard. Because he hasn’t proved the alternative hypothesis, (he might have strong evidence that there was a skunk, but he hasn’t proven it) he shouldn’t say that he accepts the alternative hypothesis.

30 + Example: Better Batteries Significance Tests: The Basics a)What conclusion can you make for the significance level α = 0.05? b) What conclusion can you make for the significance level α = 0.01? Since the P-value, 0.0276, is less than α = 0.05, the sample result is statistically significant at the 5% level. We have sufficient evidence to reject H 0 and conclude that the company’s deluxe AAA batteries last longer than 30 hours, on average. A company has developed a new deluxe AAA battery that is supposed to last longer than its regular AAA battery. However, these new batteries are more expensive to produce, so the company would like to be convinced that they really do last longer. Based on years of experience, the company knows that its regular AAA batteries last for 30 hours of continuous use, on average. The company selects an SRS of 15 new batteries and uses them continuously until they are completely drained. A significance test is performed using the hypotheses H 0 : µ = 30 hours H a : µ > 30 hours where µ is the true mean lifetime of the new deluxe AAA batteries. The resulting P- value is 0.0276. Since the P-value, 0.0276, is greater than α = 0.01, the sample result is not statistically significant at the 1% level. We do not have enough evidence to reject H 0 in this case. therefore, we cannot conclude that the deluxe AAA batteries last longer than 30 hours, on average.

31 + Remember 3 C’s Connection Conclusion Context

32 + Try P-547 Ex 13. (a) If the mean score on the SSHA for older students at this school is really 115, there is a 1.01% chance of finding a sample of 45 older students with a mean score of at least 125.7. (b) Since the P- value 0.0101 < 0.05, we would reject the null hypothesis if α = 0.05. Technically the P-value 0.0101 > 0.01 so we would fail to reject the null hypothesis if α = 0.01. However, we note that the P-value is very close to this.

33 + Type I and Type II Errors When we draw a conclusion from a significance test, we hope our conclusion will be correct. But sometimes it will be wrong. There are twotypes of MISTAKES we can make. We can reject the null hypothesis when it’s actually true, known as a Type I error, or we can fail to reject a false null hypothesis, which is a Type II error. Definition: If we reject H 0 when H 0 is true, we have committed a Type I error. If we fail to reject H 0 when H 0 is false, we have committed a Type II error. Truth about the population H 0 true H 0 false (H a true) Conclusion based on sample Reject H 0 Type I error Correct conclusion Fail to reject H 0 Correct conclusion Type II error

34 + Example: Perfect Potatoes Significance Tests: The Basics Describe a Type I and a Type II error in this setting, and explain the consequences of each. A Type I error would occur if the producer concludes that the proportion of potatoes with blemishes is greater than 0.08 when the actual proportion is 0.08 (or less). Consequence: The potato-chip producer sends the truckload of acceptable potatoes away, which may result in lost revenue for the supplier. A Type II error would occur if the producer does not send the truck away when more than 8% of the potatoes in the shipment have blemishes. Consequence: The producer uses the truckload of potatoes to make potato chips. More chips will be made with blemished potatoes, which may upset consumers. A potato chip producer and its main supplier agree that each shipment of potatoes must meet certain quality standards. If the producer determines that more than 8% of the potatoes in the shipment have “blemishes,” the truck will be sent away to get another load of potatoes from the supplier. Otherwise, the entire truckload will be used to make potato chips. To make the decision, a supervisor will inspect a random sample of potatoes from the shipment. The producer will then perform a significance test using the hypotheses H 0 : p = 0.08 H a : p > 0.08 where p is the actual proportion of potatoes with blemishes in a given truckload.

35 + Do CYU P- 539 1. A Type I error in this setting would be to say that the new batteries last longer than 30 hours when, in fact, their average lifetime is 30 hours. 2. A Type II error in this setting would be to say that the new batteries do not last longer than 30 hours when, in fact, they do last longer than 30 hours. 3.The result of a Type I error would be that the company spends the extra money to produce these new batteries when they aren’t any better than the older, cheaper, type. The result of a Type II error would be that the company would not produce the new batteries, even though they were better. Arguments could be made for either of the errors being worse. For example, with a Type I error, the company would be wasting its money producing a new battery that wasn’t any better. With a Type II error the company would be losing out on potential profits by not producing the new, better, battery.

36 + Error Probabilities We can assess the performance of a significance test by looking at the probabilities of the two types of error. That’s because statistical inferenceis based on asking, “What would happen if I did this many times?” For the truckload of potatoes in the previous example, we were testing H 0 : p = 0.08 H a : p > 0.08 where p is the actual proportion of potatoes with blemishes. Suppose that the potato-chip producer decides to carry out this test based on a random sample of 500 potatoes using a 5% significance level (α = 0.05).

37 + The shaded area in the right tail is 5%. Sample proportion values to the right of the green line at 0.0999 will cause us to reject H 0 even though H 0 is true. This will happen in 5% of all possible samples. That is, P(making a Type I error) = 0.05.

38 + Error Probabilities P(Type I error) = P( Reject H 0 ) when H 0 is really true. This is exactly the significance level of the test. Significance Tests: The Basics The significance level α of any fixed level test is the probability of a Type I error. That is, α is the probability that the test will reject the null hypothesis H 0 when H 0 is in fact true. Consider the consequences of a Type I error before choosing a significance level. Significance and Type I Error

39 + Discuss with your group : P- 540 After Significance and Type I Error, What about Type II……………… …………………………………..the test is.

40 + What about Type II errors? A significance test makes a Type II error when it fails to reject a null hypothesis that really is false. There are many values of the parameter that satisfy the alternative hypothesis, so we concentrate on one value. We can calculate the probability that a test does reject H 0 when an alternative is true. This probability is called the POWER of the test against that specific alternative. Definition: The power of a test against a specific alternative is the probability that the test will reject H 0 at a chosen significance level α when the specified alternative value of the parameter is true.

41 + Power of a test

42 +

43 +

44 + Error Probabilities The potato-chip producer wonders whether the significance test of H 0 : p = 0.08 versus H a : p > 0.08 based on a random sample of 500 potatoes has enough power to detect a shipment with, say, 11% blemished potatoes. In this case, aparticular Type II error is to fail to reject H 0 : p = 0.08 when p = 0.11. Significance Tests: The Basics Earlier, we decided to reject H 0 at α = 0.05 if our sample yielded a sample proportion to the right of the green line. What if p = 0.11? Since we reject H 0 at α= 0.05 if our sample yields a proportion > 0.0999, we’d correctly reject the shipment about 75% of the time. The power of a test against any alternative is 1 minus the probability of a Type II error for that alternative; that is, power = 1 - β. Power and Type II Error

45 + Video : Two tailed test http://www.youtube.com/watch?v=k80pME7mWRM

46 + Planning Studies: The Power of a Statistical Test How large a sample should we take when we plan to carry out a significance test? The answer depends on what alternative values of theparameter are important to detect. Significance Tests: The Basics Here are the questions we must answer to decide how many observations we need: 1.Significance level. How much protection do we want against a Type I error — getting a significant result from our sample when H 0 is actually true? 2.Practical importance. How large a difference between the hypothesized parameter value and the actual parameter value is important in practice? 3.Power. How confident do we want to be that our study will detect a difference of the size we think is important? Summary of influences on the question “How many observations do I need?” If you insist on a smaller significance level (such as 1% rather than 5%), you have to take a larger sample. A smaller significance level requires stronger evidence to reject the null hypothesis. If you insist on higher power (such as 99% rather than 90%), you will need a larger sample. Higher power gives a better chance of detecting a difference when it is really there. At any significance level and desired power, detecting a small difference requires a larger sample than detecting a large difference.

47 + Section 9.1 Significance Tests: The Basics In this section, we learned that… A significance test assesses the evidence provided by data against a null hypothesis H 0 in favor of an alternative hypothesis H a. The hypotheses are stated in terms of population parameters. Often, H 0 is a statement of no change or no difference. H a says that a parameter differs from its null hypothesis value in a specific direction (one-sided alternative) or in either direction (two-sided alternative). The reasoning of a significance test is as follows. Suppose that the null hypothesis is true. If we repeated our data production many times, would we often get data as inconsistent with H 0 as the data we actually have? If the data are unlikely when H 0 is true, they provide evidence against H 0. The P-value of a test is the probability, computed supposing H 0 to be true, that the statistic will take a value at least as extreme as that actually observed in the direction specified by H a. Summary

48 + Section 9.1 Significance Tests: The Basics Small P-values indicate strong evidence against H 0. To calculate a P-value, we must know the sampling distribution of the test statistic when H 0 is true. There is no universal rule for how small a P-value in a significance test provides convincing evidence against the null hypothesis. If the P-value is smaller than a specified value α (called the significance level), the data are statistically significant at level α. In that case, we can reject H 0. If the P-value is greater than or equal to α, we fail to reject H 0. A Type I error occurs if we reject H 0 when it is in fact true. A Type II error occurs if we fail to reject H 0 when it is actually false. In a fixed level α significance test, the probability of a Type I error is the significance level α. The power of a significance test against a specific alternative is the probability that the test will reject H 0 when the alternative is true. Power measures the ability of the test to detect an alternative value of the parameter. For a specific alternative, P(Type II error) = 1 - power. Summary

49 + Looking Ahead… We’ll learn how to test a claim about a population proportion. We’ll learn about Carrying out a significance test The one-sample z test for a proportion Two-sided tests Why confidence intervals give more information than significance tests In the next Section…


Download ppt "+ Chapter 9 Testing a Claim 9.1Significance Tests: The Basics 9.2Tests about a Population Proportion 9.3Tests about a Population Mean."

Similar presentations


Ads by Google