Problems: Q&A chapter 6, problems Chapter 6:

Slides:



Advertisements
Similar presentations
Chapter 3 Hypothesis Testing. Curriculum Object Specified the problem based the form of hypothesis Student can arrange for hypothesis step Analyze a problem.
Advertisements

BCOR 1020 Business Statistics Lecture 21 – April 8, 2008.
Inference about a Mean Part II
Chapter 9 Hypothesis Testing.
Section 9.1 Introduction to Statistical Tests 9.1 / 1 Hypothesis testing is used to make decisions concerning the value of a parameter.
1 Level of Significance α is a predetermined value by convention usually 0.05 α = 0.05 corresponds to the 95% confidence level We are accepting the risk.
June 18, 2008Stat Lecture 11 - Confidence Intervals 1 Introduction to Inference Sampling Distributions, Confidence Intervals and Hypothesis Testing.
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
Statistics 101 Chapter 10. Section 10-1 We want to infer from the sample data some conclusion about a wider population that the sample represents. Inferential.
Hypothesis testing Chapter 9. Introduction to Statistical Tests.
6.1 - One Sample One Sample  Mean μ, Variance σ 2, Proportion π Two Samples Two Samples  Means, Variances, Proportions μ 1 vs. μ 2.
LECTURE 19 THURSDAY, 14 April STA 291 Spring
Stat 1510 Statistical Inference: Confidence Intervals & Test of Significance.
Inference We want to know how often students in a medium-size college go to the mall in a given year. We interview an SRS of n = 10. If we interviewed.
The Practice of Statistics Third Edition Chapter 10: Estimating with Confidence Copyright © 2008 by W. H. Freeman & Company Daniel S. Yates.
Chapter 9 Tests of Hypothesis Single Sample Tests The Beginnings – concepts and techniques Chapter 9A.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
Lecture PowerPoint Slides Basic Practice of Statistics 7 th Edition.
Ex St 801 Statistical Methods Inference about a Single Population Mean.
Logic and Vocabulary of Hypothesis Tests Chapter 13.
© Copyright McGraw-Hill 2004
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Created by Erin Hodgess, Houston, Texas Section 7-1 & 7-2 Overview and Basics of Hypothesis Testing.
Class Seven Turn In: Chapter 18: 32, 34, 36 Chapter 19: 26, 34, 44 Quiz 3 For Class Eight: Chapter 20: 18, 20, 24 Chapter 22: 34, 36 Read Chapters 23 &
Christopher, Anna, and Casey
Chapter Nine Hypothesis Testing.
STAT 312 Chapter 7 - Statistical Intervals Based on a Single Sample
9.3 Hypothesis Tests for Population Proportions
More on Inference.
Unit 5 – Chapters 10 and 12 What happens if we don’t know the values of population parameters like and ? Can we estimate their values somehow?
Chapter 9 Hypothesis Testing.
Chapter 5 STATISTICAL INFERENCE: ESTIMATION AND HYPOTHESES TESTING
Significance Test for the Difference of Two Proportions
Lecture Nine - Twelve Tests of Significance.
Chapter 4. Inference about Process Quality
CHAPTER 6 Statistical Inference & Hypothesis Testing
Chapters 20, 21 Hypothesis Testing-- Determining if a Result is Different from Expected.
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
Statistical inference: distribution, hypothesis testing
Chapter 8: Inference for Proportions
Hypothesis Testing: Hypotheses
Statistics in Applied Science and Technology
When we free ourselves of desire,
CONCEPTS OF HYPOTHESIS TESTING
Introduction to Inference
Hypothesis Tests for a Population Mean in Practice
Chapter 9 Hypothesis Testing.
More on Inference.
Chapter Review Problems
Introduction to Inference
Chapter 9 Hypothesis Testing.
Introduction to Inference
Econ 3790: Business and Economics Statistics
Chapter Nine Part 1 (Sections 9.1 & 9.2) Hypothesis Testing
Chapter 8: Estimating with Confidence
Essential Statistics Introduction to Inference
Interval Estimation and Hypothesis Testing
Hypothesis Testing.
LESSON 18: CONFIDENCE INTERVAL ESTIMATION
Lecture 10/24/ Tests of Significance
Intro to Confidence Intervals Introduction to Inference
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Confidence Intervals
Chapter 8: Estimating with Confidence
Chapter 6 Confidence Intervals.
Chapter 8: Estimating with Confidence
Interval Estimation Download this presentation.
Statistical Test A test of significance is a formal procedure for comparing observed data with a claim (also called a hypothesis) whose truth we want to.
How Confident Are You?.
Presentation transcript:

Problems: Q&A chapter 6, problems 1-10. Chapter 6: 6.1 6.5 6.11 6.27 6.35 6.41 6.45

Note: Examples in chapter 6 are based on data that come from a normal distribution: N(, ) and most usually we are estimating  based on the unrealistic assumption that  is known. Formal methods for estimating  when  is unknown follow in later chapters.

Example: Suppose we wish to compute the true mean score on the US college entrance exam, the population mean SAT score. These scores are known to be distributed normally with mean  and standard deviation . That is, if x1, x2, .., xn, represents an SRS from the population of students taking the SAT, we say that the x’s are normally distributed with mean  and standard deviation , or, more formally: x1, x2, .., xn ~ N(, ). We don’t know the mean, . Also, in a real situation, we will not know the standard deviation, . But we will proceed by assuming that we know  and that  = 100. Also, assume that the sample size is n=500.

Our objective is to estimate  from the sample data x1, x2, .., xn. We know (from chapter 5) that an unbiased estimator of  is the sample mean We also know (from chapter 5) that the sample mean also has a normal distribution with mean, the parameter we are estimating (because the sample mean is an unbiased estimator) and variance equal to  divided by the square root of the number of observations in the SRS. In other words ~ N(, /n).

The task in chapter 6 is to use this information to make formal probability statements about the unknown value . All of these statements pertain to the fact that the sample mean has a normal distribution with known variance 2.

Three important points that follow from the normal distribution: The 68-95-99.7 rule states that 95% of the time, the sample mean will lie within  2 standard deviations of the population mean . This implies that 95% of the time the population mean  lies within  2 standard deviations of the sample mean. Hence, 95% of all sample means calculated from SRS’s will contain the true value in the interval

Suppose we take an SRS and calculate the sample mean at 495. We construct the interval 4959 = (495-9, 495+9) = (486, 504). Either (486, 504) or (486, 504). The first situation arises 95% of the time and the second situation arises 5% of the time if we repeated the experiment (take an SRS, calculate the mean, calculate the interval) a very large number of times.

Confidence interval A level C confidence interval for a parameter is an interval computed from the sample data by a method that has probability C of producing an interval containing the true value of the parameter. Convention is to use C = 0.95 or C = 0.90.

Formal construction of confidence intervals This relies on two facts. First, all normal distributions have the same standardized forms. That is, Second, the central limit theorem (chapter 5) tells us that the normal distribution is approximately correct, even if the population is not normal.

Now, we find the central C area under the standard normal curve and use Table A to construct the “critical values”, -z* and +z*. For C = 0.95, z* = 1.645. For C = 0.95, z* = 1.960. For C = 0.99, z* = 2.576.

Formally:

We call the component the “margin of error.” Opportunities for reducing the margin of error are three, namely, reducing z*, reducing , or increasing n.

Controlling the margin of error Usually, we will choose conventional levels of C, for example, C = 0.95 and C = 0.90. The population standard deviation is often out of our control. Only the sample size, n, is within our control. We can select n, to set the margin of error,

Formally:

6.2 Tests of significance

Example: (example 6.6 page 453) Someone states their weight as “187 pounds.” We take four measures over a one month period, x1 = 190.5, x2 = 189.0, x3 = 195.5, x1 = 187.0. Sample mean = 190.5, and we know that the weights are normally distributed with (unknown) mean  and (known) standard deviation  = 3. Are these measures compatible with a true population mean of 187.0.

Example: (continued) First, restate: What is the probability of observing a sample mean of 190.5 or larger from a population with a true mean 187.0? Second, formalize: Null hypothesis, Ho:  = 187.0. Alternative hypothesis, Ha:   187.0. (This is a one-sided hypothesis.)

Example: (continued) Alternatively: First, restate: What is the probability of observing a sample mean of 190.5 or larger from a population with a true mean 187.0? Second, formalize: Null hypothesis, Ho:  = 187.0. Alternative hypothesis, Ha:   187.0. (This is a two-sided hypothesis.)

In this first example, we will compute the probability of observing a sample mean of 190.5 or larger when the true population mean is 1.87 to be prob = 0.01, which is 1% and, thus small. Hence, we conclude that the data are not compatible with the hypothesis, or, more precisely, we conclude that it is only 1% of the time likely that the computed sample mean could be consistent with the null hypothesis.

Suppose we take another sample and find that the computed sample mean is 188.5. This number is much closer to the hypothesized value 187.0 and, hence, it seems more likely that this result could have been generated by the hypothesized population mean 187.0. We will compute the probability of the event “obtaining a sample mean greater than or equal to 188.5 when the true population mean is 187.0 to be 0.16. So, about 16% of the time we expect such a result. This proportion is not particularly small, so we say that the data are not inconsistent with the hull hypothesis.

Calculations: All based on the z distribution, the standard-normal N(0, 1). There are two ways to proceed. First is based on a test statistic. Second is based on P-values. Both approaches exploit the z distribution.

Formally:

Once a critical value is chosen, z Once a critical value is chosen, z*, we reject the null hypothesis in favour of the alternative hypothesis if the computed value, z, lies further in the tail than the critical value z*. Important: Note the rules for one sided tests in the greater-than direction. Note the rules for one sided tests in the less-than direction. Note the rules for two-sided tests.

For one sided tests of the greater-than variety: Ho:  = o Ha:   o We reject Ho in favor of Ha if z  z*.

For one sided tests of the less-than variety: Ho:  = o Ha:   o We reject Ho in favor of Ha if z  z*.

For one sided tests of the not-equal-to variety: Ho:  = o Ha:   o We reject Ho in favor of Ha if z  zL* or z  zU*.

Example: Two-sided significance test. Ho:  = o = 187.0 Ha:   o = 187.0 You select C = 0.95 and compute 1-C/2 = 0.025. From Table D (last line) you find zL* = -1.96, and zU* = +1.96. You note z > zU* and you reject the null in favour of the alternative.

Example: One-sided significance test. Ho:  = o = 187.0 Ha:   o = 187.0 You select C = 0.95 and compute 1-C = 0.05. From Table D (last line) you find z* = +1.645. You note z > z* and you reject the null in favour of the alternative.

Example: One-sided significance test. Ho:  = o = 187.0 Ha:   o = 187.0 You select C = 0.95 and compute 1-C = 0.05. From Table D (last line) you find z* = -1.645. You note z > z* and you do not reject the null in favour of the alternative.

Example: Two-sided significance test. Ho:  = o = 187.0 Ha:   o = 187.0 You select C = 0.95 and compute 1-C/2 = 0.025. From Table D (last line) you find zL* = -1.96, and zU* = +1.96. You note zL* < z < zU* and you do not reject the null hypothesis.

Example of P-value calculations. Remember, the probability, computed assuming Ho is true, that the test statistic would take a value as extreme or more extreme than the observed value is the P-value. The smaller the P-value, the stronger the evidence against Ho.

Example: One-sided significance test. Ho:  = o = 187.0 Ha:   o = 187.0 Now, if Ho is true, then z = 1.00 is a single observation from the z, N(0,1), distribution. The probability of observing a z value greater than 1.00 is (from table A) p = 1.00 - 0.8413 = 0.1587.

Example: Two-sided significance test. Ho:  = o = 187.0 Ha:   o = 187.0 Now, if Ho is true, then z = 2.33 is a single observation from the z, N(0,1), distribution. The probability of observing a z value greater than 2.33 is (from table A) p = 1-0.9901 = 0.0099, but, because this is a two-sided test, we say that the P-value is p = 2  0.0099 = 0.0198.

We may cover this section next week, or not. Please look at chapter 7 in advance, especially sections 7.1 and 7.2.