Sampling distribution

Slides:



Advertisements
Similar presentations
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 9 Inferences Based on Two Samples.
Advertisements

Objectives Look at Central Limit Theorem Sampling distribution of the mean.
AP Statistics Section 10.2 A CI for Population Mean When is Unknown.
Modular 13 Ch 8.1 to 8.2.
The Sampling Distribution Introduction to Hypothesis Testing and Interval Estimation.
Chapter 9 Hypothesis Testing II. Chapter Outline  Introduction  Hypothesis Testing with Sample Means (Large Samples)  Hypothesis Testing with Sample.
Clt1 CENTRAL LIMIT THEOREM  specifies a theoretical distribution  formulated by the selection of all possible random samples of a fixed size n  a sample.
Standard error of estimate & Confidence interval.
Bootstrapping applied to t-tests
Bootstrap spatobotp ttaoospbr Hesterberger & Moore, chapter 16 1.
Review of normal distribution. Exercise Solution.
Continuous Probability Distributions
Lesson Confidence Intervals about a Population Standard Deviation.
Introduction to Statistical Inference Chapter 11 Announcement: Read chapter 12 to page 299.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 9 Inferences Based on Two Samples.
9 Mar 2007 EMBnet Course – Introduction to Statistics for Biologists Nonparametric tests, Bootstrapping
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
Distribution of the Sample Means
Lecture 2 Review Probabilities Probability Distributions Normal probability distributions Sampling distributions and estimation.
1.State your research hypothesis in the form of a relation between two variables. 2. Find a statistic to summarize your sample data and convert the above.
Sample Variability Consider the small population of integers {0, 2, 4, 6, 8} It is clear that the mean, μ = 4. Suppose we did not know the population mean.
Estimation Chapter 8. Estimating µ When σ Is Known.
Sampling Distributions. Sampling Distribution Is the Theoretical probability distribution of a sample statistic Is the Theoretical probability distribution.
Limits to Statistical Theory Bootstrap analysis ESM April 2006.
: An alternative representation of level of significance. - normal distribution applies. - α level of significance (e.g. 5% in two tails) determines the.
Mystery 1Mystery 2Mystery 3.
Sampling and estimation Petter Mostad
Hypothesis Testing. Suppose we believe the average systolic blood pressure of healthy adults is normally distributed with mean μ = 120 and variance σ.
1 Chapter 23 Inferences About Means. 2 Inference for Who? Young adults. What? Heart rate (beats per minute). When? Where? In a physiology lab. How? Take.
Non-parametric Approaches The Bootstrap. Non-parametric? Non-parametric or distribution-free tests have more lax and/or different assumptions Properties:
Christopher, Anna, and Casey
Section 8.2: The Sampling Distribution of a Sample Mean
Statistical Inference
Introduction For inference on the difference between the means of two populations, we need samples from both populations. The basic assumptions.
Standard Errors Beside reporting a value of a point estimate we should consider some indication of its precision. For this we usually quote standard error.
Application of the Bootstrap Estimating a Population Mean
Introduction to Statistics for the Social Sciences SBS200 - Lecture Section 001, Fall 2016 Room 150 Harvill Building 10: :50 Mondays, Wednesdays.
Parameter Estimation.
Confidence Interval Estimation
Chapter 7 Sampling and Sampling Distributions
Basic Business Statistics (8th Edition)
3. The X and Y samples are independent of one another.
Chapter 6 Confidence Intervals.
Chapter 9 Hypothesis Testing
6-3The Central Limit Theorem.
Consolidation & Review
Chapter 7 Sampling Distributions.
Statistics in Applied Science and Technology
When we free ourselves of desire,
Inference on Mean, Var Unknown
Two-sided p-values (1.4) and Theory-based approaches (1.5)
Hypothesis Tests for a Population Mean in Practice
Confidence Interval (CI) for the Mean When σ Is Known
CI for μ When σ is Unknown
Elementary Statistics: Picturing The World
Chapter 7 Sampling Distributions.
Confidence Intervals Chapter 10 Section 1.
CENTRAL LIMIT THEOREM specifies a theoretical distribution
Why does sampling work?.
Tutorial 9 Suppose that a random sample of size 10 is drawn from a normal distribution with mean 10 and variance 4. Find the following probabilities:
Chapter 7 Sampling Distributions.
CHAPTER 15 SUMMARY Chapter Specifics
Chapter 8 Confidence Intervals.
Chapter 7 Sampling Distributions.
IE 355: Quality and Applied Statistics I Confidence Intervals
Continuous Probability Distributions
Confidence Intervals for the Mean (Large Samples)
Chapter 7 Sampling Distributions.
Statistical Inference for the Mean: t-test
How Confident Are You?.
Presentation transcript:

Sampling distribution Do not ‘read’ this. It is meant to be watched only.

Any and usually undefinable N POPULATION Any and usually undefinable N μ, σ Sample Size = N Start with just a single random sample from the population.

Any and usually undefinable N POPULATION Any and usually undefinable N μ, σ Sample Size = N Sample Size = N Sample Size = N Sample Size = N Sample Size = N All these hypothetical samples have the same N, but different parameter estimates (in this case, mean and standard deviation) for each sample.

Any and usually undefinable N POPULATION Any and usually undefinable N μ, σ Sample Size = N Sample Size = N Sample Size = N Sample Size = N Sample Size = N Note the sample means.

Any and usually undefinable N POPULATION Any and usually undefinable N μ, σ Sample Size = N Sample Size = N Sample Size = N Sample Size = N Sample Size = N Concerning means, we can think about a distribution of them, and how it would take shape.

Any and usually undefinable N POPULATION Any and usually undefinable N μ, σ If we were to create a sampling distribution, the distribution of means would have its own mean equal to μ, and standard deviation of σ/sqrt(N), and with a large N, be approximately normal. This is the Central Limit Theorem.

Any and usually undefinable N POPULATION Any and usually undefinable N μ, σ We don’t actually do this though (we only have our one sample and mean), so such a distribution is theoretical. Its mean, i.e. the population mean μ, is posited by the null hypothesis. The standard deviation would usually be estimated by our sample s.

Any and usually undefinable N POPULATION Any and usually undefinable N μ, σ μ The question now is, given this distribution, what is the probability of obtaining my sample estimate? It is the conditional probability, p(D|H0). Given the null hypothesis, what is the probability of obtaining this estimate or more extreme. Knowing the properties of the sampling distribution (e.g. normal) allows us to obtain these probabilities.

Any and usually undefinable N POPULATION Any and usually undefinable N μ, σ Lower Limit Upper Limit Alternatively, we could assume a sampling distribution centered on our estimate, obtain lower and upper limits with some confidence, and see if this confidence interval contains the population value proposed by the null hypothesis.

Any and usually undefinable N POPULATION Any and usually undefinable N μ, σ Sample Size = N Now what is bootstrapping again?

Any and usually undefinable N POPULATION Any and usually undefinable N μ, σ Sample N, Sample Size = N Now what is bootstrapping again?

Sample N, Boot Sample Size = N Boot Sample Size = N Boot Sample Size = N Boot Sample Size = N Boot Sample Size = N We get sample means from sampling with replacement from our original sample.

Sample N, Example in R. Most of the time bootstrapping is used this is done for you, but it’s good to demonstrate for yourself. After obtaining the bootstrap distribution (e.g. as we have done in lab previously): quantile(data, c(.025,.975)) #From lab example, data would be bootmean$thetastar. Lower Limit Upper Limit Now we have an empirical sampling distribution, not assumed, but based on our sample data. By finding the appropriate quantiles1, we can obtain a confidence interval as we did before (e.g. at .025 and .975 for a 95% CI), but not have to assume a theoretically normal distribution.

Summary Hopefully this will help keep some of this straight, but don’t worry if it takes awhile to get sampling distributions etc. down It’s a different way to think about things and will take some getting used to