Chapter 16 Inferential Statistics

Slides:



Advertisements
Similar presentations
HYPOTHESIS TESTING. Purpose The purpose of hypothesis testing is to help the researcher or administrator in reaching a decision concerning a population.
Advertisements

Chapter 7 Hypothesis Testing
Chapter 9 Introduction to the t-statistic
Inferential Statistics
CHAPTER 14: Confidence Intervals: The Basics
1 COMM 301: Empirical Research in Communication Lecture 15 – Hypothesis Testing Kwan M Lee.
Hypothesis Testing making decisions using sample data.
CHAPTER 21 Inferential Statistical Analysis. Understanding probability The idea of probability is central to inferential statistics. It means the chance.
Inference Sampling distributions Hypothesis testing.
Inference1 Data Analysis Inferential Statistics Research Methods Gail Johnson.
Statistical Significance What is Statistical Significance? What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant?
Statistical Significance What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant? How Do We Know Whether a Result.
1 The Basics of Regression Regression is a statistical technique that can ultimately be used for forecasting.
Chapter Sampling Distributions and Hypothesis Testing.
Chapter 9 Hypothesis Testing.
Chapter 12 Inferring from the Data. Inferring from Data Estimation and Significance testing.
Chapter 7 Probability and Samples: The Distribution of Sample Means
The problem of sampling error in psychological research We previously noted that sampling error is problematic in psychological research because differences.
Statistical hypothesis testing – Inferential statistics I.
AM Recitation 2/10/11.
1 © Lecture note 3 Hypothesis Testing MAKE HYPOTHESIS ©
Chapter 8 Introduction to Hypothesis Testing
1/2555 สมศักดิ์ ศิวดำรงพงศ์
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
F OUNDATIONS OF S TATISTICAL I NFERENCE. D EFINITIONS Statistical inference is the process of reaching conclusions about characteristics of an entire.
Chapter 11: Estimation Estimation Defined Confidence Levels
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
Topics: Statistics & Experimental Design The Human Visual System Color Science Light Sources: Radiometry/Photometry Geometric Optics Tone-transfer Function.
Lecture 14 Dustin Lueker. 2  Inferential statistical methods provide predictions about characteristics of a population, based on information in a sample.
Chapter 8 Introduction to Hypothesis Testing
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
A Broad Overview of Key Statistical Concepts. An Overview of Our Review Populations and samples Parameters and statistics Confidence intervals Hypothesis.
1 rules of engagement no computer or no power → no lesson no SPSS → no lesson no homework done → no lesson GE 5 Tutorial 5.
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
Agresti/Franklin Statistics, 1 of 122 Chapter 8 Statistical inference: Significance Tests About Hypotheses Learn …. To use an inferential method called.
The Practice of Statistics Third Edition Chapter 10: Estimating with Confidence Copyright © 2008 by W. H. Freeman & Company Daniel S. Yates.
Normal Distr Practice Major League baseball attendance in 2011 averaged 30,000 with a standard deviation of 6,000. i. What percentage of teams had between.
Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
10.1: Confidence Intervals Falls under the topic of “Inference.” Inference means we are attempting to answer the question, “How good is our answer?” Mathematically:
Interval Estimation and Hypothesis Testing Prepared by Vera Tabakova, East Carolina University.
Statistical Inference An introduction. Big picture Use a random sample to learn something about a larger population.
Introduction to Inference: Confidence Intervals and Hypothesis Testing Presentation 8 First Part.
Introduction to Inference: Confidence Intervals and Hypothesis Testing Presentation 4 First Part.
Hypothesis Testing An understanding of the method of hypothesis testing is essential for understanding how both the natural and social sciences advance.
Chapter 8 Parameter Estimates and Hypothesis Testing.
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
Ex St 801 Statistical Methods Inference about a Single Population Mean.
Stats Lunch: Day 3 The Basis of Hypothesis Testing w/ Parametric Statistics.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Please hand in homework on Law of Large Numbers Dan Gilbert “Stumbling on Happiness”
SAMPLING DISTRIBUTION OF MEANS & PROPORTIONS. SAMPLING AND SAMPLING VARIATION Sample Knowledge of students No. of red blood cells in a person Length of.
SAMPLING DISTRIBUTION OF MEANS & PROPORTIONS. SAMPLING AND SAMPLING VARIATION Sample Knowledge of students No. of red blood cells in a person Length of.
Chapter 13 Understanding research results: statistical inference.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
10.1 Estimating with Confidence Chapter 10 Introduction to Inference.
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Chapter 8: Estimating with Confidence Section 8.1 Confidence Intervals: The.
15 Inferential Statistics.
Chapter 9 Hypothesis Testing.
Unit 5: Hypothesis Testing
Statistics in Applied Science and Technology
Interval Estimation and Hypothesis Testing
Statistical Inference for Managers
CHAPTER 9 Testing a Claim
STA 291 Spring 2008 Lecture 13 Dustin Lueker.
Comparing Two Proportions
Comparing Two Proportions
How Confident Are You?.
Presentation transcript:

Chapter 16 Inferential Statistics Inferential statistics is defined as the branch of statistics that is used to make inferences about the characteristics of a population based on sample data. • The goal is to go beyond the data at hand and make inferences about population parameters. • In order to use inferential statistics, it is assumed that either random selection or random assignment was carried out (i.e., some form of randomization must is assumed).

Questions Asked in Science Is there a relationship? Answered by Null Hypothesis Significance Tests (NHST) t tests, F tests, χ2, p-values, for example What kind of relationship? (linear, curvilinear?) How strong is the relationship? Answered by effect size measures, not NHST’s

The Logic of Inferential Statistics Population: the entire universe of individuals we are interested in studying Sample: the selected subgroup that is actually observed and measured (with sample size N) Sampling Distribution of the Statistic: A theoretical distribution that describes how a statistic behaves across a large number of samples

The Three Distributions Used in Inferential Statistics I. Population Inference Selection III. Sampling Distribution of the Statistic Evaluation II. Sample

The Logic of NHST’s Samples are drawn from the population of interest Samples are observed and measured and sample statistics are calculated In order to determine whether sample statistics are consistent with a null hypothesis or not, sample statistics are compared to a model that describes what is expected for sample statistics This model is the sampling distribution of the statistic

Sampling Distributions Sampling distributions allow us to make "probability" statements in inferential statistics. • A sampling distribution is defined as "The theoretical probability distribution of the values of a statistic that result when all possible random samples of a particular size are drawn from a population." • One specific type of sampling distribution is called the sampling distribution of the mean. To generate this distribution by hand (not done in practice), you would randomly select a sample, calculate the mean, randomly select another sample, calculate the mean, and continue this process until you have calculated the means for a very large number of samples. You could then construct a histogram to show the sampling distribution of the mean. This distribution would tell you how likely or probable it is to get any particular value of the mean.

• No matter what the shape of the population distribution or the sample, the sampling distribution of the mean is normally distributed (as long as your sample size is about 30 or more for your sampling). • Note that the mean of the sampling distribution of the mean is equal to the population mean! That tells you that repeated sampling will, over the long run, estimate the correct mean. The variance shows you that sample means will tend to be somewhat different from the true population mean in an individual sample.

Although I just described the sampling distribution of the mean, it is important to remember that a sampling distribution can be obtained for any statistic. For example, you could also obtain the following sampling distributions: • Sampling distribution of the percentage (or proportion). • Sampling distribution of the variance. • Sampling distribution of the correlation. • Sampling distribution of the regression coefficient. • Sampling distribution of the difference between two means.

The standard deviation of a sampling distribution is called the standard error (SE). In other words, the standard error is just a special kind of standard deviation and you learned what a standard deviation was in the last chapter. The smaller the standard error, the more precisely the statistic has been estimated. Sample size has a powerful effect on precision of estimation:

Estimation The key estimation question is "Based on my random sample, what is my estimate of the population parameter?” • The basic idea is that you are going to use your sample data to provide information about the population. There are actually two types of estimation. • They can be first understood through the following analogy: Let's say that you take your car to your local car dealer's service department and you ask the service manager how much it will cost to repair your car. If the manager says it will cost you $500 then she is providing a point estimate. If the manager says it will cost somewhere between $400 and $600 then she is providing an interval estimate.

In other words, a point estimate is a single number, and an interval estimate is a range of numbers. • A point estimate is the value of your sample statistic (e.g., your sample mean or sample correlation), and it is used to estimate the population parameter (e.g., the population mean or the population correlation) • For example, if you take a random sample from adults living an the United States and you find that the average income is $45,000, then your best guess or your point estimate for the population of adults in the U.S. will be $45,000. • Again, whenever you engage in point estimation, all you need to do is to use the value of your sample statistic as your "best guess" (i.e., as your estimate) of the (unknown) population parameter.

Oftentimes, we like to put an interval around our point estimates so that we realize that the actual population value is somewhat different from our point estimate because sampling error is always present in sampling. • An interval estimate (also called a confidence interval) is a range of numbers inferred from the sample that has a known probability of capturing the population parameter over the long run (i.e., over repeated sampling). The probability relates to the area in a normal curve

99.7% 95% 68% μ – 3σ μ – 2σ μ – σ μ μ + σ μ + 2σ μ + 3σ

• The "beauty" of confidence intervals is that we know the probability of the true population parameter. • Specifically, with a 95 percent confidence interval, you are able to be "95% confident" that it will include the population parameter. • For example, you might take the point estimate of annual income of U.S. adults of $45,000 and surround it by a 95% confidence interval and find the interval is $43,000 to $47,000. You can be "95% confident" that the average income is somewhere between $43,000 and $47,000.

Hypothesis Testing (NHST) is the branch of inferential statistics concerned with how well the sample data support a null hypothesis. • First note that the null hypothesis is usually the prediction that there is no relationship in the population or no difference between groups • The alternative hypothesis is the logical opposite of the null hypothesis and says there is a relationship in the population or the groups are different • The researcher’s interest is in the alternative hypothesis • The question answered in hypothesis testing: "Is the value of my sample statistic unlikely enough (assuming that the null hypothesis is true) for me to reject the null hypothesis and tentatively accept the alternative hypothesis?“ • Note that it is the null hypothesis that is directly tested in hypothesis testing (not the alternative hypothesis).

You may be wondering, when do you actually reject the null hypothesis and make the decision to tentatively accept the alternative hypothesis? • You reject the null hypothesis when the probability of your result assuming a true null is very small. That is, you reject the null when your sample result would be very unlikely if the null was true. • In particular, you set a significance level (also called the alpha level) to use in your research study to decide when to reject the null as a plausible description of the sample.

By convention, the significance level (alpha) is set at p = By convention, the significance level (alpha) is set at p = .05 or smaller When your the sample statistic is compared to the sampling distribution, a probability value (p-value) is obtained for your sample statistic. If that p-value is less than the decision rule (.05), the null hypothesis is rejected. If your p-value is larger (more probable) than .05, no conclusion is reached.

• The probability value (p-value) is a number that tells you the probability of your result or a more extreme result when it is assumed that there is no relationship in the population (i.e., when you are assuming that the null hypothesis is true which is what we do in hypothesis testing and in jurisprudence). • The significance level (alpha) is the decision rule, chosen in advance (a priori), that you will use to conclude that the null should be rejected. By convention in education and the social sciences, alpha is usually chosen as .05 or a smaller value. • The logic of the NHST is the following: If my sample result is so rare (p < .05) that is very unlikely to have occurred when the null is true, then the null must be false and something else must be occurring (i.e., there really is a relationship or a difference between groups)

• Statistical significance does not tell you whether you have practical significance. • If a finding is statistically significant then you can claim that the evidence suggests that the observed result (e.g., your observed correlation or your observed difference between two means) was probably not just due to chance. • An effect size measure can provide additional important information to help interpret the strength of a statistically significant relationship. An effect size indicator is defined as a measure of the strength of a relationship. • A finding is practically significant when the difference between the means or the size of the correlation is big enough, in your opinion, to be of practical use. ((For example, a correlation of .15 would probably not be practically significant, even if it was statistically significant. On the other hand, a correlation of .85 would probably be practically significant. )) • Practical significance requires you to make a non-quantitative decision and to think about what the impact or utility of a result would be in application