Null Hypothesis Testing

Slides:



Advertisements
Similar presentations
Introduction to Hypothesis Testing
Advertisements

Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
Statistics.  Statistically significant– When the P-value falls below the alpha level, we say that the tests is “statistically significant” at the alpha.
Hypothesis Testing making decisions using sample data.
STAT 135 LAB 14 TA: Dongmei Li. Hypothesis Testing Are the results of experimental data due to just random chance? Significance tests try to discover.
Statistical Significance What is Statistical Significance? What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant?
HYPOTHESIS TESTING Four Steps Statistical Significance Outcomes Sampling Distributions.
Hypothesis Testing Steps of a Statistical Significance Test. 1. Assumptions Type of data, form of population, method of sampling, sample size.
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Statistical Significance What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant? How Do We Know Whether a Result.
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
PSY 1950 Confidence and Power December, Requisite Quote “The picturing of data allows us to be sensitive not only to the multiple hypotheses that.
S519: Evaluation of Information Systems Social Statistics Inferential Statistics Chapter 8: Significantly significant.
PY 427 Statistics 1Fall 2006 Kin Ching Kong, Ph.D Lecture 6 Chicago School of Professional Psychology.
PSY 307 – Statistics for the Behavioral Sciences
Inferential Statistics
Inferential Statistics
Thomas Songer, PhD with acknowledgment to several slides provided by M Rahbar and Moataza Mahmoud Abdel Wahab Introduction to Research Methods In the Internet.
Chapter 4 Hypothesis Testing, Power, and Control: A Review of the Basics.
1 © Lecture note 3 Hypothesis Testing MAKE HYPOTHESIS ©
Tuesday, September 10, 2013 Introduction to hypothesis testing.
Chapter 8 Introduction to Hypothesis Testing
Hypothesis Testing.
Let’s flip a coin. Making Data-Based Decisions We’re going to flip a coin 10 times. What results do you think we will get?
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
1 Statistical Inference Greg C Elvers. 2 Why Use Statistical Inference Whenever we collect data, we want our results to be true for the entire population.
Hypothesis Testing: One Sample Cases. Outline: – The logic of hypothesis testing – The Five-Step Model – Hypothesis testing for single sample means (z.
Chapter 8 Introduction to Hypothesis Testing
Individual values of X Frequency How many individuals   Distribution of a population.
S519: Evaluation of Information Systems Social Statistics Inferential Statistics Chapter 8: Significantly significant.
Significance testing and confidence intervals Col Naila Azam.
No criminal on the run The concept of test of significance FETP India.
Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove.
Chapter 8 Introduction to Hypothesis Testing ©. Chapter 8 - Chapter Outcomes After studying the material in this chapter, you should be able to: 4 Formulate.
1 Chapter 8 Introduction to Hypothesis Testing. 2 Name of the game… Hypothesis testing Statistical method that uses sample data to evaluate a hypothesis.
Issues concerning the interpretation of statistical significance tests.
Module 15: Hypothesis Testing This modules discusses the concepts of hypothesis testing, including α-level, p-values, and statistical power. Reviewed.
Copyright © 2011 Pearson Education, Inc. Putting Statistics to Work.
9.3/9.4 Hypothesis tests concerning a population mean when  is known- Goals Be able to state the test statistic. Be able to define, interpret and calculate.
Inferential Statistics Inferential statistics allow us to infer the characteristic(s) of a population from sample data Slightly different terms and symbols.
Introduction to Testing a Hypothesis Testing a treatment Descriptive statistics cannot determine if differences are due to chance. Sampling error means.
Descriptive and Inferential Statistics Descriptive statistics The science of describing distributions of samples or populations Inferential statistics.
Chapter 13 Understanding research results: statistical inference.
Presented by Mo Geraghty and Danny Tran June 10, 2016.
Critical Appraisal Course for Emergency Medicine Trainees Module 2 Statistics.
Learning Objectives Describe the hypothesis testing process Distinguish the types of hypotheses Explain hypothesis testing errors Solve hypothesis testing.
15 Inferential Statistics.
Hypothesis Testing I The One-sample Case
Dr.MUSTAQUE AHMED MBBS,MD(COMMUNITY MEDICINE), FELLOWSHIP IN HIV/AIDS
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
Statistical inference: distribution, hypothesis testing
CHAPTER 9 Testing a Claim
Hypothesis Testing: Hypotheses
Inferential Statistics:
Week 11 Chapter 17. Testing Hypotheses about Proportions
CHAPTER 9 Testing a Claim
Statistical Inference
CHAPTER 9 Testing a Claim
CHAPTER 12 Inference for Proportions
CHAPTER 12 Inference for Proportions
Interpreting Epidemiologic Results.
CHAPTER 9 Testing a Claim
Testing Hypotheses I Lesson 9.
CHAPTER 9 Testing a Claim
Section 11.1: Significance Tests: Basics
Type I and Type II Errors
Statistical Test A test of significance is a formal procedure for comparing observed data with a claim (also called a hypothesis) whose truth we want to.
Rest of lecture 4 (Chapter 5: pg ) Statistical Inferences
CHAPTER 9 Testing a Claim
STA 291 Spring 2008 Lecture 17 Dustin Lueker.
Presentation transcript:

Null Hypothesis Testing Nickerson (2000) Psych Methods

Common misconceptions p(D|H) = p(H|D) Bayes Theorem shows how they are actually related. We must consider the prior probability of the hypothesis. What is the probability that someone has a disease given that they have this test result p(H|D) vs. probability of a test result given they have this disease p(D|H). If the disease is rare, chances are good they do not have it even tho the result is positive. https://www.ted.com/talks/peter_donnelly_shows_how_stats_fool_juries

Picture of the problem Zika in Florida If lab score > 80, test for Zika is positive. Test is true positive 90 pct, false positive 10 pct. Do Not have Zika Have Zika p(positive|Zika) = .90, P(Zika|positive) < .50 If we slide the cutoff score down toward 75, p(pos|Zika) will approach 1, and p(Zika|positive) will diminsh.

Nickerson 1 Belief that p is the probability that the null is true and 1-p is the probability that the alternative is true P is the likelihood, the probability of the data given that the null is true. This is Bayes Theorem.

Nickerson 2 Belief that rejection of the null establishes the truth of the theory that predicts it. Predict that the sun will (not) rise Affirming the consequent – typically many things can result in the same outcome.

Nickerson 3 Small p means results are replicable What does replication mean? Often ambiguous. Significant? Power – often small studies are chosen for publication because of significant results. Unless the population effect size is large, further significant results are unlikely Howell 8.7. Aronson et al (1998) stereotype threat study had significant result. But power calculations show power to be .52. About 50/50 chance to find sig result in a replication.

Nickerson 4 Confusion between statistical and practical (or clinical) significance Small effect size can show small p value of sample size is large. Present effect sizes along with your significance tests.

Nickerson 5 Belief that alpha is the probability that if one has rejected the null, one has made a Type I error. This is only true if the result is nil (ES =0) in the population. Very unlikely unless random numbers. ‘Nature abhors a vacuum.’ Google attributes this quote to Aristotle.

Nickerson 6 Belief that failing to reject the null hypothesis is equivalent to demonstrating it to be true. This is a common misunderstanding. Vote counting in meta-analysis. The kernel of truth is that sometimes you have to make a decision, e.g., which strain to plant, whether to implement a new training program. One good reason to avoid accepting the null is that it is really easy to get nonsignificant results. Use a really small sample, choose poor measures, provide a weak treatment. Alpha makes it hard to get a good result by incompetence; beta makes it easy.

Nickerson 7 All or none decisions; arbitrary boundary There is a difference between a best estimate and a decision to be made on the basis of the data. Provide a point estimate. But also provide a confidence interval to communicate the uncertainty that remains after data collection.

Nickerson 8 Use non nil hypotheses Be specific about alternatives Range null Report p values Report precision of measurement Report confidence intervals Report effect sizes. Two main questions: What is the point prediction or estimate What is the magnitude of effect

ASA statement What is a p-value? Informally, a p-value is the probability under a specified statistical model that a statistical summary of the data (for example, the sample mean difference between two compared groups) would be equal to or more extreme than its observed value.

1. P-values can indicate how incompatible the data are with a specified statistical model. 2. P-values do not measure the probability that the studied hypothesis is true, or the probability that the data were produced by random chance alone.

3. Scientific conclusions and business or policy decisions should not be based only on whether a p-value passes a specific threshold. 4. Proper inference requires full reporting and transparency

5. A p-value, or statistical significance, does not measure the size of an effect or the importance of a result. 6. By itself, a p-value does not provide a good measure of evidence regarding a model or hypothesis. No single index should substitute for scientific reasoning.