Introduction to sample size and power calculations How much chance do we have to reject the null hypothesis when the alternative is in fact true? (what’s.

Slides:



Advertisements
Similar presentations
Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
Advertisements

Chapter 12: Testing hypotheses about single means (z and t) Example: Suppose you have the hypothesis that UW undergrads have higher than the average IQ.
Testing a Claim about a Proportion Assumptions 1.The sample was a simple random sample 2.The conditions for a binomial distribution are satisfied 3.Both.
INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE
1. Estimation ESTIMATION.
The Two Sample t Review significance testing Review t distribution
Review: What influences confidence intervals?
Fundamentals of Hypothesis Testing. Identify the Population Assume the population mean TV sets is 3. (Null Hypothesis) REJECT Compute the Sample Mean.
Sample size computations Petter Mostad
Cal State Northridge  320 Ainsworth Sampling Distributions and Hypothesis Testing.
Chapter 3 Normal Curve, Probability, and Population Versus Sample Part 2.
Statistics for the Social Sciences Psychology 340 Fall 2006 Review For Exam 1.
Chapter Sampling Distributions and Hypothesis Testing.
T-Tests Lecture: Nov. 6, 2002.
Independent Sample T-test Often used with experimental designs N subjects are randomly assigned to two groups (Control * Treatment). After treatment, the.
PY 427 Statistics 1Fall 2006 Kin Ching Kong, Ph.D Lecture 5 Chicago School of Professional Psychology.
Chapter 9 Hypothesis Testing.
Chapter 9 Hypothesis Testing II. Chapter Outline  Introduction  Hypothesis Testing with Sample Means (Large Samples)  Hypothesis Testing with Sample.
Today Concepts underlying inferential statistics
Calculating sample size for a case-control study
Probability Population:
Chapter 9 Hypothesis Testing II. Chapter Outline  Introduction  Hypothesis Testing with Sample Means (Large Samples)  Hypothesis Testing with Sample.
Statistical Inference I: Hypothesis testing; sample size
Inferential Statistics
Statistical Analysis. Purpose of Statistical Analysis Determines whether the results found in an experiment are meaningful. Answers the question: –Does.
Hypothesis Testing and T-Tests. Hypothesis Tests Related to Differences Copyright © 2009 Pearson Education, Inc. Chapter Tests of Differences One.
INFERENTIAL STATISTICS – Samples are only estimates of the population – Sample statistics will be slightly off from the true values of its population’s.
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
Jeopardy Hypothesis Testing T-test Basics T for Indep. Samples Z-scores Probability $100 $200$200 $300 $500 $400 $300 $400 $300 $400 $500 $400.
Hypothesis Testing II The Two-Sample Case.
Section 10.1 ~ t Distribution for Inferences about a Mean Introduction to Probability and Statistics Ms. Young.
Single Sample Inferences
1 Power and Sample Size in Testing One Mean. 2 Type I & Type II Error Type I Error: reject the null hypothesis when it is true. The probability of a Type.
Chapter 8 Hypothesis Testing I. Chapter Outline  An Overview of Hypothesis Testing  The Five-Step Model for Hypothesis Testing  One-Tailed and Two-Tailed.
Hypothesis Testing: One Sample Cases. Outline: – The logic of hypothesis testing – The Five-Step Model – Hypothesis testing for single sample means (z.
Chapter 9 Hypothesis Testing II: two samples Test of significance for sample means (large samples) The difference between “statistical significance” and.
Copyright © 2012 by Nelson Education Limited. Chapter 7 Hypothesis Testing I: The One-Sample Case 7-1.
Statistics and Research methods Wiskunde voor HMI Bijeenkomst 3 Relating statistics and experimental design.
Individual values of X Frequency How many individuals   Distribution of a population.
6.1 - One Sample One Sample  Mean μ, Variance σ 2, Proportion π Two Samples Two Samples  Means, Variances, Proportions μ 1 vs. μ 2.
Stats 95.
Psy B07 Chapter 4Slide 1 SAMPLING DISTRIBUTIONS AND HYPOTHESIS TESTING.
1 Lecture note 4 Hypothesis Testing Significant Difference ©
Introduction to Inferential Statistics Statistical analyses are initially divided into: Descriptive Statistics or Inferential Statistics. Descriptive Statistics.
DIRECTIONAL HYPOTHESIS The 1-tailed test: –Instead of dividing alpha by 2, you are looking for unlikely outcomes on only 1 side of the distribution –No.
Confidence Intervals Lecture 3. Confidence Intervals for the Population Mean (or percentage) For studies with large samples, “approximately 95% of the.
Chapter 9 Introduction to the t Statistic. 9.1 Review Hypothesis Testing with z-Scores Sample mean (M) estimates (& approximates) population mean (μ)
Testing Differences between Means, continued Statistics for Political Science Levin and Fox Chapter Seven.
26134 Business Statistics Tutorial 11: Hypothesis Testing Introduction: Key concepts in this tutorial are listed below 1. Difference.
Simple examples of the Bayesian approach For proportions and means.
Hypothesis Testing and the T Test. First: Lets Remember Z Scores So: you received a 75 on a test. How did you do? If I said the mean was 72 what do you.
Statistical Inference I: Hypothesis testing; sample size.
Hypothesis Testing Introduction to Statistics Chapter 8 Feb 24-26, 2009 Classes #12-13.
AGENDA Review In-Class Group Problems Review. Homework #3 Due on Thursday Do the first problem correctly Difference between what should happen over the.
Sampling Distribution (a.k.a. “Distribution of Sample Outcomes”) – Based on the laws of probability – “OUTCOMES” = proportions, means, test statistics.
Psych 230 Psychological Measurement and Statistics Pedro Wolf October 21, 2009.
Hypothesis test flow chart
Sample Size Needed to Achieve High Confidence (Means)
Hypothesis Testing and Statistical Significance
Copyright © 2009 Pearson Education, Inc t LEARNING GOAL Understand when it is appropriate to use the Student t distribution rather than the normal.
Chapter 9 Introduction to the t Statistic
The ability to find a difference when one really exists.
AP Biology Intro to Statistics
CHAPTER 6 Statistical Inference & Hypothesis Testing
Hypothesis Testing Review
Inferences on Two Samples Summary
Hypothesis Tests for Proportions
Hypothesis Testing.
Elements of a statistical test Statistical null hypotheses
Power Section 9.7.
Presentation transcript:

Introduction to sample size and power calculations How much chance do we have to reject the null hypothesis when the alternative is in fact true? (what’s the probability of detecting a real effect?)

Can we quantify how much power we have for given sample sizes?

Null Distribution: difference=0. Clinically relevant alternative: difference=10%. Rejection region. Any value >= 6.5 (0+3.3*1.96) study 1: 263 cases, 1241 controls For 5% significance level, one-tail area=2.5% (Z  /2 = 1.96) Power= chance of being in the rejection region if the alternative is true=area to the right of this line (in yellow) 

Rejection region. Any value >= 6.5 (0+3.3*1.96) Power= chance of being in the rejection region if the alternative is true=area to the right of this line (in yellow) study 1: 263 cases, 1241 controls Power here:

Critical value= 0+10*1.96=20 Power closer to 15% now. 2.5% area Z  /2 =1.96 study 1: 50 cases, 50 controls

Critical value= *1.96 = 1 Power is nearly 100%! Study 2: 18 treated, 72 controls, STD DEV = 2 Clinically relevant alternative: difference=4 points

Critical value= *1.96 = 5 Power is about 40% Study 2: 18 treated, 72 controls, STD DEV=10

Critical value= *1.96 = 1 Power is about 50% Study 2: 18 treated, 72 controls, effect size=1.0 Clinically relevant alternative: difference=1 point

Factors Affecting Power 1. Size of the effect 2. Standard deviation of the characteristic 3. Bigger sample size 4. Significance level desired

average weight from samples of 100 Null Clinically relevant alternative 1. Bigger difference from the null mean

average weight from samples of Bigger standard deviation

average weight from samples of Bigger Sample Size

average weight from samples of Higher significance level Rejection region.

Sample size calculations Based on these elements, you can write a formal mathematical equation that relates power, sample size, effect size, standard deviation, and significance level… **WE WILL DERIVE THESE FORMULAS FORMALLY SHORTLY**

Simple formula for difference in means Sample size in each group (assumes equal sized groups) Represents the desired power (typically.84 for 80% power). Represents the desired level of statistical significance (typically 1.96). Standard deviation of the outcome variable Effect Size (the difference in means)

Simple formula for difference in proportions Sample size in each group (assumes equal sized groups) Represents the desired power (typically.84 for 80% power). Represents the desired level of statistical significance (typically 1.96). A measure of variability (similar to standard deviation) Effect Size (the difference in proportions)

Derivation of sample size formula….

Critical value= 0+.52*1.96=1 Power close to 50% Study 2: 18 treated, 72 controls, effect size=1.0

SAMPLE SIZE AND POWER FORMULAS Critical value= 0+standard error (difference)*Z  /2 Power= area to right of Z  =

Power is the area to the right of Z . OR power is the area to the left of - Z . Since normal charts give us the area to the left by convention, we need to use - Z  to get the correct value. Most textbooks just call this “ Z  ”; I’ll use the term Z power to avoid confusion.

All-purpose power formula…

Derivation of a sample size formula… Sample size is embedded in the standard error….

Algebra…

Sample size formula for difference in means

Examples Example 1: You want to calculate how much power you will have to see a difference of 3.0 IQ points between two groups: 30 male doctors and 30 female doctors. If you expect the standard deviation to be about 10 on an IQ test for both groups, then the standard error for the difference will be about: = 2.57

Power formula… P(Z ≤ -.79) =.21; only 21% power to see a difference of 3 IQ points.

Example 2: How many people would you need to sample in each group to achieve power of 80% (corresponds to Z  =.84) 174/group; 348 altogether

Sample Size needed for comparing two proportions: Example: I am going to run a case-control study to determine if pancreatic cancer is linked to drinking coffee. If I want 80% power to detect a 10% difference in the proportion of coffee drinkers among cases vs. controls (if coffee drinking and pancreatic cancer are linked, we would expect that a higher proportion of cases would be coffee drinkers than controls), how many cases and controls should I sample? About half the population drinks coffee.

Derivation of a sample size formula: The standard error of the difference of two proportions is:

Here, if we assume equal sample size and that, under the null hypothesis proportions of coffee drinkers is.5 in both cases and controls, then s.e.(diff)= Derivation of a sample size formula:

For 80% power… There is 80% area to the left of a Z-score of.84 on a standard normal curve; therefore, there is 80% area to the right of Would take 392 cases and 392 controls to have 80% power! Total=784

Question 2: How many total cases and controls would I have to sample to get 80% power for the same study, if I sample 2 controls for every case? Ask yourself, what changes here?

Different size groups… Need: 294 cases and 2x294=588 controls. 882 total. Note: you get the best power for the lowest sample size if you keep both groups equal (882 > 784). You would only want to make groups unequal if there was an obvious difference in the cost or ease of collecting data on one group. E.g., cases of pancreatic cancer are rare and take time to find.

General sample size formula

General sample size needs when outcome is binary:

Compare with when outcome is continuous:

Question How many subjects would we need to sample to have 80% power to detect an average increase in MCAT biology score of 1 point, if the average change without instruction (just due to chance) is plus or minus 3 points (=standard deviation of change)?

Standard error here=

Therefore, need: (9)( ) 2 /1 = 70 people total Where D=change from test 1 to test 2. (difference)

Sample size for paired data:

Paired data difference in proportion: sample size: