Sample size calculation

Slides:



Advertisements
Similar presentations
Introduction to Hypothesis Testing
Advertisements

Chapter 10 Introduction to Inference
Significance testing Ioannis Karagiannis (based on previous EPIET material) 18 th EPIET/EUPHEM Introductory course
Significance testing and confidence intervals Ágnes Hajdu EPIET Introductory course
Sample size calculation
Critical review of significance testing F.DAncona from a Alain Morens lecture 2006.
Power of a test. power The power of a test (against a specific alternative value) Is a tests ability to detect a false hypothesis Is the probability that.
Hypothesis Test II: t tests
Hypothesis Testing W&W, Chapter 9.
9.4 t test and u test Hypothesis testing for population mean Example : Hemoglobin of 280 healthy male adults in a region: Question: Whether the population.
Power and sample size.
Study Size Planning for Observational Comparative Effectiveness Research Prepared for: Agency for Healthcare Research and Quality (AHRQ)
Type I & Type II errors Brian Yuen 18 June 2013.
Type I and Type II Errors. Ms. Betts Chapter 9 Quiz.
Pitfalls of Hypothesis Testing + Sample Size Calculations.
Recap of confidence intervals If the 95%CI does not include the hypothesized μ, we conclude that our sample is statistically different from the assumed.
Sample size estimation
+ Chapter 10 Section 10.4 Part 2 – Inference as Decision.
Hypothesis Testing Steps of a Statistical Significance Test. 1. Assumptions Type of data, form of population, method of sampling, sample size.
Basic Elements of Testing Hypothesis Dr. M. H. Rahbar Professor of Biostatistics Department of Epidemiology Director, Data Coordinating Center College.
Sample Size and Statistical Power Epidemiology 655 Winter 1999 Jennifer Beebe.
Inferential Statistics
Sample size calculation
Hadpop Calculations. Odds ratio What study applicable? Q. It is suggested that obesity increases the chances on an individual becoming infected with erysipelas.
Warm-up Day of 8.1 and 8.2 Quiz and Types of Errors Notes.
Hypothesis Testing.
Academic Viva POWER and ERROR T R Wilson. Impact Factor Measure reflecting the average number of citations to recent articles published in that journal.
Chapter 7 Statistical Issues in Research Planning and Evaluation.
1 Power and Sample Size in Testing One Mean. 2 Type I & Type II Error Type I Error: reject the null hypothesis when it is true. The probability of a Type.
Instructor Resource Chapter 5 Copyright © Scott B. Patten, Permission granted for classroom use with Epidemiology for Canadian Students: Principles,
Warm-up Day of 8.1 and 8.2 Review. 8.2 P#20, 23 and 24 P#20 a. and b. c. Since the p-hat is along the line for reasonably likely events.
Hypothesis testing and Decision Making Formal aspects of hypothesis testing.
1 ConceptsDescriptionHypothesis TheoryLawsModel organizesurprise validate formalize The Scientific Method.
Inferential Statistics Body of statistical computations relevant to making inferences from findings based on sample observations to some larger population.
Ch 10 – Intro To Inference 10.1: Estimating with Confidence 10.2 Tests of Significance 10.3 Making Sense of Statistical Significance 10.4 Inference as.
Using Inference to MAKE DECISIONS The Type I and Type II Errors in Hypothesis Testing.
Chapter 9: Power I can make really good decisions. Chapter 9: Power Target Goal: I can make really good decisions. 9.1d h.w: pg 548: 23, 25.
MeanVariance Sample Population Size n N IME 301. b = is a random value = is probability means For example: IME 301 Also: For example means Then from standard.
Introduction to hypothesis testing Hypothesis testing is about making decisions Is a hypothesis true or false? Ex. Are women paid less, on average, than.
Hypothesis Testing and Statistical Significance
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
More about tests and intervals CHAPTER 21. Do not state your claim as the null hypothesis, instead make what you’re trying to prove the alternative. The.
Critical Appraisal Course for Emergency Medicine Trainees Module 2 Statistics.
+ Homework 9.1:1-8, 21 & 22 Reading Guide 9.2 Section 9.1 Significance Tests: The Basics.
Sample size calculation
CHAPTER 9 Testing a Claim
Significance testing and confidence intervals
Chapter 6  PROBABILITY AND HYPOTHESIS TESTING
Power of a test.
Hypothesis Testing.
Random error, Confidence intervals and P-values
When we free ourselves of desire,
Chapter 9: Hypothesis Testing
AP Statistics: Chapter 21
Section 10.3 Making Sense of Statistical Significance
CHAPTER 12 Inference for Proportions
CHAPTER 12 Inference for Proportions
More About Tests Notes from
Chapter 7: Statistical Issues in Research planning and Evaluation
Power of a Hypothesis Test
Power of a test.
CHAPTER 9 Testing a Claim
CHAPTER 9 Testing a Claim
CHAPTER 9 Testing a Claim
Inference as Decision Section 10.4.
Power and Error What is it?.
AP STATISTICS LESSON 10 – 4 (DAY 2)
Type I and Type II Errors
Statistical Power.
Inference about Population Mean
Presentation transcript:

Sample size calculation Lecture on measures of disease occurence Ioannis Karagiannis based on previous EPIET material

Objectives: sample size To understand: Why we estimate sample size Principles of sample size calculation Ingredients needed to estimate sample size

Why estimate sample size? How many people are needed to answer your study objectives? Pointless if power is too small; Waste of resources if sample size needed is too large.

How large should your sample size be? A national Salmonella outbreak has occurred with several hundred cases; You plan a case-control study to identify if consumption of food X is associated with infection; How many cases and controls should you recruit?

How much power do you have? An outbreak of 14 cases of a mysterious disease has occurred in cohort 17; You suspect exposure to food item Y is associated with illness and plan to undertake a cohort study under the kind auspices of coordinators; With the available cases, how much power will you have to detect a RR of 1.5?

Issues in sample size estimation Estimate sample needed to measure the factor of interest Trade-off between study size and resources Sample size determined by various factors: significance level (α) power (1-β) expected prevalence of factor of interest

Significance testing: null and alternative hypotheses Null hypothesis (H0) There is no difference Any difference is due to chance Alternative hypothesis (H1) There is a true difference

Examples of null hypotheses Case-control study H0: OR=1 “the odds of exposure among cases are the same as the odds of exposure among controls” Cohort study H0: RR=1 “the AR among the exposed is the same as the AR among the unexposed”

Significance and power Truth H0 true No difference H0 false Difference Decision Cannot reject H0 Correct decision Type II error = β Reject H0 Type I error level = α significance power = 1-β

Significance level (p-value) probability of finding a difference (RR≠1, reject H0), when no difference exists; α or type I error; usually set at 5%; p-value used to reject H0 (significance level);  NB: a hypothesis is never “accepted”

Type II error and power β is the type II error probability of not finding a difference, when a difference really does exist Power is (1-β) and is usually set to 80% probability of finding a difference when a difference really does exist (=sensitivity)

Power Easiest ways to increase power are to: increase sample size increase desired difference (or effect size) required  NB: increasing the desired difference in RR/OR means move it away from 1! increase significance level desired (α error) to e.g. 10%  Less wide confidence intervals

The effect of sample size Consider 3 cohort studies looking at exposure to oysters with N=10, 100, 1000 In all 3 studies, 60% of the exposed are ill compared to 40% of unexposed (RR = 1.5)

Table A (N=10) Became ill Yes Total AR Ate oysters 3 5 3/5 No 2 2/5 10 5/10 RR=1.5, 95% CI: 0.4-5.4, p=0.53

Table B (N=100) Became ill Yes Total AR Ate oysters 30 50 30/50 No 20 20/50 100 50/100 RR=1.5, 95% CI: 1.0-2.3, p=0.046

Table C (N=1000) Became ill Yes No AR Ate oysters 300 500 300/500 200 200/500 Total 1000 500/1000 RR=1.5, 95% CI: 1.3-1.7, p<0.001

Sample size and power In Table A, with n=10 sample, there was no significant association with oysters, but there was with a larger sample size. This illustrates the relationship between power and sample size

Steps in estimating sample size for descriptive survey Identify major study variable Determine type of estimate (%, mean, ratio,...) Indicate expected frequency of factor of interest Literature, expert opinion Decide on desired precision of the estimate

Cohort sample size: issues to consider Risk ratio worth detecting Expected frequency of disease in unexposed population Ratio of unexposed to exposed Desired level of significance (α) Power of the study (1-β)

Cohort: Episheet Power calculation Risk of α error 5% Population exposed 100 Exp freq disease in unexposed 5% Ratio of unexposed to exposed 1:1 RR to detect ≥1.5 The power of a statistical test is the probability that the test will reject a false null hypothesis, or in other words that it will not make a Type II error. The higher the power, the greater the chance of obtaining a statistically significant result when the null hypothesis is false. Statistical power depends on the significance criterion, the size of the difference or the strength of the similarity (that is, the effect size) in the population, and the sensitivity of the data.

Case-control sample size: issues to consider Number of cases Number of controls per case OR ratio worth detecting % of exposed persons in source population Desired level of significance (α) Power of the study (1-β)

Case-control: Power calculation Risk of α error 5% Number of cases 200 Proportion of controls exposed 5% OR to detect ≥1.5 No. controls:case 1:1 The power of a statistical test is the probability that the test will reject a false null hypothesis, or in other words that it will not make a Type II error. The higher the power, the greater the chance of obtaining a statistically significant result when the null hypothesis is false. Statistical power depends on the significance criterion, the size of the difference or the strength of the similarity (that is, the effect size) in the population, and the sensitivity of the data.

Statistical Power of a Case-Control Study for different control-to-case ratios and odds ratios (50 cases) a probability is a number between 0 and 1; the probability of an event or proposition and its complement must add up to 1; and the joint probability of two events or propositions is the product of the probability of one of them and the probability of the second, conditional on the first. Representation and interpretation of probability values The probability of an event is generally represented as a real number between 0 and 1, inclusive. An impossible event has a probability of exactly 0, and a certain event has a probability of 1, but the converses are not always true: probability 0 events are not always impossible, nor probability 1 events certain. The rather subtle distinction between "certain" and "probability 1" is treated at greater length in the article on "almost surely". Most probabilities that occur in practice are numbers between 0 and 1, indicating the event's position on the continuum between impossibility and certainty. The closer an event's probability is to 1, the more likely it is to occur. For example, if two mutually exclusive events are assumed equally probable, such as a flipped coin landing heads-up or tails-up, we can express the probability of each event as "1 in 2", or, equivalently, "50%" or "1/2". Probabilities are equivalently expressed as odds, which is the ratio of the probability of one event to the probability of all other events. The odds of heads-up, for the tossed coin, are (1/2)/(1 - 1/2), which is equal to 1/1. This is expressed as "1 to 1 odds" and often written "1:1". Odds a:b for some event are equivalent to probability a/(a+b). For example, 1:1 odds are equivalent to probability 1/2, and 3:2 odds are equivalent to probability 3/5. There remains the question of exactly what can be assigned probability, and how the numbers so assigned can be used; this is the question of probability interpretations.

Conclusions Don’t forget to undertake sample size/power calculations Use all sources of currently available data to inform your estimates Try several scenarios Adjust for non-response

Acknowledgements Nick Andrews Richard Pebody Viviane Bremer