Chapter 17 Comparing Two Proportions

Slides:



Advertisements
Similar presentations
January Structure of the book Section 1 (Ch 1 – 10) Basic concepts and techniques Section 2 (Ch 11 – 15): Inference for quantitative outcomes Section.
Advertisements

Comparing Two Proportions (p1 vs. p2)
Lecture 3 Outline: Thurs, Sept 11 Chapters Probability model for 2-group randomized experiment Randomization test p-value Probability model for.
Inference for Regression
Chap 9: Testing Hypotheses & Assessing Goodness of Fit Section 9.1: INTRODUCTION In section 8.2, we fitted a Poisson dist’n to counts. This chapter will.
Categorical Data. To identify any association between two categorical data. Example: 1,073 subjects of both genders were recruited for a study where the.
Copyright (c) Bani K. Mallick1 STAT 651 Lecture #17.
HS 1678: Comparing Two Means1 Two Independent Means Unit 8.
HS 167Basics of Hypothesis Testing1 (a)Review of Inferential Basics (b)Hypothesis Testing Procedure (c)One-Sample z Test (σ known) (d)One-sample t test.
Chapter 17 Comparing Two Proportions
Lecture 5 Outline – Tues., Jan. 27 Miscellanea from Lecture 4 Case Study Chapter 2.2 –Probability model for random sampling (see also chapter 1.4.1)
Lecture 5 Outline: Thu, Sept 18 Announcement: No office hours on Tuesday, Sept. 23rd after class. Extra office hour: Tuesday, Sept. 23rd from 12-1 p.m.
Chi-square Test of Independence
7/2/2015Basics of Significance Testing1 Chapter 15 Tests of Significance: The Basics.
Chapter 9 Hypothesis Testing.
5-3 Inference on the Means of Two Populations, Variances Unknown
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
Inference about Population Parameters: Hypothesis Testing
Richard M. Jacobs, OSA, Ph.D.
Lecture 9: p-value functions and intro to Bayesian thinking Matthew Fox Advanced Epidemiology.
1 Chapter 20 Two Categorical Variables: The Chi-Square Test.
The Chi-Square Test Used when both outcome and exposure variables are binary (dichotomous) or even multichotomous Allows the researcher to calculate a.
How Can We Test whether Categorical Variables are Independent?
Inference for regression - Simple linear regression
Chapter 13: Inference in Regression
Overview of Statistical Hypothesis Testing: The z-Test
Fundamentals of Hypothesis Testing: One-Sample Tests
Inference for proportions - Comparing 2 proportions IPS chapter 8.2 © 2006 W.H. Freeman and Company.
September 15. In Chapter 18: 18.1 Types of Samples 18.2 Naturalistic and Cohort Samples 18.3 Chi-Square Test of Association 18.4 Test for Trend 18.5 Case-Control.
September In Chapter 14: 14.1 Data 14.2 Scatterplots 14.3 Correlation 14.4 Regression.
More About Significance Tests
+ Chapter 9 Summary. + Section 9.1 Significance Tests: The Basics After this section, you should be able to… STATE correct hypotheses for a significance.
8.1 Inference for a Single Proportion
September 15. In Chapter 11: 11.1 Estimated Standard Error of the Mean 11.2 Student’s t Distribution 11.3 One-Sample t Test 11.4 Confidence Interval for.
Evidence-Based Medicine 3 More Knowledge and Skills for Critical Reading Karen E. Schetzina, MD, MPH.
Confidence Intervals Nancy D. Barker, M.S.. Statistical Inference.
October 15. In Chapter 9: 9.1 Null and Alternative Hypotheses 9.2 Test Statistic 9.3 P-Value 9.4 Significance Level 9.5 One-Sample z Test 9.6 Power and.
Biostatistics Class 6 Hypothesis Testing: One-Sample Inference 2/29/2000.
October 15. In Chapter 19: 19.1 Preventing Confounding 19.2 Simpson’s Paradox 19.3 Mantel-Haenszel Methods 19.4 Interaction.
1October In Chapter 17: 17.1 Data 17.2 Risk Difference 17.3 Hypothesis Test 17.4 Risk Ratio 17.5 Systematic Sources of Error 17.6 Power and Sample.
Contingency tables Brian Healy, PhD. Types of analysis-independent samples OutcomeExplanatoryAnalysis ContinuousDichotomous t-test, Wilcoxon test ContinuousCategorical.
The binomial applied: absolute and relative risks, chi-square.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 10 Comparing Two Groups Section 10.4 Analyzing Dependent Samples.
November 15. In Chapter 12: 12.1 Paired and Independent Samples 12.2 Exploratory and Descriptive Statistics 12.3 Inference About the Mean Difference 12.4.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Fundamentals of Hypothesis Testing: One-Sample Tests Statistics.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
BPS - 5th Ed. Chapter 221 Two Categorical Variables: The Chi-Square Test.
Essential Statistics Chapter 161 Review Part III_A_Chi Z-procedure Vs t-procedure.
Issues concerning the interpretation of statistical significance tests.
METHODS IN BEHAVIORAL RESEARCH NINTH EDITION PAUL C. COZBY Copyright © 2007 The McGraw-Hill Companies, Inc.
EGR 252 S10 JMB Ch.10 Part 3 Slide 1 Statistical Hypothesis Testing - Part 3  A statistical hypothesis is an assertion concerning one or more populations.
1 Chapter 11: Analyzing the Association Between Categorical Variables Section 11.1: What is Independence and What is Association?
Chap 8-1 Fundamentals of Hypothesis Testing: One-Sample Tests.
Chapter 10 The t Test for Two Independent Samples
CHAPTER 27: One-Way Analysis of Variance: Comparing Several Means
1 G Lect 7a G Lecture 7a Comparing proportions from independent samples Analysis of matched samples Small samples and 2  2 Tables Strength.
© Copyright McGraw-Hill 2004
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 10 Comparing Two Groups Section 10.1 Categorical Response: Comparing Two Proportions.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Lab 8: Types of Studies and Study Designs Lab Workbook (pp. 37 – 40)
A short introduction to epidemiology Chapter 6: Precision Neil Pearce Centre for Public Health Research Massey University Wellington, New Zealand.
Hypothesis Tests u Structure of hypothesis tests 1. choose the appropriate test »based on: data characteristics, study objectives »parametric or nonparametric.
Chapter 10: The t Test For Two Independent Samples.
The binomial applied: absolute and relative risks, chi-square
Week 11 Chapter 17. Testing Hypotheses about Proportions
Chapter 18 Cross-Tabulated Counts
Risk ratios 12/6/ : Risk Ratios 12/6/2018 Risk ratios StatPrimer.
Intro to Confidence Intervals Introduction to Inference
Chapter 18 Part C: Matched Pairs
Presentation transcript:

Chapter 17 Comparing Two Proportions 4/17/2017 Chapter 17 Comparing Two Proportions April 17 Basic Biostat

In Chapter 17: 17.1 Data 17.2 Proportion Difference (Risk Difference) 17.3 Hypothesis Test 17.4 Proportion Ratio (Risk Ratio) 17.5 Systematic Sources of Error 17.6 Power and Sample Size

§17.1 Data Two independent groups Binary response Epidemiologic jargon: Group 1 = “exposed group” and Group 2 = “nonexposed group” Count “successes” in each group and convert to proportions

Sample Proportions Proportion in the exposed group: Proportion in the nonexposed group:

§17.2 Proportion Difference (Risk Difference) The risk difference is the absolute difference in incidence proportions in the groups.

In large samples, the sampling distribution of the risk difference is approximately Normal

Confidence Interval, Risk Difference Plus-four confidence interval method for a difference in proportions. This method is accurate in samples as small as 5 per group.

95% CI for p1 – p2, Example WHI data a1 = 751, n1 = 8503, a2 = 623, n2 = 8102

95% CI for p1 – p2, Example The plus-four method is similar to Wilson’s score method. Here’s output from from WinPepi > Compare2.exe > Program B showing results from the traditional large-sample method and Wilson score CI for the illustrative example.

§17.3 Hypothesis Test We test the proportions for a significant (“nonrandom”) difference Two methods are covered in this chapter z test (large sample) Fisher’s exact procedure (small samples) A third method called the chi-square test is covered in the next chapter

z Test A. Hypotheses. H0: p1 = p2 against Ha:p1 ≠ p2 [One-sided: Ha: p1 > p2 or Ha: p1 < p2] B. Test statistic. C. P-value. Convert zstat to P-value [Table B or F]

conclude: highly significant

z Test: Notes z statistic dissection numerator is observed difference denominator is standard error when p1 = p2 A continuity correction can be optionally applied (p. 382) This z test is equivalent to the chi-square test of association (Chapter 18) In small samples (fewer than 5 successes expected in either group), avoid the z test and use the exact Fisher or Mid-P procedure

Fisher’s Exact Test (2-by-2) Before conducting Fisher’s test, data are rearranged to form a 2-by-2 table : Successes Failures Total Group 1 a1 b1 n1 Group 2 a2 b2 n2 m1 m2 N Recall that and

WHI Data, 2-by-2 Format + − Total Estrogen + 751 7755 8506 Estrogen − 623 7479 8102 1374 15234 16608

Fisher’s Exact Test, Procedure A. Hypotheses. H0: p1 = p2 vs. Ha: p1 ≠ p2 [one sided Ha: p1 > p2 or Ha: p1 < p2] B. Test statistic. Observed counts in 2-by-2 table C. P-value. Use computer program (e.g., WinPepi > Compare2.exe > Program B). The mathematical basis of the test is described on pp. 386–7.

Fisher’s Exact Test, Example The incidence of colonic necrosis in an exposed group is 2 of 117. The incidence in a non-exposed group is 0 of 862. Is this difference statistically significant? A. H0: p1 = p2 against Ha: p1 ≠ p2 B. Data. + − Total Group 1 2 115 117 Group 2 862 977 979

C. P = 0. 014 (WinPepi output shown here) C. P = 0.014 (WinPepi output shown here). The evidence against H0 is “significant.”

§17.4 Proportion (Risk) Ratio Let RR refer to an risk ratio or prevalence ratios Interpretation The RR is a risk multiplier, e.g., an RR of 2 suggests that the exposure doubles risk When p1 = p2 , RR = 1. This is the “baseline RR,” indicating no association.

RR Example, WHI Data + − Total Estrogen + 751 7755 8506 Estrogen − 623 7479 8102 The indicates a positive association; specifically, 15% higher risk (in relative terms) with exposure.

Note natural log scale of sampling distribution (1– α)100% CI for the RR Note natural log scale of sampling distribution

90% CI for RR, WHI Example + − Total Estrogen + 751 7755 8506 623 7479 8102

CI for RR, Computerized Results + − Total Estrogen + 751 7755 8506 Estrogen − 623 7479 8102 Output from WinPepi > Compare2.exe > Program B. See prior slide for hand calculations

§17.5 Systematic Error (Advanced Topic) In observational studies, systematic errors are often more important than random sampling error Three types of systematic error are considered: Confounding Information bias Selection bias

Confounding Confounding = the mixing together of the effects of the explanatory variable with the effects of “lurking” variables. Consider this example: The WHI estrogen experiment found increased morbidity and mortality in estrogen users Earlier, non-experimental studies found the opposite: lower morbidity and mortality in users Plausible explanation: In non-experimental studies, estrogen users (self-selected) were more likely to have “lurking” lifestyles factors that contributed to better health, i.e., confounding

Information Bias Information bias is due to the mismeasurement or misclassification of variables in the study. Misclassification may be nondifferential (occurs to the same extent in the groups) or differential (one groups experiences a greater degree of misclassification than the other) Nondifferential misclassification tends to bias results toward the null (or have no effect). Differential misclassification can bias results in either direction.

Nondifferential & Differential Misclassification - Examples

Selection Bias Selection bias ≡ systematic error related to the manner in which study participants are selected for study Example. If we shoot an arrow into the broad side of a barn and later draw a bull’s-eye where it had landed, have we really identified anything worth noting?

17.6 Power and Sample Size Power and sample sizes analysis for comparing proportions requires us to understand relationships between these factors: r ≡ sample size allocation ratio n1 / n2 1−β ≡ power (type II error) α ≡ significance level (type I error) p1 ≡ expected proportion, group 1 p2 ≡ expected proportion in group 2, or some measure of effect size, such as the expected RR

17.6 Power and Sample Size Because of the complexity of calculations (pp. 396 – 402), use software… Here’s WinPepi’s Compare2 Sample size menu.