Fitting Marginal Structural Models Eleanor M Pullenayegum Asst Professor Dept of Clin. Epi & Biostatistics

Slides:



Advertisements
Similar presentations
June 25, 2006 Propensity Score Adjustment in Survival Models Carolyn Rutter Group Health Cooperative AcademyHealth, Seattle WA.
Advertisements

Agency for Healthcare Research and Quality (AHRQ)
Multilevel Event History Modelling of Birth Intervals
Controlling for Time Dependent Confounding Using Marginal Structural Models in the Case of a Continuous Treatment O Wang 1, T McMullan 2 1 Amgen, Thousand.
A workshop introducing doubly robust estimation of treatment effects
Review of Identifying Causal Effects Methods of Economic Investigation Lecture 13.
1 Arlene Ash QMC - Third Tuesday September 21, 2010 (as amended, Sept 23) Analyzing Observational Data: Focus on Propensity Scores.
Study Designs in Epidemiologic
Estimating Direct Effects of New HIV Prevention Methods. Focus: the MIRA Trial UCSF: Helen Cheng, Nancy Padian, Steve Shiboski, Ariane van der Straten.
Observational Designs Oncology Journal Club April 26, 2002.
Informative Censoring Addressing Bias in Effect Estimates Due to Study Drop-out Mark van der Laan and Maya Petersen Division of Biostatistics, University.
PHSSR IG CyberSeminar Introductory Remarks Bryan Dowd Division of Health Policy and Management School of Public Health University of Minnesota.
Extension Article by Dr Tim Kenny
Biostatistics ~ Types of Studies. Research classifications Observational vs. Experimental Observational – researcher collects info on attributes or measurements.
 Confounders are usually controlled with the “standard” response regression model.  The standard model includes confounders as covariates in the response.

Chapter 51 Experiments, Good and Bad. Chapter 52 Experimentation u An experiment is the process of subjecting experimental units to treatments and observing.
Clustered or Multilevel Data
Chapter 11 Survival Analysis Part 2. 2 Survival Analysis and Regression Combine lots of information Combine lots of information Look at several variables.
Adjusting for Time-Varying Confounding in Survival Analysis Jennifer S. Barber Susan A. Murphy Natalya Verbitsky University of Michigan.
N The Experimental procedure involves manipulating something called the Explanatory Variable and seeing the effect on something called the Outcome Variable.
Modeling clustered survival data The different approaches.
Using Regression Models to Analyze Randomized Trials: Asymptotically Valid Tests Despite Incorrect Regression Models Michael Rosenblum, UCSF TAPS Fellow.
Mitchell Hoffman UC Berkeley. Statistics: Making inferences about populations (infinitely large) using finitely large data. Crucial for Addressing Causal.
Foster Care Reunification: The use of hierarchical modeling to account for sibling and county correlation Emily Putnam-Hornstein, MSW Center for Social.
Assessing Survival: Cox Proportional Hazards Model Peter T. Donnan Professor of Epidemiology and Biostatistics Statistics for Health Research.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 12: Multiple and Logistic Regression Marshall University.
Study Design and Analysis in Epidemiology: Where does modeling fit? Meaningful Modeling of Epidemiologic Data, 2010 AIMS, Muizenberg, South Africa Steve.
GEE and Generalized Linear Mixed Models
Multiple Choice Questions for discussion
Advanced Statistics for Interventional Cardiologists.
Essentials of survival analysis How to practice evidence based oncology European School of Oncology July 2004 Antwerp, Belgium Dr. Iztok Hozo Professor.
Dr Laura Bonnett Department of Biostatistics. UNDERSTANDING SURVIVAL ANALYSIS.
Biostatistics Case Studies 2005 Peter D. Christenson Biostatistician Session 4: Taking Risks and Playing the Odds: OR vs.
TWO-STAGE CASE-CONTROL STUDIES USING EXPOSURE ESTIMATES FROM A GEOGRAPHICAL INFORMATION SYSTEM Jonas Björk 1 & Ulf Strömberg 2 1 Competence Center for.
ECON ECON Health Economic Policy Lab Kem P. Krueger, Pharm.D., Ph.D. Anne Alexander, M.S., Ph.D. University of Wyoming.
Lecture 8: Generalized Linear Models for Longitudinal Data.
Assessing Survival: Cox Proportional Hazards Model
October 15H.S.1 Causal inference Hein Stigum Presentation, data and programs at:
Assessing ETA Violations, and Selecting Attainable/Realistic Parameters Causal Effect/Variable Importance Estimation and the Experimental Treatment Assumption.
Andrew Thomson on Generalised Estimating Equations (and simulation studies)
Excepted from HSRP 734: Advanced Statistical Methods June 5, 2008.
Biostatistics Case Studies 2008 Peter D. Christenson Biostatistician Session 5: Choices for Longitudinal Data Analysis.
Data Analysis – Statistical Issues Bernd Genser, PhD Instituto de Saúde Coletiva, Universidade Federal da Bahia, Salvador
Estimating Causal Effects from Large Data Sets Using Propensity Scores Hal V. Barron, MD TICR 5/06.
Introduction to Survival Analysis Utah State University January 28, 2008 Bill Welbourn.
1 1 Using Marginal Structural Model to Estimate and Adjust for Causal Effect of Post- discontinuation Chemotherapy on Survival in Cancer Trials Y. Wang,
HSRP 734: Advanced Statistical Methods July 31, 2008.
Epidemiologic design from a sampling perspective Epidemiology II Lecture April 14, 2005 David Jacobs.
Propensity Score Matching for Causal Inference: Possibilities, Limitations, and an Example sean f. reardon MAPSS colloquium March 6, 2007.
Empirical Efficiency Maximization: Locally Efficient Covariate Adjustment in Randomized Experiments Daniel B. Rubin Joint work with Mark J. van der Laan.
BIOST 536 Lecture 11 1 Lecture 11 – Additional topics in Logistic Regression C-statistic (“concordance statistic”)  Same as Area under the curve (AUC)
Study Designs for Clinical and Epidemiological Research Carla J. Alvarado, MS, CIC University of Wisconsin-Madison (608)
Osteoarthritis Initiative Analytic Strategies for the OAI Data December 6, 2007 Charles E. McCulloch, Division of Biostatistics, Dept of Epidemiology and.
Right Hand Side (Independent) Variables Ciaran S. Phibbs June 6, 2012.
Right Hand Side (Independent) Variables Ciaran S. Phibbs.
Describing the risk of an event and identifying risk factors Caroline Sabin Professor of Medical Statistics and Epidemiology, Research Department of Infection.
11/20091 EPI 5240: Introduction to Epidemiology Confounding: concepts and general approaches November 9, 2009 Dr. N. Birkett, Department of Epidemiology.
1 Basic epidemiological study designs and its role in measuring disease exposure association M. A. Yushuf Sharker Assistant Scientist Center for Communicable.
Assessing the Total Effect of Time-Varying Predictors in Prevention Research Bethany Bray April 7, 2003 University of Michigan, Dearborn.
Instructor Resource Chapter 15 Copyright © Scott B. Patten, Permission granted for classroom use with Epidemiology for Canadian Students: Principles,
Matching. Objectives Discuss methods of matching Discuss advantages and disadvantages of matching Discuss applications of matching Confounding residual.
Types of Studies. Aim of epidemiological studies To determine distribution of disease To examine determinants of a disease To judge whether a given exposure.
1 Study Design Imre Janszky Faculty of Medicine, ISM NTNU.
PH240A Epidemiology and the Curse of Dimensionality Alan Hubbard Division of Biostatistics U.C. Berkeley
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 13: Multiple, Logistic and Proportional Hazards Regression.
STA248 week 121 Bootstrap Test for Pairs of Means of a Non-Normal Population – small samples Suppose X 1, …, X n are iid from some distribution independent.
Survival time treatment effects
The Centre for Longitudinal Studies Missing Data Strategy
Counterfactual models Time dependent confounding
Presentation transcript:

Fitting Marginal Structural Models Eleanor M Pullenayegum Asst Professor Dept of Clin. Epi & Biostatistics

Outline Causality and observational data Inverse-Probability weighting and MSMs Fitting an MSM Goodness-of-fit Assumptions/ Interpretation

Causality in Medical Research Often want to establish a causal association between a treatment/exposure and an event Difficult to do with observational data due to confounding Gold-standard for causal inferences is the randomized trial Randomize half the patients to receive the treatment/exposure, and half to receive usual care Deals with measured and unmeasured confounders

Randomized trials are not always possible Sometimes, they are unethical cannot do a randomized trial on the effects of second-hand smoke on lung cancer or a randomized trial of the effects of living near power stations Sometimes, they are not feasible Study of a rare disease (funding is an issue)

Observational Studies Observe rather than experiment (or interfere!) Recruit some people who are exposed to second- hand smoke and some who are not Study communities living close to power lines vs. those who don’t Confounding is a major concern For 1 st example, are workplace environment, home environment, age, gender, income similar between exposed and unexposed? For 2 nd example, are education, family income, air pollution similar between cases and controls?

Handling Confounding Match exposed and unexposed on key confounders e.g. for every family living close to a power station, attempt to find a control family living in a similar neighbourhood with a similar income Adjust for confounders for the smoking example, adjust for age, gender, level of education, income, type of work, family history of cancer etc. Cannot deal with unmeasured confounders

Causal Pathways There are some things we cannot adjust for When studying the effect of a lipid-lowering drug on heart disease, we can’t adjust for LDL-cholesterol level

Causal Pathways Drug LDL-cholesterolHeart Disease LDL-cholesterol mediates the effect of the drug Cannot adjust for variables that are on the causal pathway between exposure and outcome.

Motivating Example Juvenile Dermatomyositis (JDM) is a rare but serious skin/muscle disease in children Standard treatment is with steroids (Prednisone), however these have unpleasant side-effects Intravenous immunoglobulin (IVIg) is a possible alternative treatment DAS measures disease activity

JDM Dataset 81 kids, 7 on IVIg at baseline, 23 on IVIG later Outcome is time to quiescence Quiescence happens when DAS=0 IVIg tends to be given when the child is doing particularly badly (high DAS) DAS is a counfounder

Causal Pathway for JDM study DAS confounds IVIg and outcome DAS is on the causal pathway DAS t IVIg t IVIg t+1 DAS t+1 Time-to-Quiescence … …

A Thought Experiment Suppose that at each time t, we could create an identical copy of each child i. Then if the real child received IVIG, we would give the copy control and vice versa We could then compare the child to its copy Solves confounding by matching: the child is matched with the copy If treatment varies on a monthly basis and we follow for 5 years, we would have copies

Counterfactuals Clearly, this is impossible. But we can use the idea Define the counterfactuals for child i to be the outcomes for each of the imaginary copies Idea: treat the counterfactuals as missing data

Inverse-Probability Weighting Inverse-Probability Weighting (IPW) is a way of re-weighting the dataset to account for selective observation E.g. if we have missing data, then we weight the observed data by the inverse of the probability of being observed Why does this work? Suppose we have a response Y ij, treatment indicator x ij and R ij =1 if Y ij observed, 0 o/w

Inverse-Probability Weighting Suppose we want to fit the marginal model Usually, we solve the GEE equation If we use just the observed data, we solve LHS does not have mean 0

Inverse-Probability Weighting If we replace  by with  ij, the conditional probability of observing Y ij, then What to condition on? Must condition on Y ij If MAR, then conditionally independent given previous Y

Marginal Structural Models MSMs use inverse-probability weighting to deal with the unobserved (“missing”) counterfactuals We cannot adjust for confounders… …but using IPW, can re-weight the dataset so that treatment and covariates are unconfounded i.e. mean covariate levels are the sample between treated and untreated patients So can do a simple marginal analysis

Probability-of-Treatment Model Weighting is based on the Probability-of- Treatment model Treatment is longitudinal For each child at each time, need probability of receiving the observed treatment trajectory Probability is conditional on past responses and confounders Assume independent of current response

JDM Example Probability of being on IVIg at baseline (logistic regression) Probability of transitioning onto IVIg (Cox PH) Probability of transitioning off IVIg (Cox PH) Suppose a child initiates IVIG at 8 months and is still on IVIG at 12 months. What is the probability of the observed treatment pattern?

Trratment probability No IVIg 08 Initiate IVIg 12 Still on IVIG P(not on IVIg at baseline) P(no transition before month 8) P(transition at month 8) P(no transition off before month 12)

Model Fitting First identified covariates univariately Then entered those that were sig. into model and refined (by removing those that were no longer sig.) IVIg at baseline: Functional status (any vs. none) OR 11.6, 95% CI 1.94 to 69.7; abnormal swallow/voice OR 6.28, 95% CI to IVIg termination: no covariates

Assessing goodness-of-fit If the IPT weights are correct, in the re- weighted population, treatment and covariates are unconfounded This property is crucial testable …so we should test it!

Goodness-of-fit in the JDM study Biggest concern is that kids are doing badly when they start IVIg If inverse-probability weights are correct, then at each time t, amongst patients previously IVIg- naïve, IVIg is not associated with covariates. Will look at differences in mean covariate values by current IVIg status amongst patients previously IVIg-naïve Data are longitudinal, so use a GEE analysis, adjusting for time

Model 1 – HRs for Treatment Initiation Hazard Ratios and 95% confidence intervals for initiating treatment

Model 2 -Revised Treatment initiation Hazard Ratios and 95% confidence intervals for initiating treatment

New goodness-of-fit

Back to basics Some patients start IVIg because they are steroid- resistant (early-starters) Others start because they are steroid-dependent (late-starters) Repeat model-fitting process separately for early and late starters

Refined two-stage model

Efficacy Results

Other concerns with MSMs Format of treatment effect (e.g. constant over time, PH etc.) Unmeasured counfounders Lack of efficiency Experimental Treatment Assignment

Efficiency IPW reduces bias but also reduces efficiency The further the weights are from 1, the worse the efficiency Can stabilise the weights: Estimating equations will still be zero-mean if we multiply  ijj by a factor depending on j and treatment In JDM study, we used  ijj =R ij P(Rx history)/P(Rx history|confounders)

Efficiency – other techniques Doubly robust methods (Bang & Robins) Could have used a more information-rich outcome Did a secondary analysis using DAS as the outcome – got far more precise (and more positive) results

Experimental Treatment Assignment In order for MSMs to work, there must be some experimentality in the way treatment is assigned Intuitively, if we can predict perfectly who will get what treatment, then we have complete confounding Mathematically, if  ij is 0 then we’re in trouble! Actually, we get into trouble if  ij = 0 or 1

Testing the ETA – simple checks At each time j, review the distribution of covariates amongst those who are on treatment vs. those who are not. Review the distribution of the weights check  bounded away from 0/1 In the JDM example, also check distn of transition probabilities

Testing the ETA – more advanced methods Bootstrapping Wang Y, Petersen ML, Bangsberg D, van der Laan MJ. Diagnosing bias in the inverse probability of treatment weighted estimator resulting from violation of experimental treatment assignment. UC Berkeley Division of Biostatistics working paper series, 2006.

Implementing MSMs For time-to-event outcome, can do weighted PH regression in R Used the svycoxph function from the survey package For continuous (or binary) outcome, use weighted GEE Used proc genmod in SAS with scgwt Weighted GEEs are not straightforward in R STATA could probably handle either type of outcome

MSMs - potentials Often good observational databases exist Should do what we can with them before using large amounts of money to do trials Can deal with a time-varying treatment Conceptually fairly straightforward Do not have to model correlation structure in responses

MSMs - limitations There may always be unmeasured confounders Relies heavily on probability-of-treatment model being correct Experimental ETA violations can often occur (particularly with small sample sizes) Somewhat inefficient Doubly robust methods may help Not a replacement for an RCT

Key points MSMs can help to establish causal associations from observational data Make some strong assumptions Need goodness-of-fit for measured confounders Will never find the right model Aim to find good models

References Robins JM, Hernan MA, Brumback B. Marginal structural models and causal inference in epidemiology. Epidemiology 2000; 11: Bang H, Robins JM (2005). Doubly Robust Estimation in Missing Data and Causal Inference Models. Biometrics 61 (4), 962–973. Pullenayegum EM, Lam C, Manlhiot C, Feldman BM. Fitting Marginal Structural Models: Estimating covariate-treatment associations in the re-weighted dataset can guide model fitting. Journal of Clinical Epidemiology. Wang Y, Petersen ML, Bangsberg D, van der Laan MJ. Diagnosing bias in the inverse probability of treatment weighted estimator resulting from violation of experimental treatment assignment. UC Berkeley Division of Biostatistics working paper series, 2006.