Empirical Efficiency Maximization: Locally Efficient Covariate Adjustment in Randomized Experiments Daniel B. Rubin Joint work with Mark J. van der Laan.

Slides:



Advertisements
Similar presentations
COMPUTER INTENSIVE AND RE-RANDOMIZATION TESTS IN CLINICAL TRIALS Thomas Hammerstrom, Ph.D. USFDA, Division of Biometrics The opinions expressed are those.
Advertisements

Assessing the Effects of Time-varying Predictors or Treatments: A Conceptual Discussion Daniel Almirall VA Medical Center, HSRD Duke Medical Center, Dept.
Brief introduction on Logistic Regression
M. Kathleen Kerr “Design Considerations for Efficient and Effective Microarray Studies” Biometrics 59, ; December 2003 Biostatistics Article Oncology.
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
Estimating Direct Effects of New HIV Prevention Methods. Focus: the MIRA Trial UCSF: Helen Cheng, Nancy Padian, Steve Shiboski, Ariane van der Straten.
Logistic Regression.
Informative Censoring Addressing Bias in Effect Estimates Due to Study Drop-out Mark van der Laan and Maya Petersen Division of Biostatistics, University.
Chapter 8 Estimation: Additional Topics
x – independent variable (input)
 Confounders are usually controlled with the “standard” response regression model.  The standard model includes confounders as covariates in the response.
Maximum likelihood (ML) and likelihood ratio (LR) test
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Model and Variable Selections for Personalized Medicine Lu Tian (Northwestern University) Hajime Uno (Kitasato University) Tianxi Cai, Els Goetghebeur,
Lecture 23: Tues., Dec. 2 Today: Thursday:

1 Unsupervised Learning With Non-ignorable Missing Data Machine Learning Group Talk University of Toronto Monday Oct 4, 2004 Ben Marlin Sam Roweis Rich.
Maximum likelihood (ML)
Common Problems in Writing Statistical Plan of Clinical Trial Protocol Liying XU CCTER CUHK.
Planning Survival Analysis Studies of Dynamic Treatment Regimes Z. Li & S.A. Murphy UNC October, 2009.
Evaluating Hypotheses
Dynamic Treatment Regimes, STAR*D & Voting D. Lizotte, E. Laber & S. Murphy Psychiatric Biostatistics Symposium May 2009.
Chapter 11 Survival Analysis Part 2. 2 Survival Analysis and Regression Combine lots of information Combine lots of information Look at several variables.
Model Selection in Semiparametrics and Measurement Error Models Raymond J. Carroll Department of Statistics Faculty of Nutrition and Toxicology Texas A&M.
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
Independent Sample T-test Classical design used in psychology/medicine N subjects are randomly assigned to two groups (Control * Treatment). After treatment,
Modeling clustered survival data The different approaches.
Using Regression Models to Analyze Randomized Trials: Asymptotically Valid Tests Despite Incorrect Regression Models Michael Rosenblum, UCSF TAPS Fellow.
Maximum likelihood (ML)
Adaptive Designs for Clinical Trials
Sample Size Determination Ziad Taib March 7, 2014.
Assessing Survival: Cox Proportional Hazards Model Peter T. Donnan Professor of Epidemiology and Biostatistics Statistics for Health Research.
Survival Analysis A Brief Introduction Survival Function, Hazard Function In many medical studies, the primary endpoint is time until an event.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 12: Multiple and Logistic Regression Marshall University.
GEE and Generalized Linear Mixed Models
Overall agenda Part 1 and 2  Part 1: Basic statistical concepts and descriptive statistics summarizing and visualising data describing data -measures.
Modeling Menstrual Cycle Length in Pre- and Peri-Menopausal Women Michael Elliott Xiaobi Huang Sioban Harlow University of Michigan School of Public Health.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Assessing Survival: Cox Proportional Hazards Model
Estimation Kline Chapter 7 (skip , appendices)
April 4 Logistic Regression –Lee Chapter 9 –Cody and Smith 9:F.
1 THE ROLE OF COVARIATES IN CLINICAL TRIALS ANALYSES Ralph B. D’Agostino, Sr., PhD Boston University FDA ODAC March 13, 2006.
Maximum Likelihood Estimation Methods of Economic Investigation Lecture 17.
MBP1010 – Lecture 8: March 1, Odds Ratio/Relative Risk Logistic Regression Survival Analysis Reading: papers on OR and survival analysis (Resources)
Stat 112: Notes 2 Today’s class: Section 3.3. –Full description of simple linear regression model. –Checking the assumptions of the simple linear regression.
Lecture 9: Analysis of intervention studies Randomized trial - categorical outcome Measures of risk: –incidence rate of an adverse event (death, etc) It.
Introduction Sample Size Calculation for Comparing Strategies in Two-Stage Randomizations with Censored Data Zhiguo Li and Susan Murphy Institute for Social.
Introduction to logistic regression and Generalized Linear Models July 14, 2011 Introduction to Statistical Measurement and Modeling Karen Bandeen-Roche,
CHAPTER 17 O PTIMAL D ESIGN FOR E XPERIMENTAL I NPUTS Organization of chapter in ISSO –Background Motivation Finite sample and asymptotic (continuous)
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Simulation Study for Longitudinal Data with Nonignorable Missing Data Rong Liu, PhD Candidate Dr. Ramakrishnan, Advisor Department of Biostatistics Virginia.
Using Propensity Score Matching in Observational Services Research Neal Wallace, Ph.D. Portland State University February
Effect of the Reference Set on Frequency Inference Donald A. Pierce Radiation Effects Research Foundation, Japan Ruggero Bellio Udine University, Italy.
Case Selection and Resampling Lucila Ohno-Machado HST951.
Love does not come by demanding from others, but it is a self initiation. Survival Analysis.
Machine Learning 5. Parametric Methods.
Estimation Kline Chapter 7 (skip , appendices)
Review of statistical modeling and probability theory Alan Moses ML4bio.
Machine Learning CUNY Graduate Center Lecture 6: Linear Regression II.
Rerandomization to Improve Covariate Balance in Randomized Experiments Kari Lock Harvard Statistics Advisor: Don Rubin 4/28/11.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
Nonparametric Statistics
Direct method of standardization of indices. Average Values n Mean:  the average of the data  sensitive to outlying data n Median:  the middle of the.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 13: Multiple, Logistic and Proportional Hazards Regression.
Jeffrey E. Korte, PhD BMTRY 747: Foundations of Epidemiology II
Logistic Regression.
Yu Du, PhD Research Scientist Eli Lilly and Company
Presentation transcript:

Empirical Efficiency Maximization: Locally Efficient Covariate Adjustment in Randomized Experiments Daniel B. Rubin Joint work with Mark J. van der Laan

Outline Review adjustment in experiments. Locally efficient estimation. Problems with standard methods. New method addressing problems. Abstract formulation. Back to experiments, with simulations and numerical results. Application to survival analysis.

Randomized Experiments (no covariates yet)

Randomized Experiments

Example: Women recruited. Randomly assigned to diaphragm or no diaphragm. See if they get HIV. Example: Men recruited. Randomly assigned to circumcision or no circumcision. See if they get HIV.

Randomized Experiments Randomization allows causal inference. No confounding. Differences in outcomes between treatment and control groups are due to the treatment. Unverifiable assumptions needed for causal inference in observational studies.

Counterfactual Outcome Framework: Neyman-Rubin Model

Causal Parameters

Causal Parameters (binary response)

Estimating Causal Parameters in Randomized Experiments

Randomized Experiments with Covariates Same setup, only now demographic or clinical measurements taken prior to randomization. Question: With extra information, can we more precisely estimate causal parameters? Answer: Yes. (Fisher, 1932). Subject’s covariate has information about how he or she would have responded in both arms.

Covariate Adjustment (has at least two meanings) 1: Gaining precision in randomized experiments. 2: Accounting for confounding in observational studies. This talk only deals with the first meaning.

Covariate Adjustment. Not very difficult when covariates divide subjects into a handful of strata. Have to think with a single continuous covariate (e.g. age). Modern studies can collect a lot of baseline information. Gene expression profile, complete medical history, biometric measurements. Important longstanding problem, but lots of confusion. Not “solved.” Recent work by others.

Covariate Adjustment Pocock et al. (2002) recently surveyed 50 clinical trial reports. Of 50 reports, 36 used covariate adjustment for estimating causal parameters, and 12 emphasized adjusted over unadjusted analysis. “Nevertheless, the statistical emphasis on covariate adjustment is quite complex and often poorly understood, and there remains confusion as to what is an appropriate statistical strategy.”

Recent Work on this Problem Koch et al. (1998). - Modifications to ANCOVA. Tsiatis et al. (2000, 2007a, 2007b). - Locally efficient estimation. Moore and van der Laan (2007). - Targeted maximum likelihood. Freedman (2007a, 2007b). - Classical methods under misspecification.

Covariate Adjustment Can choose to ignore baseline measurements. Why might extra precision be worth it? Narrow confidence intervals for treatment effect. Stop trials earlier. Subgroup analysis, having smaller sample sizes.

Covariate Adjustment

Example: Intention-to-treat

Example: Log Odds Ratio

Locally Efficient Estimation Primarily motivated by causal inference problems in observational studies. Origin in Robins and Rotnitzky (1992), Robins, Rotnitzky, and Zhao (1994). Surveyed in van der Laan and Robins (2003), Tsiatis (2006).

Locally Efficient Estimation

Locally Efficient Estimation in Randomized Experiments Working model for treatment distribution given covariates known by design. So what does local efficiency tell us? Model outcome distribution given (covariates, treatment). We’ll be asymptotically efficient if working model is correct, but still asymptotically normal otherwise. But what does this mean if there’s no reason to believe the working model? Unadjusted estimators are also asymptotically normal. What about precision?

Empirical Efficiency Maximization Working model for outcome distribution given (treatment, covariates) typically fit with likelihood-based methods. Often linear, logistic, or Cox regression models. “Factorization of likelihood” means such estimates lead to double robustness in observational studies. But such robustness is superfluous in controlled experiments. We try to find the working model element resulting in the parameter estimate with smallest asymptotic variance.

Interlude: Abstract Formulation

Asymptotics

Interlude: Abstract Formulation

Back to Randomized Experiments

Randomized Experiments

Working Model Loss Function

Connection to High Dimensional or Complex Data Suppose a high dimensional covariate is related to the outcome, and we would like to adjust for it to gain precision. Many steps in data processing can be somewhat arbitrary (e.g. dimension reduction, smoothing, noise removal). With cross-validation, new loss function can guide selection of tuning parameters governing this data processing.

Numerical Asymptotic Efficiency Calculations

1=Unadjusted estimator ignoring covariates. 2=Likelihood-based locally efficient. 3=Empirical efficiency maximization. 4=Efficient.

Intention-to-treat Parameter

Numerical Asymptotic Efficiency Calculations

1=Unadjusted difference in means. 2=Standard likelihood-based locally efficient estimator. 3=empirical efficiency maximization. 4=efficient estimator.

Treatment Effects

Sneak Peak: Survival Analysis

Form locally efficient estimate. Working model for full data distribution now likely a proportional hazards model. For estimating (e.g. five-year survival), will be asymptotically efficient if model correct. Otherwise still asymptotically normal. But locally efficient estimator can be worse than Kaplan-Meier if model is wrong.

Sneak Peak: Survival Analysis

Sneak Peak: Survival Analysis. Generated Data. Want five-year survival.

1=Kaplan-Meier. 2=Likelihood-based locally efficient. 3=Empirical efficiency maximization. 4=Efficient.

Summary Robins and Rotnitzky’s locally efficient estimation developed for causal inference in observational studies. In experiments, estimators can gain or lose efficiency, depending on validity of working model. Often there might be no reason to have any credence in a working model. A robustness result implies we can better fit working model (or select tuning parameters in data processing) with a nonstandard loss function (the squared influence curve).

Troublesome Issues When working regression model is linear, our procedure usually reduces to existing procedures. When nonlinear, fitting regression model with our loss function can entail non- convex optimization. Proving convergence to working model elements for well-known logistic and Cox regression models.

References Empirical Efficiency Maximization: Improved Locally Efficient Covariate Adjustment. In press in International Journal of Biostatistics. Technical reports at and