Presentation is loading. Please wait.

Presentation is loading. Please wait.

Clash of Causal Inference Techniques

Similar presentations


Presentation on theme: "Clash of Causal Inference Techniques"— Presentation transcript:

1 Clash of Causal Inference Techniques
Clash London Calling album cover from Basil Rathbone as Sherlock Holmes photo By Employee(s) of Universal Studios (Photograph in possession of SchroCat) [Public domain], via Wikimedia Commons Terrence Willett and Nathan Pellegrin RP Group conference Kellogg west April 2014

2

3 Outcomes Describe purpose of regression and propensity score matching (PSM) Explain data requirements and basic procedure of regression and PSM Compare and contrast regression and PSM Identify additional resources for further exploration

4 Why causal inference If you need to use statistics, then you should design a better experiment. –attributed to Rutherford Most education research is observational/correlational, not experimental. Atom from

5 Common scenario Did participation in an activity, class or support result in higher outcomes for students than would have happened had they not participated? Students self-selected to participate and/or were recruited to participate. When participants are compared to non-participants, differences in outcomes can be attributed to differences in background variables or motivation. Can we determine if the participation caused a change in outcomes? No, but…

6 Regression 𝑦= 𝑎 0 + 𝑎 1 𝑥 1 + 𝑎 2 𝑥 2 …+ 𝑎 𝑛 𝑥 𝑛 +𝑒
Classic correlational technique Covariates used in model to attempt to control for differences in background variables or motivation Background variables can include measures of or proxies for skill level, social capital, or socio-economic status Measures of self-motivation often unavailable Models are imperfect and generally must be combined with other evidence to more completely describe the possible influence of an intervention, program or strategy

7 Propensity Score Matching
One of several ways to create a matched comparison group of non-participants intended to be similar to participant group for a valid comparison Likelihood or other techniques used to create a score indicating the likelihood that a particular non-participant would have been a participant based on similarity to one or more participants

8 The Counterfactual (potential outcomes) framework
∆ = 𝑌 𝑡 − 𝑌 𝑐 This shows the causal effect of some event (participation, treatment, intervention) on an individual. 𝑌 𝑡 is the potential outcome under treatment. 𝑌 𝑐 is the potential outcome under non-treatment (control, counterfactual) and thus cannot be observed in our universe*. *In light of recently published evidence for inflationary theories of the cosmos, we note that we may be living in a multiverse (Alan Guth, 2014). Counterfactual conditions may obtain in alternate universes.

9 The potential outcomes matrix
Yt Yc Actual Treatment Status (T) 1 = observable = not observable

10 Average Treatment effect (ATE)
Participants and non-participants differ systematically (w.r.t. demographics, trajectories, risk profiles, self-selection, etc.) Different people respond differently to treatment (differential response) These facts must be taken into account when modeling/computing treatment effects. This means all four cells of the matrix must be estimated in order to obtain an average treatment effect ! How ? Programs (treatments) are often designed for and targeted toward those who are expected to gain the most from participation (aka “positive selection”); the potential treatment effect can be different for participants and non-participants.

11 Symbolic derivation of average effects
1 ∆ = 𝑌 𝑖 𝑡 − 𝑌 𝑖 𝑐 2 𝑌 𝑖 = 𝑇 𝑖 𝑌 𝑖 𝑡 − (1−𝑇 𝑖 ) 𝑌 𝑖 𝑐 3 𝐸[∆] =𝐸[ 𝑌 𝑖 𝑡]−𝐸[ 𝑌 𝑖 𝑐] 4 𝐸 𝑁 [ 𝑌 𝑖 | 𝑇 𝑖 =1]− 𝐸 𝑁 [ 𝑌 𝑖 | 𝑇 𝑖 =0] 5 𝐸 ∆ = 𝜌𝐸 𝑌 𝑖 𝑡 𝑇 𝑖 =1 + 1−𝜌 𝐸 𝑌 𝑖 𝑡 𝑇 𝑖 =0 −{𝜌𝐸 𝑌 𝑖 𝑐 𝑇 𝑖 =1 + 1−𝜌 𝐸 𝑌 𝑖 𝑐 𝑇 𝑖 =0 } 6 𝐸 ∆ 𝑇 𝑖 =1 =𝐸 𝑌 𝑖 𝑡 𝑇 𝑖 =1 −𝐸 𝑌 𝑖 𝑐 𝑇 𝑖 =1 7 𝐸 ∆ 𝑇 𝑖 =0 =𝐸 𝑌 𝑖 𝑡 𝑇 𝑖 =0 −𝐸 𝑌 𝑖 𝑐 𝑇 𝑖 =0 ATE ATT ATU

12 Conditional independence (perfect stratification) (SELection on observables)
If our observations include information on every one of the variables influencing likelihood of participation or differential responses*, then it is possible to avoid omitted variable bias and so achieve CI (PS, SOO). And if we have CI, this is like randomized assignment …. 𝐸 𝑌 𝑡 𝑇=1, 𝑿 = 𝐸 𝑌 𝑡 𝑇=0, 𝑿 𝐸 𝑌 𝑐 𝑇=1, 𝑿 = 𝐸 𝑌 𝑐 𝑇=0, 𝑿 * How often do we encounter such datasets in institutional research? The equations at the bottom capture the conditions that are achieved through random assignment

13 Assuming CI… Potential Outcome Yt Yc Actual Treatment Status (T) 1
= observable = not observable Y(t) Y(c)

14 Average Treatment effect (ATE)
PSM offers a way to plug values into (5), (6) and (7) to obtain unbiased estimates. However, if we have CI, then why not just something like…. 𝑌 𝑖 =𝛼+𝛾 𝑇 𝑖 +𝜷′ 𝑿 𝑖 + 𝜀 𝑖 ? (and many other regression techniques) There is no firm and fast answer to this question. Decisions are based on pragmatic considerations. However….. Programs (treatments) are often designed for and targeted toward those who are expected to gain the most from participation (aka “positive selection”); the potential treatment effect can be different for participants and non-participants.

15 example

16 Pros and cons Regressions can be “easier” to run but harder to explain to a general audience PSM can be more time consuming to conduct but easier to explain to a general audience Regressions tend to perform better with large data sets while PSM tends to perform better with few observations provided the non-participant group has sufficient numbers of individuals with the key confounding variables Regressions have been used for many years and are well described mathematically with broad consensus on proper error terms PSM is newer and there is not consensus on optimal matching procedures or proper error terms Regression will use all cases with non-missing data while PSM may only a subset of cases from the pool of non-participants All analytic methods suffer if key variables are not available Conclusions can often be the same with either method

17 How to run PSM Create data file (95% of effort)
Match participants and non-participants on a set of control variables to create a comparison group with similar proportions on all characteristics (i.e. comparison group would have a similar percent female, Hispanic, low income, etc. as compared to the participant group) This step is referred to as “balancing” and generally must be repeated several times to obtain balance on all variables of interest either by adjusting matching criteria or removing variables Run comparative analyses, which can include simple t-tests, post-PSM regressions, or other techniques Major packages that conduct PSM include STATA, R, and SAS STATA version 12 and older have psmatch2, v13 has teffects psmatch Note SPSS/PASW does not do PSM directly but there is an R plugin for SPSS

18 An alternative strategy to deciding this battle (warning: Empiricists approaching)

19 Alternative perspectives
Estimating program effects based on observational data can also be understood as… Delimiting the error “built-into” inductive reasoning about causes (the problem of induction) An “inverse problem” in the study of social dynamics. A missing data problem An optimization problem From these alternate perspectives there is a large menu of methods and extensions, including Euclidean, Mahalanobis and Gower’s distance, nearest neighbor, cosine similarity, kernal functions, genetic algorithms, imputation, dimensionality reduction, and other supervised/unsupervised learning algorithms A criterion that can be applied to regression, PSM and other methods is this: how do they perform at predicting new (future) observations? (false positives, false negatives, correlated errors) In empirical investigations we do not answer this question just once using new replications or sampling under varying conditions. It is applicable in any situation where predictions are made concerning new (possibly future) observations. We update our models based on new evidence. And in this respect, regression and PSM methods can both be used as tools of discovery; as ways to extend our understanding of the processes producing patterns we find in sets of observations (and in streams of information, generally). SO: CHOOSE THE METHOD/MODEL WHICH YIELDS THE SMALLEST PREDICTION ERROR. That may decide a battle in a particular setting (or occasion), but the war between methods will go on …. A number of the studies I reviewed used a comparison scheme in which a dataset is constructed for a certain value of ATE, and then estimates of ATE are recovered via PSM and regression. However OLS regression is typically used, and since it has been recognized for decades that OLS has limitations. Many techniques have been developed to address these limitations (ridge regression, quantile regression, bootstrapping, non-linear models….). “empirical investigations “ includes applied econ and “the quant side of business”.

20 “The inability to predict outliers implies the inability to predict the course of history” ― Nassim Nicholas Taleb, The Black Swan: The Impact of the Highly Improbable “If you insist on strict proof (or strict disproof) in the empirical sciences, you will never benefit from experience, and never learn from it how wrong you are.“ — Karl Raimund Popper, The Logic of Scientific Discovery: Logik Der Forschung (2002), 28.

21 Further reading Angrist, J. D., & Pischke, J. (2008). Mostly Harmless Econometrics: An Empiricists Companion Morgan, S., Harding, D. (2006) Matching Estimators of Causal Effects: From Stratification and Weighting to Practical Data Analysis Routines Caliendo and Kopeinig Practical Guide for PSM Rosenbaum, P. R., & Rubin, D. B. (1983). The central role of the propensity score in observational studies for causal effects. Biometrika, 70, Padgett, R.; Salisbury, M.; An, B.; & Pascarella, E. (2010). Required, Practical, or Unnecessary? An Examination and Demonstration of Propensity Score Matching Using Longitudinal Secondary Data. New Directions for Institutional Research – Assessment Supplement (pp ). San Francisco, CA: Jossey-Bass. Soledad Cepeda, M.; Boston, R.; Farrar, J., & Strom, B. (2003). Comparison of Logistic Regression versus Propensity Score When the Number of Events Is Low and There Are Multiple Confounders. American Journal of Epidemiology, 158,

22 Thank you Terrence Willett Director of Planning, Research, and Knowledge Systems Cabrillo College Nathan Pellegrin Data Processing Specialist Peralta District


Download ppt "Clash of Causal Inference Techniques"

Similar presentations


Ads by Google