Www.worldbank.org/hdchiefeconomist The World Bank Human Development Network Spanish Impact Evaluation Fund.

Slides:



Advertisements
Similar presentations
Explanation of slide: Logos, to show while the audience arrive.
Advertisements

Explanation of slide: Logos, to show while the audience arrive.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Designing an impact evaluation: Randomization, statistical power, and some more fun…
The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Explanation of slide: Logos, to show while the audience arrive.
Holland on Rubin’s Model Part II. Formalizing These Intuitions. In the 1920 ’ s and 30 ’ s Jerzy Neyman, a Polish statistician, developed a mathematical.
The counterfactual logic for public policy evaluation Alberto Martini hard at first, natural later 1.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Correlation and regression Dr. Ghada Abo-Zaid
The World Bank Human Development Network Spanish Impact Evaluation Fund.
PHSSR IG CyberSeminar Introductory Remarks Bryan Dowd Division of Health Policy and Management School of Public Health University of Minnesota.
Study Design Data. Types of studies Design of study determines whether: –an inference to the population can be made –causality can be inferred random.
Chapter 2 – Tools of Positive Analysis
TOOLS OF POSITIVE ANALYSIS
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Chapter 2: The Research Enterprise in Psychology
1. An Overview of the Data Analysis and Probability Standard for School Mathematics? 2.
Chapter 2: The Research Enterprise in Psychology
Chapter 2 The Research Enterprise in Psychology. n Basic assumption: events are governed by some lawful order  Goals: Measurement and description Understanding.
Fundamentals of Data Analysis Lecture 4 Testing of statistical hypotheses.
Understanding Statistics
Chapter 1: The Research Enterprise in Psychology.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Experimental Design making causal inferences Richard Lambert, Ph.D.
Chapter 2 The Research Enterprise in Psychology. Table of Contents The Scientific Approach: A Search for Laws Basic assumption: events are governed by.
CAUSAL INFERENCE Shwetlena Sabarwal Africa Program for Education Impact Evaluation Accra, Ghana, May 2010.
Slides to accompany Weathington, Cunningham & Pittenger (2010), Chapter 3: The Foundations of Research 1.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Chapter 1 Measurement, Statistics, and Research. What is Measurement? Measurement is the process of comparing a value to a standard Measurement is the.
Econometrics ECO 54 History of Economic Thought Udayan Roy.
Estimating Causal Effects from Large Data Sets Using Propensity Scores Hal V. Barron, MD TICR 5/06.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Causal Inference Nandini Krishnan Africa Impact Evaluation.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
ECON 3039 Labor Economics By Elliott Fan Economics, NTU Elliott Fan: Labor 2015 Fall Lecture 21.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Propensity Score Matching for Causal Inference: Possibilities, Limitations, and an Example sean f. reardon MAPSS colloquium March 6, 2007.
LECTURE 1 - SCOPE, OBJECTIVES AND METHODS OF DISCIPLINE "ECONOMETRICS"
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Nigeria Impact Evaluation Community of Practice Abuja, Nigeria, April 2, 2014 Measuring Program Impacts Through Randomization David Evans (World Bank)
Chapter 2 The Research Enterprise in Psychology. Table of Contents The Scientific Approach: A Search for Laws Basic assumption: events are governed by.
McGraw-Hill/Irwin Copyright © 2008 by The McGraw-Hill Companies, Inc. All rights reserved. CHAPTER 2 Tools of Positive Analysis.
Christel M. J. Vermeersch November 2006 Session V Instrumental Variables.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Impact Evaluation Sebastian Galiani November 2006 Causal Inference.
1 General Elements in Evaluation Research. 2 Types of Evaluations.
Causal Model Ying Nian Wu UCLA Department of Statistics July 13, 2007 IPAM Summer School.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Public Finance and Public Policy Jonathan Gruber Third Edition Copyright © 2010 Worth Publishers 1 of 24 Copyright © 2010 Worth Publishers.
What is Impact Evaluation … and How Do We Use It? Deon Filmer Development Research Group, The World Bank Evidence-Based Decision-Making in Education Workshop.
How Psychologists Do Research Chapter 2. How Psychologists Do Research What makes psychological research scientific? Research Methods Descriptive studies.
Copyright © 2015 Inter-American Development Bank. This work is licensed under a Creative Commons IGO 3.0 Attribution-Non Commercial-No Derivatives (CC-IGO.
Impact Evaluation Methods Randomization and Causal Inference Slides by Paul J. Gertler & Sebastian Martinez.
Ten things about Experimental Design AP Statistics, Second Semester Review.
Experimental Evaluations Methods of Economic Investigation Lecture 4.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
Technical Track Session III
Explanation of slide: Logos, to show while the audience arrive.
Explanation of slide: Logos, to show while the audience arrive.
Introduction to Econometrics
Statistical Data Analysis
Explanation of slide: Logos, to show while the audience arrive.
Explanation of slide: Logos, to show while the audience arrive.
David Mann David Stapleton (Mathematica Policy Research) Alice Porter
Positive analysis in public finance
Introduction to Producing Data
Presentation transcript:

The World Bank Human Development Network Spanish Impact Evaluation Fund

This material constitutes supporting material for the "Impact Evaluation in Practice" book. This additional material is made freely but please acknowledge its use as follows: Gertler, P. J.; Martinez, S., Premand, P., Rawlings, L. B. and Christel M. J. Vermeersch, 2010, Impact Evaluation in Practice: Ancillary Material, The World Bank, Washington DC ( The content of this presentation reflects the views of the authors and not necessarily those of the World Bank. CAUSAL INFERENCE Technical Track Session I

Policy questions are causal in nature Does school decentralization improve school quality? Does one more year of education cause higher income? Does conditional cash transfers cause better health outcomes in children? How do we improve student learning? Cause-effect relationships are a part of what policy makers do:

Policy questions are causal in nature Does school decentralization improve school quality? Does one more year of education cause higher income? Does conditional cash transfers cause better health outcomes in children? How do we improve student learning? Cause-effect relationships are a part of what policy makers do: But the statistics you learnt in school/university do not address this…

Standard Statistical Analysis Likelihood and other estimation techniques. To infer parameters of a distribution from samples drawn of that distribution. With the help of such parameters, one can: o Infer association among variables, o Estimate the likelihood of past and future events, o Update the likelihood of events in light of new evidence or new measurement. Aim Uses Tools

Standard Statistical Analysis For this to work well, experimental conditions must remain the same. But our policy questions were: o If I decentralize schools, will quality improve? o If I find a way to make a child go to school longer, will she earn more money? o If I start giving cash to families, will their children become healthier? o If I train teachers, will students learn more? Condition

Standard Statistical Analysis For this to work well, experimental conditions must remain the same. But our policy questions were: o If I decentralize schools, will quality improve? o If I find a way to make a child go to school longer, will she earn more money? o If I start giving cash to families, will their children become healthier? o If I train teachers, will students learn more? Condition The conditions change!!

Causal Analysis For causal questions, we need to infer aspects of the data generation process. We need to be able to deduce: 1. the likelihood of events under static conditions, (as in Standard Statistical Analysis) and also 2. the dynamics of events under changing conditions.

Causal Analysis “dynamics of events under changing conditions” includes: 1. Predicting the effects of interventions. 2. Predicting the effects of spontaneous changes. 3. Identifying causes of reported events.

Causation vs. Correlation Standard statistical analysis/probability theory: o The word “cause” is not in its vocabulary. o Allows us to say is that two events are mutually correlated, or dependent. This is not enough for policy makers o They look at rationales for policy decisions: if we do XXX, then will we get YYY? o We need a vocabulary for causality.

THE RUBIN CAUSAL MODEL Vocabulary for Causality

Population & Outcome Variable Define the population by U. Each unit in U is denoted by u. The outcome of interest is Y. Also called the response variable. For each u  U, there is an associated value Y(u).

“ Causes/Treatment For simplicity, we assume that there are just two possible states: o Unit u is exposed to treatment and o Unit u is exposed to comparison. Causes are those things that could be treatments in hypothetical experiments. - Rubin

The Treatment Variable Let D be a variable that indicates the state to which each unit in U is exposed. Where does D come from? o In a controlled study: constructed by the experimenter. o In an uncontrolled study: determined by factors beyond the experimenter’s control. D = 1If unit u is exposed to treatment 0If unit u is exposed to comparison

Linking Y and D Thus, we need two response variables: o Y 1 (u) is the outcome if unit u is exposed to treatment. o Y 0 (u) is the outcome if unit u is exposed to comparison. Y=response variable D= treatment variable The response Y is potentially affected by whether u receives treatment or not.

The effect of treatment on outcome Response variable Y Treatment variable D D = 1If unit u is exposed to treatment 0If unit u is exposed to comparison Y 1 (u) is the outcome if unit u is exposed to treatment is the outcome if unit u is exposed to comparison Y 0 (u) For any unit u, treatment causes the effect δ u = Y 1 (u) - Y 0 (u)

But there is a problem: For any unit u, treatment causes the effect δ u = Y 1 (u) - Y 0 (u) Fundamental problem of causal inference o For a given unit u, we observe either Y 1 (u) or Y 0 (u), o it is impossible to observe the effect of treatment on u by itself! We do not observe the counterfactual If we give u treatment, then we cannot observe what would have happened to u in the absence of treatment.

So what do we do? Instead of measuring the treatment effect on unit u, we identify the average treatment effect for the population U (or for sub-populations)

Estimating the ATE Most econometrics methods attempt to construct from observational data consistent estimators of: E U (Y 1 ) = Y̅ 1 and E U (Y 0 )= Y̅ 0 Although E U (Y 1 ) and E U (Y 0 ) cannot both be calculated, they can be estimated. So, (1) Replace the impossible-to-observe treatment effect of D on a specific unit u, (2) with the possible-to-estimate average treatment effect of D over a population U of such units.

A simple estimator of ATE U So we are trying to estimate: ATE U = E U (Y 1 ) - E U (Y 0 ) = Y̅ 1 - Y̅ 0 (1) Consider the following simple estimator: δ̅̂ = [ Y̅̂ 1 | D = 1] - [ Y̅̂ 0 | D =0 ] Note o Equation (1) is defined for the whole population. o Equation (2) is an estimator to be computed on a sample drawn from that population. (2)

An important lemma

Proof: Try this at home

Fundamental Conditions Thus, a sufficient condition for the simple estimator to consistently estimate the true ATE is that: [Y̅ 1 |D=1]=[Y̅ 1 |D=0] The average outcome under treatment Y̅ 1 is the same for the treatment (D=1) and the comparison (D=0) groups [Y̅ 0 |D=1]=[Y̅ 0 |D=0] The average outcome under comparison Y̅ 0 is the same for the treatment (D=1) and the comparison (D=0) groups And

When will those conditions be satisfied? Intuitively, there can be no correlation between (1) Whether someone gets the treatment and (2) How much that person potentially benefits from the treatment. It is sufficient that treatment assignment D be uncorrelated with the potential outcome distributions of Y 0 and Y 1 The easiest way to achieve this uncorrelatedness is through random assignment of treatment.

Another way of looking at it With some algebra, it can be shown that:

Another way of looking at it (in words) There are two sources of bias that need to be eliminated from estimates of causal effects : o Baseline difference/Selection bias o Heterogeneous response to the treatment Most of the methods available only deal with selection bias.

“ Treatment on the Treated Average Treatment Effect is not always the parameter of interest… often, it is the average treatment effect for the treated that is of substantive interest:

Treatment on the Treated “No baseline difference between the treatment and comparison groups.” If we need to estimate Treatment on the Treated: Then the simple estimator consistently estimates Treatment on the Treated if: (2)

References Judea Pearl (2000): Causality: Models, Reasoning and Inference, Cambridge University press. (Book) Chapters 1, 5 and 7. Trygve Haavelmo (1944): “The probability approach in econometrics”, Econometrica 12, pp. iii-vi Arthur Goldberger (1972): “Structural Equations Methods in the Social Sciences”, Econometrica 40, pp Donald B. Rubin (1974): “Estimating causal effects of treatments in randomized and nonrandomized experiments”, Journal of Educational Psychology 66, pp Paul W. Holland (1986): “Statistics and Causal Inference”, Journal of the American Statistical Association 81, pp , with discussion.

Thank You

? Q & A ?