Impact Evaluation Methods. Randomized Trials Regression Discontinuity Matching Difference in Differences.

Slides:



Advertisements
Similar presentations
Designing an impact evaluation: Randomization, statistical power, and some more fun…
Advertisements

REGRESSION, IV, MATCHING Treatment effect Boualem RABTA Center for World Food Studies (SOW-VU) Vrije Universiteit - Amsterdam.
Review of Identifying Causal Effects Methods of Economic Investigation Lecture 13.
#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Regression Discontinuity 10/13/09. What is R.D.? Regression--the econometric/statistical tool social scientists use to analyze multivariate correlations.
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
The counterfactual logic for public policy evaluation Alberto Martini hard at first, natural later 1.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Regression Discontinuity 10/13/08. What is R.D.? Regression--the econometric/statistical tool social scientists use to analyze multivariate correlations.
Differences-in- Differences November 10, 2009 Erick Gong Thanks to Null & Miguel.
Public Policy & Evidence: How to discriminate, interpret and communicate scientific research to better inform society. Rachel Glennerster Executive Director.
The Fundamental Problem of Causal Inference Alexander Tabarrok January 2007.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
Making Impact Evaluations Happen World Bank Operational Experience 6 th European Conference on Evaluation of Cohesion Policy 30 November 2009 Warsaw Joost.
PAI786: Urban Policy Class 2: Evaluating Social Programs.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 AIM-CDD Using Randomized Evaluations.
Non Experimental Design in Education Ummul Ruthbah.
Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.
1-1 Copyright © 2015, 2010, 2007 Pearson Education, Inc. Chapter 23, Slide 1 Chapter 23 Comparing Means.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Non-Experimental Methods Florence Kondylis.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
CAUSAL INFERENCE Shwetlena Sabarwal Africa Program for Education Impact Evaluation Accra, Ghana, May 2010.
The days ahead Monday-Wednesday –Training workshop on how to measure the actual reduction in HIV incidence that is caused by implementation of MC programs.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Beyond surveys: the research frontier moves to the use of administrative data to evaluate R&D grants Oliver Herrmann Ministry of Business, Innovation.
Jean Yoon December 15, 2010 Research Design. Outline Causality and study design Quasi-experimental methods for observational studies –Covariate matching.
ECON 3039 Labor Economics By Elliott Fan Economics, NTU Elliott Fan: Labor 2015 Fall Lecture 21.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Public Policy Analysis ECON 3386 Anant Nyshadham.
Propensity Score Matching for Causal Inference: Possibilities, Limitations, and an Example sean f. reardon MAPSS colloquium March 6, 2007.
Public Policy Analysis ECON 3386 Anant Nyshadham.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Nigeria Impact Evaluation Community of Practice Abuja, Nigeria, April 2, 2014 Measuring Program Impacts Through Randomization David Evans (World Bank)
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
What is randomization and how does it solve the causality problem? 2.3.
Framework of Preferred Evaluation Methodologies for TAACCCT Impact/Outcomes Analysis Random Assignment (Experimental Design) preferred – High proportion.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Randomized Evaluations: Applications Kenny Ajayi September 22, 2008 Global Poverty and Impact Evaluation.
Randomized Assignment Difference-in-Differences
REBECCA M. RYAN, PH.D. GEORGETOWN UNIVERSITY ANNA D. JOHNSON, M.P.A. TEACHERS COLLEGE, COLUMBIA UNIVERSITY ANNUAL MEETING OF THE CHILD CARE POLICY RESEARCH.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Do European Social Fund labour market interventions work? Counterfactual evidence from the Czech Republic. Vladimir Kváča, Czech Ministry of Labour and.
Impact Evaluation Methods Randomization and Causal Inference Slides by Paul J. Gertler & Sebastian Martinez.
The Evaluation Problem Alexander Spermann, University of Freiburg 1 The Fundamental Evaluation Problem and its Solution SS 2009.
ENDOGENEITY - SIMULTANEITY Development Workshop. What is endogeneity and why we do not like it? [REPETITION] Three causes: – X influences Y, but Y reinforces.
Differences-in-Differences
Quasi Experimental Methods I
Differences-in-Differences
Quasi Experimental Methods I
Impact Evaluation Methods
Impact evaluation: The quantitative methods with applications
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
Comparing Two Means: Paired Data
Impact Evaluation Methods
1 Causal Inference Counterfactuals False Counterfactuals
Impact Evaluation Toolbox
Matching Methods & Propensity Scores
Impact Evaluation Methods: Difference in difference & Matching
Evaluating Impacts: An Overview of Quantitative Methods
Explanation of slide: Logos, to show while the audience arrive.
Class 2: Evaluating Social Programs
Comparing Two Means: Paired Data
Class 2: Evaluating Social Programs
Applying Impact Evaluation Tools: Hypothetical Fertilizer Project
Presentation transcript:

Impact Evaluation Methods

Randomized Trials Regression Discontinuity Matching Difference in Differences

The Goal Causality We did program X, and because of it, Y happened.

The Goal Causal Inference Y happened because of X, not for some other reason. Thus it makes sense to think that if we did X again in a similar setting, Y would happen again.

Getting to Causality In a more research-friendly universe, we’d be able to observe a single person (call him Fred) after we both gave and didn’t give him the treatment. Y treated Fred -Y untreated Fred

Getting to Causality In the reality-based community, finding this Y treated Fred -Y untreated Fred “counterfactual” is impossible. Is the solution to get more people?

Getting to Causality With more people, we can calculate Average (treated)-Average(untreated). But what if there’s an underlying difference between the treated and untreated?

Getting to Causality Confounding Factors/Selection Bias/Omitted Variable Bias Textbook Example: If textbooks were deliberately given to the most needy schools, the simple difference is incorrect. If textbooks were already present in the schools where parents cared a lot about education, the simple difference is incorrect.

Problem Solved If we randomize the treatment, on average, treatment and control groups should be the same in all respects, and there won’t be selection bias. Check that it’s true for all observables. Hope that it’s therefore true for all unobservables.

Math You’d Rather Not See See Clair’s slides from September 15 -omitted variable bias Very accessible reading from same week by Duflo, Glennerster & Kremer. -selection bias

Randomization Randomize who gets treated. Check if it came out OK. Basically, that’s it.

Randomization Examples: Progresa-Cash if kids go to school Moving to Opportunity-voucher to move to better neighborhood Fertilizer & Hybrid Seed Loan maturity & Interest rate Deworming

Regression Discontinuity Being involved in a program is clearly not random. Smarter kids get get scholarships. Kids in smaller classes learn better. Big firms are more likely to unionize.

Regression Discontinuity Being involved in a program is clearly not random. Or is it? Scholarship cutoff +1 girl vs. scholarship cutoff-1 girl Isreali 41 kid school vs. Isreali 40 kid school Union-yes 50%+1 school vs. Union-yes 50% -1 school

Regression Discontinuity Being involved in a program is clearly not random. Or is it? Scholarship cutoff +1 girl vs. scholarship cutoff-1 girl Isreali 41 kid school vs. Isreali 40 kid school Union-yes 50%+1 school vs. Union-yes 50% -1 school

So how do we actually do this? 1.Draw two pretty pictures 1.Eligibility criterion (test score, income, or whatever) vs. Program Enrollment 2.Eligibility criterion vs. Outcome

So how do we actually do this? 2. Run a simple regression. (Yes, this is basically all we ever do, and the stats programs we use can run the calculation in almost any situation, but before we do it, it’s necessary to make sure the situation is appropriate and draw the graphs so that we can have confidence that our estimates are actually causal.) Outcome as a function of test score (or whatever), with a binary (1 if yes, 0 if no) variable for program enrollment.

As Good As Random, Sort Of Randomize who gets treated (within a bandwidth). Check if it came out OK (within a bandwidth). (within a bandwidth) Basically, that’s it (within a bandwidth).

Difference in Differences Change for the treated - Change for the control (t1-t0)-(c1-c0) t1-t0-c1+c0 t1-c1-t0+c0 t1-c1-(t0-c0) Which is the same as…

Examples Malaria Bleakley, Hoyt. Malaria Eradication in the Americas: A Retrospective Analysis of Childhood Exposure. Working paper. Land Reform Besley, Timothy and Robin Burgess. Land Reform, Poverty Reduction, and Growth: Evidence from India. Quarterly Journal of Economics. May 2000,

Matching Match each treated participant to one or more untreated participant based on observable characteristics. Assumes no selection on unobservables Condense all observables into one “propensity score,” match on that score.

Matching After matching treated to most similar untreated, subtract the means, calculate average difference

Matching Examples: Does piped water reduce diarrhea? Jalan, Jyotsna and Martin Ravallion. Does Piped Water Reduce Diarrhea for Children in Rural India? Journal of Econometrics. January 2003, Anti-poverty program in Argentina Jalan, Jyotsna and Martin Ravallion. Estimating the Benefit Incidence of an Antipoverty Program by Propensity Score Matching. Journal of Business and Economic Statistics. January 2003,

Matching Matching algorithm can be performed in many ways. Guido Imbens’ webpage

Summary The weakest (easiest) assumption is the best assumption. Randomization wins. Real scientists use it too.

Proof by One Example LaLonde, Robert. Evaluating the Econometric Evaluations of Training Programs with Experimental Data. American Economic Review, September Run a randomization and analyze it well. Then pretend you don’t have all the data that you do, construct fake comparison groups using the census, and show that none of your crazy methods get you right answer.