An introduction to Impact Evaluation and application to the Ethiopia NFSP Workshop on Approaches to Evaluating The Impact of the National Food Security.

Slides:



Advertisements
Similar presentations
Armenias Millennium Challenge Account: Assessing Impacts Ken Fortson, MPR Ester Hakobyan, MCA Anahit Petrosyan, MCA Anu Rangarajan, MPR Rebecca Tunstall,
Advertisements

An impact evaluation of Ethiopias Food Security Program John Hoddinott, IFPRI (in collaboration with Dan Gilligan, Alemayehu Seyoum and Samson Dejene)
Designing an impact evaluation: Randomization, statistical power, and some more fun…
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Overview M&E Capacity Strengthening Workshop, Maputo 19 and 20 September 2011.
Knowing if the RBF mechanism is working Incorporating Rigorous Impact Evaluation into your HRBF program Sebastian Martinez World Bank.
Mywish K. Maredia Michigan State University
#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
Project Monitoring Evaluation and Assessment
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
TOOLS OF POSITIVE ANALYSIS
SAMPLING AND STATISTICAL POWER Erich Battistin Kinnon Scott Erich Battistin Kinnon Scott University of Padua DECRG, World Bank University of Padua DECRG,
What do we know about gender and agriculture in Africa? Markus Goldstein Michael O’Sullivan The World Bank Cross-Country Workshop for Impact Evaluations.
Ethiopia Productive Social Safety Net. Program description This program aims to provide –Predictable, multi-year assistance to –chronically the food insecure.
Experiments and Observational Studies.  A study at a high school in California compared academic performance of music students with that of non-music.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 AIM-CDD Using Randomized Evaluations.
ISTEP: Technology Field Research in Developing Communities Instructor: M. Bernardine Dias CAs: Sarah Belousov and Ermine Teves Spring 2009.
Copyright © 2010 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Experiments and Observational Studies. Observational Studies In an observational study, researchers don’t assign choices; they simply observe them. look.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 13 Experiments and Observational Studies.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Non-Experimental Methods Florence Kondylis.
Evaluation methods and tools (Focus on delivery mechanism) Jela Tvrdonova, 2014.
Measuring Impact: Experiments
Use of impact evaluation for operational research PREM Week 2007 Arianna Legovini, Africa Impact Evaluation Initiative (AFTRL)
Private involvement in education: Measuring Impacts Felipe Barrera-Osorio HDN Education Public-Private Partnerships in Education, Washington, DC, March.
Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.
PPA 502 – Program Evaluation Lecture 2c – Process Evaluation.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
Impact Evaluation for Evidence-Based Policy Making
What is randomization and how does it solve the causality problem? 2.3.
AADAPT Workshop for Impact Evaluation in Agriculture and Rural Development Goa, India 2009 With generous support from Gender Action Plan.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Measuring Impact 1 Non-experimental methods 2 Experiments
Africa RISING M&E Expert Meeting Addis Ababa, 5-7 September 2012.
Development Impact Evaluation in Finance and Private Sector 1.
Can a Market-Assisted Land Redistribution Program Improve the Lives of the Poor? Evidence from Malawi Gayatri Datar (World Bank, IEG) Ximena V. Del Carpio.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Implementing an impact evaluation under constraints Emanuela Galasso (DECRG) Prem Learning Week May 2 nd, 2006.
Randomized Assignment Difference-in-Differences
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Prof. (FH) Dr. Alexandra Caspari Rigorous Impact Evaluation What It Is About and How It Can Be.
Chapter 9 Audit Sampling – Part a.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Randomization.
Monitoring and evaluation Objectives of the Session  To Define Monitoring, impact assessment and Evaluation. (commonly know as M&E)  To know why Monitoring.
Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead Specialist Africa Impact Evaluation Initiative.
Do European Social Fund labour market interventions work? Counterfactual evidence from the Czech Republic. Vladimir Kváča, Czech Ministry of Labour and.
Impact Evaluation Methods Randomization and Causal Inference Slides by Paul J. Gertler & Sebastian Martinez.
Common Pitfalls in Randomized Evaluations Jenny C. Aker Tufts University.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
Monitoring and evaluation 16 July 2009 Michael Samson UNICEF/ IDS Course on Social Protection.
Introduction to Impact Evaluation The Motivation Emmanuel Skoufias The World Bank PRMPR PREM Learning Week: April 21-22, 2008.
Operational Aspects of Impact Evaluation
Impact Evaluation for Real Time Decision Making
Measuring Results and Impact Evaluation: From Promises into Evidence
Right-sized Evaluation
Explanation of slide: Logos, to show while the audience arrive.
Impact evaluation: The quantitative methods with applications
Development Impact Evaluation in Finance and Private Sector
Implementation Challenges
Randomization This presentation draws on previous presentations by Muna Meky, Arianna Legovini, Jed Friedman, David Evans and Sebastian Martinez.
Evaluating Impacts: An Overview of Quantitative Methods
Randomization This presentation draws on previous presentations by Muna Meky, Arianna Legovini, Jed Friedman, David Evans and Sebastian Martinez.
Sampling for Impact Evaluation -theory and application-
Operational Aspects of Impact Evaluation
Applying Impact Evaluation Tools: Hypothetical Fertilizer Project
Presentation transcript:

An introduction to Impact Evaluation and application to the Ethiopia NFSP Workshop on Approaches to Evaluating The Impact of the National Food Security Program, Addis Ababa, January 16, 2006 Arianna Legovini, The World Bank

ET NFSP Arianna Legovini2 Content 1. 1.Context 2. 2.Motivation 3. 3.Defining impact 4. 4.The evaluation problem 5. 5.Problems with naïve evaluations 6. 6.Application to the NFSP

ET NFSP Arianna Legovini3 1. Context  Most social programs are targeted to a specific group of the population.  Within that group some units (individuals, households, communities) get the program and others do not, or some get it at the beginning of program implementation and others get it later, … depending on budget constraints, implementation capacity, and targeting efficiency.

ET NFSP Arianna Legovini4 2. Motivating rigorous impact evaluation  Monitoring tells us whether things are moving in the right direction but cannot tell us the cause of those movements  Rigorous impact evaluation causally links activities to results.  There are three main functions for impact evaluation:   It is a political tool to manage public opinion   It is a fiscal tool to allocate budget resources more efficiently, and   It is a management tool to improve program design, results and sustainability

ET NFSP Arianna Legovini5 Political tool  Public opinion can easily shift in response to few publicized failures  Empirical evidence on program impacts give governments the ability to effectively manage public opinion (media, parliament)

ET NFSP Arianna Legovini6 Fiscal tool  Shift in budget allocation are often made on hunches of what might deliver more results  Impact evaluation turns hunches into quantitative estimates of the cost- effectiveness of alternative programs in reaching desired objectives  It provides the basis for shifting resources towards more effective programs

ET NFSP Arianna Legovini7 Management tool  This is where impact evaluation can make huge contributions  By testing alternative ways of doing things within the same program, impact evaluation can measure the relative effectiveness of alternatives  Feedback can be used to modify program design during implementation to ensure success.  It is the ultimate managing-by-results tool.

ET NFSP Arianna Legovini8 3. Defining Impact  In the language of the impact evaluation profession, the impact of a program is the amount of change in any outcome (short, medium or long-term) which is caused by the program.  Impact can be measured at any time:  Even just after program announcement (e.g. change in expectations may lead to changes in asset prices even before a program is implemented), or  Any time after implementation (e.g. some changes are instantaneous and some take time to realize BUT if nothing moves in the short-run, you can rest assured nothing will move in the long run either)

ET NFSP Arianna Legovini9 4. The evaluation problem  The impact of the program is the difference between the outcome with the program and the outcome without the program.  BUT we only observe outcomes with OR without the program, not both at the same time.  This means that the problem of measuring impact is a problem of missing data—not because we do not collect them, but because they don’t exist.

ET NFSP Arianna Legovini10  Impact evaluation is a set of rigorous analytical methods applied to solve the missing data problem by building a counterfactual and measuring observed outcomes against it.

ET NFSP Arianna Legovini11 Observed changes and changes against a counterfactual

ET NFSP Arianna Legovini12 Solving the missing data problem  Need to build a control or comparison group that looks the same at time zero as the group getting the program (treatment group)  What does “look the same” mean? It means that, on average, the comparison and treatment group only differ by the fact that one gets the program and the other does not.

ET NFSP Arianna Legovini13 Building a control Experimental approach : Randomization  Give all units in the eligible population an equal chance to participate; and compare those who win the lottery with those who did not win the lottery  Give all units in the eligible population an equal chance to be included in either phase I or phase II of the program; and then use units in phase II as controls for phase I

ET NFSP Arianna Legovini14 Building a comparison Non-experimental approach :  Matching: match characteristics of participants and non-participants at time zero; and then compare changes in outcomes in treated population against changes in outcomes in matched observations  Regression discontinuity: assume that because of measurement error, units just above or just below the eligibility threshold are the same on average; and compare participants just below the threshold with non-participants just above the threshold  Pipeline comparisons: compare current participants to prospective participants  Instrumental variables: identify exogenous variable that influences participation but not outcomes to identify exogenous variation in outcomes due to the program

ET NFSP Arianna Legovini15 Estimating impact  Single-difference (outcome of treatment minus outcome of control at time t+1) can be used in experimental setting with sufficiently large samples  Double-difference (before and after and against a control) is used in experimental setting with small samples and non- experimental settings to control for selection bias

ET NFSP Arianna Legovini16 Data requirements  Surveys with representative samples of beneficiaries and non-beneficiaries  Questionnaires include outcomes of interest (consumption, income, assets etc), questions about the program in question and questions about other programs

ET NFSP Arianna Legovini17 Data requirements  Experimental approach with large samples only requires survey at time t+1 (single diff). In this case baseline is needed only if we are interested in measuring targeting efficiency.  Experimental with small sample and non- experimental generally require baseline and follow up panel data (double diff)

ET NFSP Arianna Legovini18 5. Problems with naive evaluation techniques  Comparisons of beneficiaries before and after: outcomes change for many reasons, how to determine which portion of that change is due to the program?  Comparisons ex-post relative to the rest of the population: if program targets the worse off, comparisons ex-post will compare beneficiaries with non-beneficiaries that were initially somewhat better off and wrongly conclude program did not have positive impacts

ET NFSP Arianna Legovini19 6. Application to the National Food Security Program  The NFSP M&E plan:  Is a good monitoring plan using administrative data with a good logical framework  However, it has no plan for assessing targeting efficiency and the plan for evaluation is flawed (no comparison)  The sampling for the baseline only surveys beneficiaries and the questionnaire does not permit to estimate consumption

ET NFSP Arianna Legovini20 Current NFSP evaluation plan  Plan to track changes in outcomes before and after and against (unspecified) standards  BUT program outcomes can improve or worsen because of e.g. the weather. These effects might be much larger and completely overshadow the effects of the program. Adverse weather will have us conclude that the program is ineffective irrespectively

ET NFSP Arianna Legovini21 Strengthening NFSP impact evaluation  Define for which activities we want to estimate impact and ask the right questions to improve program overtime  Implement framework to answer these questions by adding variation of treatment of beneficiaries  Build comparison groups via randomized experiments or the estimation of a counterfactual using statistical methods  Expand survey sample to include comparison, include a consumption module in the questionnaire, and collect panel data  Estimate impacts using double difference estimators

ET NFSP Arianna Legovini22 Impact of which NFSP activities?  Productive safety-nets:  Public works  Workfare/cash or food transfers  Unconditional transfers to labor-poor households  Resettlement  Other food security activities  Financial services  Agricultural extension  Water  Land conservation, etc.

ET NFSP Arianna Legovini23 Opportunities to find a comparison  Use phase-in of the program to compare current beneficiaries with prospective beneficiaries since not all 15 M are beneficiaries at this time (pipeline comparison)  Compare woredas or kebeles with the program with similar woredas or kebeles without the program (matching)  Compare beneficiary households that barely qualified for the program with non-beneficiary households that barely did not (regression discontinuity design)

ET NFSP Arianna Legovini24 Data and analysis  The questionnaire should include questions to capture the outcomes of the program.  Currently the questionnaire includes questions on assets, savings and credit, and agricultural activities  Including a consumption module would provide the basis for measuring welfare and graduation in a reliable and quantitative manner.  Collect panel data (baseline and follow up survey on the same sample of beneficiaries and non-beneficiaries).  Estimate impacts

ET NFSP Arianna Legovini25 Yet, the most compelling reason to do impact evaluation is to learn how to improve program effectiveness over time  This requires adding variation in program treatment to determine which mechanisms work best to achieve e.g. graduation

ET NFSP Arianna Legovini26 Examples of what can be learned: is farm only best?  Is a wider menu of farm and non-farm support more effective than farm support only?—Take a random sample of localities where we include support to non-farm activities; and compare results overtime.

ET NFSP Arianna Legovini27 More examples of what can be learned—how much is enough?  To secure household graduation, is it more effective to spread support thin across all eligible households (or communities) or to focus NFSP effort on a portion of the households (or communities) this year, another portion next year, etc?—randomly select a small group of beneficiaries, double the fiscal effort on these households, and monitor whether graduation rates more than double or not

ET NFSP Arianna Legovini28 More examples of what can be learned—credit pick up rates and effectiveness?  Does one size fit all? Would modifying the size of the credit package improve credit pick up rates?— randomly select some communities where household can choose the size of the credit; and compare with one size credit communities  To improve the effectiveness of credit, would a package of credit-training-technical assistance improve credit effectiveness? (evidence from the rest of the world says yes)—randomly select some communities where training and/or TA is attached to credit; and compare with credit only communities

ET NFSP Arianna Legovini29 More examples of what can be learned—Quality of productive community assets  Does the threat of audit significantly increase the quality of productive community assets?—take a random sample of communities and communicate intention to audit project implementation (Evidence from Indonesia says yes)

ET NFSP Arianna Legovini30 Concluding remarks  Credible impact estimates require a strengthened framework for evaluation. This requires taking the following steps:  Agreeing on the questions that can be asked  Designing a framework that can answer these questions (program delivery, evaluation method, sampling, questionnaire)  Implementing the framework, analyzing and reporting periodically  Improving program overtime based on evidence