OVE’s Experience with Impact Evaluations Paris June, 2005.

Slides:



Advertisements
Similar presentations
1 Table design Module 3 Session 2. 2 Objectives of this session By the end of this session, you will be able to: appreciate the different type of objectives.
Advertisements

Designing an impact evaluation: Randomization, statistical power, and some more fun…
The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Explanation of slide: Logos, to show while the audience arrive.
Mywish K. Maredia Michigan State University
Research Methodology For reader assistance, have an introductory paragraph in which attention is given to the organization of the section in relation to.
Introduction and the Context The Use and value of Urban Planning.
Estimation and Reporting of Heterogeneity of Treatment Effects in Observational Comparative Effectiveness Research Prepared for: Agency for Healthcare.
Incorporating considerations about equity in policy briefs What factors are likely to be associated with disadvantage? Are there plausible reasons for.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Basic Business Statistics.
GENEVA JULY, 2015 The limits of cash transfer in addressing child labour.
Poverty: Facts, Causes and Consequences Hilary Hoynes University of California, Davis California Symposium on Poverty October 2009.
Rural Poverty and Hunger (MDG1) Kevin Cleaver Director of Agriculture and Rural Development November 2004.
Measuring Progress: Strategies for Monitoring and Evaluation Rebecca Stoltzfus.
The Impact of Court Decentralization on Domestic Violence Against Women Raúl Andrade Jimena Montenegro March 2009.
RESEARCH A systematic quest for undiscovered truth A way of thinking
Non Experimental Design in Education Ummul Ruthbah.
Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.
5110 Zeller Guidelines for research proposal
The Impact of Air Pollution on Infant Mortality: Evidence from Geographic Variation in Pollution Shocks Induced by a Recession Kenneth Y. Chay and Michael.
T tests comparing two means t tests comparing two means.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Overview of Evaluation Designs. Learning objectives By the end of this presentation, you will be able to: Explain evaluation design Describe the differences.
Tips for Researchers on Completing the Data Analysis Section of the IRB Application Don Allensworth-Davies, MSc Statistical Manager, Data Coordinating.
Targeted Interventions in Health Care: The case of PROMIN Sebastian Galiani Mercedes Fernandez Ernesto Schargrodsky.
Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.
Beyond surveys: the research frontier moves to the use of administrative data to evaluate R&D grants Oliver Herrmann Ministry of Business, Innovation.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department.
Directorate for Food, Agriculture, and Fisheries 1 ORGANISATION FOR ECONOMIC CO-OPERATION AND DEVELOPMENT ORGANISATION DE COOPÉRATION ET DE DEVELOPMENT.
Evaluation Proposal Defense Observations and Suggestions Yibeltal Kiflie August 2009.
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
Evaluation Designs Adrienne DiTommaso, MPA, CNCS Office of Research and Evaluation.
Framework of Preferred Evaluation Methodologies for TAACCCT Impact/Outcomes Analysis Random Assignment (Experimental Design) preferred – High proportion.
Current practices in impact evaluation Howard White Independent Evaluation Group World Bank.
T tests comparing two means t tests comparing two means.
Comments on: The Evaluation of an Early Intervention Policy in Poor Schools Germano Mwabu June 9-10, 2008 Quebec City, Canada.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
GOVERNMENT OF THE KINGDOM OF LESOTHO Water and Sewerage Company (WASCO) Greater Maseru Water Supply Feasibility Study & Preliminary Design Results of Socio-Economics.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
What is Impact Evaluation … and How Do We Use It? Deon Filmer Development Research Group, The World Bank Evidence-Based Decision-Making in Education Workshop.
Research Design Quantitative. Quantitative Research Design Quantitative Research is the cornerstone of evidence-based practice It provides the knowledge.
Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead Specialist Africa Impact Evaluation Initiative.
IMPLEMENTATION AND PROCESS EVALUATION PBAF 526. Today: Recap last week Next week: Bring in picture with program theory and evaluation questions Partners?
Conditional Cash Transfer Pilot Upper Egypt. Background Currently Egypt has a welfare regime that is comprehensive and generous but which has some limitations.
Assessing the Impact of Informality on Wages in Tanzania: Is There a Penalty for Women? Pablo Suárez Robles (University Paris-Est Créteil) 1.
Chapter 23: Overview of the Occupational Therapy Process and Outcomes
The SPF-I: towards a joint position on Social Protection in Europe and Central Asia SHEILA MARNIE UNDP.
1 TARGETING HEALTH INSURANCE TO THE POOR IN COLOMBIA By Tarsicio Castañeda Reaching the Poor Conference The World Bank, February 18-20, 2004.
Writing a sound proposal
Regression to Predict Absences on MAP Scores
Research Department Inter-American Development Bank
Measuring Results and Impact Evaluation: From Promises into Evidence
Commercial Agriculture Development Project Impact Evaluation
TRACER STUDIES—Assessments and Evaluations
Clinical Studies Continuum
Don’t waste a good investment
Presentation at the African Economic Conference
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
Resource Mobilisation Strategy for Slum Upgrading
Matching Methods & Propensity Scores
Evaluating Impacts: An Overview of Quantitative Methods
Sampling for Impact Evaluation -theory and application-
Research Design Quantitative.
An Introduction to Evaluating Federal Title Funding
Title Team Members.
Presentation transcript:

OVE’s Experience with Impact Evaluations Paris June, 2005

Impact Evaluations  Alternative definitional models: –time elapsed since intervention –Counterfactual comparison  OVE adopted the counterfactual approach, and further limited the initial sample to programs with partial coverage.  Partial coverage allows observation of treatment effects through comparison of treated and untreated groups

Policy  The general evaluative question proposed by the IDB’s ex post policy is “… the extent to which the development objectives of IDB-financed projects have been attained.”  This questions is most convincingly answered through treatment effect evaluations

Selecting Projects  Random selection is appropriate for accountability-oriented evaluations  Purposive selection of projects of similar design across countries is better for generating learning regarding the model underlying the interventions  Clusters of like projects permit meta-evaluations of models

Projects Selected  Chose purposive cluster sampling strategy but some stand-alone projects. A total of 16 projects were selected  Clusters: (i) Neighborhood Improvement projects and (ii) Land Titling Projects.  Stand-alone: Cash Transfer (Argentina), Potable Water (Ecuador); Agricultural Subsidies and Cash Transfer Programs (Mexico’s Procampo and Opportunities programs); Social Investment Fund (Panama)  Stand-alones serve as pilots for future clusters

Both Performance Monitoring and Treatment Effect Are Required Treatment Effect includes randomized design; propensity score matching, controlled comparison, discontinuous regressions.

Limits to Treatment Effect Evaluations

Experience  The required information supposedly generated through standardized performance monitoring is absent in a large majority of IDB projects examined  10 of the 16 selected projects had inadequate data for treatment effect evaluation  6 of the 16 could be retrofitted with sufficient data to attempt a treatment effect evaluation  Retrofitting implied significant data collection costs, costs that could have been avoided had adequate performance monitoring been in place over the life of the project.

The Bank’s Current Portfolio Of 593 active projects in mid -2004:  97 (16%) claim existence of information for at least one development outcome, of which  27 have the information in an electronic form, of which for  5 the information is held in the Bank, of which  2 appear to be collecting data for treatment effect evaluation

Experience: Limits to Retrofitting  The questions answered are dependent on the information found rather than on the relevance and usefulness of the hypotheses being tested: the tail wagging the dog.  It severely limits the set of control variables’ information thus reduces the veracity of treatment effect findings  retrofitted data may not correspond to the development outcomes declared by the projects. A project can be evaluated using intended and unintended effects, but should at least consider as a minimum the intended ones.

Experience Confirms the Value of Treatment Effect Evaluation WaterSewerRubbishIlliteracyIncomeRentChild Mortality Homicide Rate Naive Treatment In just one project (Neighborhood Improvement, Rio-Brazil) comparing naïve and treatment effect the following held regarding naïve and treatment effects: positive/negative, negative/positive; greater/smaller; smaller/greater; and the same.

So pictures need to be interpreted with caution BeforeAfter

Experience  “…the six treatment effect evaluations undertaken during 2004 do show that the Bank’s interventions have a significant development effect for at least one declared development objective. These findings suggest that the Bank may be currently understating its contribution to development.”

EXPERIENCE: Findings Land Titling  “ Beneficiaries of Land Regularization projects saw property values for their land increase …. However, for the other purported development effects (greater productivity, increased investment, and greater access to credit), no unambiguous treatment effects were found.  Ramifications for project design: for small and poor producers to benefit from a pro-market regime, titling alone is not sufficient  Transaction costs and market distortions that limit access to credit must be also simultaneously be addressed

EXPERIENCE: Findings Potable Water  heterogeneity of results important. a regressive relationship between treatment effect and income, where more educated (and wealthier) households did better than less educated (and poorer) households  Ramification for project design: projects should include or be coordinated with, as a hypothesis to be tested, a health education component together with potable water expansion Bottom 25%25%-50%50%-75%Top 25% Expenditure level proportional change All SampleAt least Primary Impact on infant mortality

EXPERIENCE: Findings from cash transfer and agricultural subsidy programs Issue: Do conditions attached to cash transfers produce more change than the transfers alone  Income effect alone may be substantial, and conditionalities are costly to administer and monitor  In a comparison between two programs in Mexico with and without conditionality the following ramifications for project design were found: –Conditionality (school and clinic attendance) does result in an effect over and above the income effect of the transfer. –Transfers to the mother as opposed to the father matters as the effects are greater when the transfer is to the mother

EXPERIENCE: Low Costs Treatment effect evaluations can be done inexpensively, if attention is paid to data at the time of design and during implementation Data collection costs can be substantial if retrofitted, but still within reasonable limits. Costs ranged from $28,000 to $92,000 per evaluation, much lower than the “norm”: small budget high returns

Summary  Initial experience with treatment effect impact evaluations provided considerable knowledge relevant for future project design  Costs were moderate, and can be expected to be lower in the future if the performance monitoring system is improved  Data has value to researchers, and cost-sharing in data collection was possible in several cases  Treatment effect evaluation provides the only convincing basis for asserting development effectiveness