Prof. (FH) Dr. Alexandra Caspari Rigorous Impact Evaluation What It Is About and How It Can Be.

Slides:



Advertisements
Similar presentations
Mywish K. Maredia Michigan State University
Advertisements

#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
Good Evaluation Planning – and why this matters Presentation by Elliot Stern to Evaluation Network Meeting January 16 th 2015.
Designing Influential Evaluations Session 5 Quality of Evidence Uganda Evaluation Week - Pre-Conference Workshop 19 th and 20 th May 2014.
Assessing Program Impact Chapter 8. Impact assessments answer… Does a program really work? Does a program produce desired effects over and above what.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Research Design and Validity Threats
SOWK 6003 Social Work Research Week 4 Research process, variables, hypothesis, and research designs By Dr. Paul Wong.
Making Impact Evaluations Happen World Bank Operational Experience 6 th European Conference on Evaluation of Cohesion Policy 30 November 2009 Warsaw Joost.
TOOLS OF POSITIVE ANALYSIS
I want to test a wound treatment or educational program in my clinical setting with patient groups that are convenient or that already exist, How do I.
Research Design for Quantitative Studies
Quantitative Research Designs
Methodology Describe Context & setting Design Participants Sampling Power Analysi s Interventions Outcome (study variables) Data Collection Procedures.
TRANSLATING RESEARCH INTO ACTION What is Randomized Evaluation? Why Randomize? J-PAL South Asia, April 29, 2011.
Health Programme Evaluation by Propensity Score Matching: Accounting for Treatment Intensity and Health Externalities with an Application to Brazil (HEDG.
Are Disasters Any Different? Challenges and Opportunities for Post-Disaster Impact Evaluation Alison Buttenheim, Princeton University Howard White, 3ie.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Semester 2: Lecture 9 Analyzing Qualitative Data: Evaluation Research Prepared by: Dr. Lloyd Waller ©
Chapter 3 Research Methods Used to Study Child Behavior Disorders.
1 The Need for Control: Learning what ESF achieves Robert Walker.
Copyright ©2008 by Pearson Education, Inc. Pearson Prentice Hall Upper Saddle River, NJ Foundations of Nursing Research, 5e By Rose Marie Nieswiadomy.
Assisting GPRA Report for MSP Xiaodong Zhang, Westat MSP Regional Conference Miami, January 7-9, 2008.
Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.
Beyond surveys: the research frontier moves to the use of administrative data to evaluate R&D grants Oliver Herrmann Ministry of Business, Innovation.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Causal Inference Nandini Krishnan Africa Impact Evaluation.
1 Experimental Research Cause + Effect Manipulation Control.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
Impact Evaluations and Development Draft NONIE Guidance on Impact Evaluation Cairo Conference: Perspectives on Impact Evaluation Tuesday, March 31, 2009.
Ch. 2 Tools of Positive Economics. Theoretical Tools of Public Finance theoretical tools The set of tools designed to understand the mechanics behind.
1 Copyright © 2011 by Saunders, an imprint of Elsevier Inc. Chapter 8 Clarifying Quantitative Research Designs.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
Nigeria Impact Evaluation Community of Practice Abuja, Nigeria, April 2, 2014 Measuring Program Impacts Through Randomization David Evans (World Bank)
Chapter 3 should describe what will be done to answer the research question(s), describe how it will be done and justify the research design, and explain.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
What is randomization and how does it solve the causality problem? 2.3.
An introduction to Impact Evaluation and application to the Ethiopia NFSP Workshop on Approaches to Evaluating The Impact of the National Food Security.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Measuring Impact 1 Non-experimental methods 2 Experiments
Chapter 10 Finding Relationships Among Variables: Non-Experimental Research.
Diagram of a quasi-experimental design with two groups
1 Module 3 Designs. 2 Family Health Project: Exercise Review Discuss the Family Health Case and these questions. Consider how gender issues influence.
Evaluation Designs Adrienne DiTommaso, MPA, CNCS Office of Research and Evaluation.
Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Current practices in impact evaluation Howard White Independent Evaluation Group World Bank.
The Experiment Chapter 7. Doing Experiments In Everyday Life Experiments in psychology use the same logic that guides experiments in biology or engineering.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Randomized Assignment Difference-in-Differences
Evaluation Research Dr. Guerette. Introduction Evaluation Research – Evaluation Research – The purpose is to evaluate the impact of policies The purpose.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Outcomes Evaluation A good evaluation is …. –Useful to its audience –practical to implement –conducted ethically –technically accurate.
Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead Specialist Africa Impact Evaluation Initiative.
Impact Evaluation Methods Randomization and Causal Inference Slides by Paul J. Gertler & Sebastian Martinez.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
Monitoring and evaluation 16 July 2009 Michael Samson UNICEF/ IDS Course on Social Protection.
Evaluation What is evaluation?
1 An introduction to Impact Evaluation (IE) for HIV/AIDS Programs March 12, 2009 Cape Town Léandre Bassolé ACTafrica, The World Bank.
William M. Trochim James P. Donnelly Kanika Arora 8 Introduction to Design.
Chapter 9 Scrutinizing Quantitative Research Design.
Introduction to Impact Evaluation The Motivation Emmanuel Skoufias The World Bank PRMPR PREM Learning Week: April 21-22, 2008.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Impact evaluation of actions for jobseekers under the current OP ESF- Flemish Community : beyond classical parameters for success Expert Hearing.
Impact Evaluation Methods
Evaluating Impacts: An Overview of Quantitative Methods
Positive analysis in public finance
Presentation transcript:

Prof. (FH) Dr. Alexandra Caspari Rigorous Impact Evaluation What It Is About and How It Can Be Done In Practice Alexandra Caspari, Frankfurt/Main Germany Conference »Perspectives on Impact Evaluation: Approaches to Assessing Development Effectiveness« 31 st March – 2 nd April 2009, Cairo

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 1 University of Applied Sciences Historical Review – The Evaluation Gap  MDGs (2000), ‘Paris Declaration on Aid Effectiveness’ (2005), and ‘Agenda for Action’ (Accra, 2008):  Increasing attention to Impact Evaluations  Lack of knowledge about effectiveness of projects and programs  2006: Report “When will we ever learn?” of the CGD ‘Evaluation Gap Working Group’  gap in quantity and quality of impact evaluations: -too few impact evaluations are being carried out and -those conducted often unable to properly assess impact because of methodological shortcomings  Recommendation: ‘Collective Action’  International Initiatives (NONIE, 3IE, …)

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 2 University of Applied Sciences What is Impact Evaluation?  OECD/DAC (2002): “positive and negative, primary and secondary long-term effects produced by a development intervention, directly or indirectly, intended or unintended”  emphasises on ‘produced by’: -measures impact with clear causation (causal attribution) -considers the counterfactual, i.e. the question “What difference did this program make?” “What would have happened without the intervention?” Rigorous Impact Evaluation (RIE):  Distinction against more “usual evaluations” by adding “rigorous”  focus on clear causation  use of adequate methods (to meet methodological shortcomings)  most important point: selection of the evaluation design to consider the counterfactual

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 3 University of Applied Sciences The Counterfactual  Causal effect: An actual effect δ i caused by a treatment T (a program) is the difference between the outcome Y i1 under a treatment T (T=1), i.e. program participant, minus the alternative outcome Y i0 that would have happened without the treatment T (T=0), i.e. non-participant :  Impact is not directly observable: -one can observe any given individual either as a treated person (participant) or untreated person (non-participant) but not both states -if individual i is participating in a program (T=1), then the outcome Y i0 is unobservable -this unobservable outcome Y i0 is called counterfactual  Analyzing the difference between the observed outcome and the unobserved potential outcome by choosing the best evaluation design

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 4 University of Applied Sciences Considering the Counterfactual  often used non-experimental designs: ● : observation, P: participants (treated), t: time (first, second observation), X: project intervention one-group pre-test post-test design (a) time t2t2 t1t1 P measured impact impact indicator  measured impact =  the counterfactual is not considered!  with non-experimental designs causal attribution is not possible!

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 5 University of Applied Sciences Considering the Counterfactual  necessary: experimental or quasi-experimental designs  adequate comparison group (‘with-and without comparison’)  „Real“ Experiments / Randomized Controlled Trials (RCTs): (Laboratory)Experiments: -random assignment of individuals to treatment (P) and control group (C)  groups differ solely due to chance -treatment and conditions are known/checkable Field experiments: -take place in real-world settings -anyhow treatment and control groups are assigned at random  Quasi-Experiments: -no random assignment -has a source of randomization that is “as if” randomly assigned -control group is often reconstructed ex-post

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 6 University of Applied Sciences Considering the Counterfactual ● : observation, P: participants (treaded), C: control group (non-treated), D: difference, t: time (first, second observation), X: project intervention over- estimated impact one-group pre-test post-test design (a) time t2t2 t1t1 P measured impact impact indicator static group comparison (4) time impact indicator t2t2 t1t1 C P measured impact = D t2 (single difference) time impact indicator t2t2 t1t1 C P D t1 pre-test post-test control group design (1)/(2) (double difference) measured impact = D t2 – D t1 D t2

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 7 University of Applied Sciences Approaches to Impact Evaluation  appropriate impact evaluation designs are often reject as unnecessarily sophisticated or because of ethical concerns  various realistic ways in which quasi-experimental designs can be introduced in an ethically and politically acceptable manner: -Matching on Observables -Regression Discontinuity -Propensity Score Matching (PSM) -Pipeline Approach -Multiple Comparison Group Design

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 8 University of Applied Sciences Possible Approaches in Practice  Matching on Observables: -characteristics (access tor services, economic level, type of housing, etc.) on which the comparison group should match the program group (individuals, households or areas) are identified carefully -often easily observable or identifiable characteristics -unobservable differences has to be kept in mind -control group is build out of those individuals, households or areas which match best -quasi-experimental design “pretest-posttest-comparison with post- test non-equivalent control group” (3) or at least “static group comparison” (4) is possible  single-difference (SD) possible

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 9 University of Applied Sciences Possible Approaches in Practice  Regression Discontinuity: -if a program is assigned using a clear threshold for eligibility comprised for one ore more criteria (age, income less than…) -control group is built out of those just above the threshold and hence not eligible for the program -those individuals will have comparable characteristics -quasi-experimental design “pre-test post-test non-equivalent control group design” (2) possible!  double-difference (DD) possible!

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 10 University of Applied Sciences Possible Approaches in Practice  Pipeline Approach: -if large programs (housing or community infrastructure, immunization, …) are introduced in phases over several years -when there are no major differences between the characteristics of families, communities scheduled for each phase and -when there is no selection criteria for participants of the first phase (the poorest families, communities, …)  participants of phase 2 & 3 = control group for participants phase 1  quasi-experimental design “pre-test post-test non-equivalent control group design” (2) possible!  double-difference (DD) possible

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 11 University of Applied Sciences Important Remarks  The international discussion about RIE refers just to a small aspect of evaluation: the causal attribution of impact  Impact is measured at the level of target groups/participants  because target groups are typically large, for this evaluation step quantitative methods are necessary (representativeness vs. profundity)  other evaluation methods are not condemned!  causal attribution is necessary but not sufficient  ‘black box’ remains: why does a program have impact (or does not)  comprehensive meaningful and reliable impact evaluations need the use of mixed method, i.e. use of quantitative and qualitative methods

Fachhochschule Frankfurt am Main – Alexandra Caspari, 31/03/2009, CairoSlide 12 University of Applied Sciences  Reference: Caspari, Alexandra/Barbu, Ragnhild (2008): Wirkungsevaluierungen Zum Stand der internationalen Diskussion und dessen Relevanz für die Evaluierung der deutschen Entwicklungszusammenarbeit