Presentation is loading. Please wait.

Presentation is loading. Please wait.

Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.

Similar presentations


Presentation on theme: "Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14."— Presentation transcript:

1

2 Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14

3 I am cutting rocks http://img359.imageshack.us/img359/7104/picture420bt2.jpg The essence of theory of change – linking activities to intended outcomes I am building a temple

4 Theory of change “the process through which it is expected that inputs will be converted to expected outputs, outcome and impact” DfID Further Business Case Guidance “Theory of Change”

5 Theory of change Start with a RESULTS CHAIN

6 The results chain: tips Activities OutputsOutcomes We produce Influence Contribute to We control Clients Clients control We are accountable for We expectShould occurr 100% attribution Some attribution Partial attribution Readily changed Less flexibility to change Long term Delivered annually By end of program Long-term

7 Monitoring – activities and outputs

8 Personal Monitoring Tools

9 No monitoring - blind and deaf

10 Monitoring and Evaluation MonitoringEfficiency  Measures how productively inputs (money, time, personnel, equipment) are being used in the creation of outputs (products, results)  An efficient organisation is one that achieves its objectives with the least expenditures of resources Evaluation Effectiveness  Measures the degree to which results / objectives have been achieved  An effective organisation is one that achieves its results and objectives

11 10 All Most Some Inputs Outputs Outcomes MONITORING focused on project process (per individual project) EVALUATION focused on effectiveness of project process (for many projects)  Resources  Staff  Funds  Facilities  Supplies  Training  Project deliverables achieved  “Count” (quantified) what has been done  Short and intermediate effects.  Long term effects and changes

12 Resist temptation, there must be a better way! Clear objectives Few key indicators Quick simple methods Existing data sources Participatory method Short feed-back loops Action results!

13 Monitoring/Evaluation objectives must be SMART Specific Measurable Achievable Realistic Timed (see 10 Easy Mistakes, page 5)

14 Evaluation: who evaluates whom? The value of a joint approach

15 The Logical Chain 1.Define Objectives (and Methodology) 2.Supply Inputs 3.Achieve Outputs 4.Generate Outcome 5.Identify and Measure Indicators 6.Evaluate by comparing Objectives with Indicators 7.Redefine Objectives (and Methodology)

16 www.brac.net Impact Evaluation An assessment of the causal effect of a project, program or policy beneficiaries. Uses a counterfactual… Impacts = Outcomes - What would have happened anyway

17 www.brac.net When to use Impact Evaluation? Evaluate impact when project is: Innovative Replicable/ scalable Strategically relevant for reducing poverty Evaluation will fill knowledge gap Substantial policy impact Use evaluation within a program to test alternatives and improve programs

18 www.brac.net Impact Evaluation Answers What was the effect of the program on outcomes? How much better of the beneficiaries because of the program policies? How would outcome change if changed program design? Is the program cost-effective?

19 www.brac.net Different Methods to measure impact evaluation  Randomised Assignment – experimental Non Experimental:  Matching  Difference-in-Difference  Regression Discontinuity Design  Instrumental Variable / Random Promotion

20 www.brac.net Randomization The “gold standard” in evaluating the effects of interventions It allows us to form a “treatment” and “control” groups – identical characteristics – differ only by intervention Counterfactual: randomized-out group

21 www.brac.net Matching Matching uses large data sets and heavy statistical techniques to construct the best possible artificial comparison group for a given treatment group. Selected basis of similarities in observed characteristics Assumes no selection bias based on unobservable characteristics. Counterfactual: matched comparison group

22 www.brac.net Difference-in-difference Compares the change in outcomes overtime between the treatment group and comparison group Controls for the factors constant overtime in both groups ‘parallel trends’ in both the groups in the absence of the program Counter-factual: changes over time for the non- participants

23 www.brac.net DesignWhen to use Randomization ► Whenever possible ► When an intervention will not be universally implemented Random Promotion ► When an intervention is universally implemented Regression Discontinuity ► If an intervention is assigned based on rank Diff-in-diff ► If two groups are growing at similar rates Matching ► When other methods are not possible ► Matching at baseline can be very useful Uses of Different Design

24 www.brac.net Qualitative and Quantitative Methods Qualitative methods focus on how results were achieved (or not). They can be very helpful for process evaluation. It is often very useful to conduct a quick qualitative study before planning an experimental (RCT) study.

25 Thank you!


Download ppt "Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14."

Similar presentations


Ads by Google