Designing a Random Assignment Social Experiment In the U.K.; The Employment Retention and Advancement Demonstration (ERA)

Slides:



Advertisements
Similar presentations
Active labour market measures and entrepreneurship in Poland Rafał Trzciński Impact Evaluation Spring School Hungary,
Advertisements

Knowing if the RBF mechanism is working Incorporating Rigorous Impact Evaluation into your HRBF program Sebastian Martinez World Bank.
#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
The Bayh-Dole Act of 1980: Policy Model for Other Industrial Economies? David C. Mowery Haas School of Business U.C. Berkeley & NBER Bhaven N. Sampat University.
Designing Influential Evaluations Session 5 Quality of Evidence Uganda Evaluation Week - Pre-Conference Workshop 19 th and 20 th May 2014.
Unit 1 Section 1.3.
The French Youth Experimentation Fund (Fonds d’Expérimentation pour la Jeunesse – FEJ) Mathieu Valdenaire (DJEPVA - FEJ) International Workshop “Evidence-based.
Assessing Program Impact Chapter 8. Impact assessments answer… Does a program really work? Does a program produce desired effects over and above what.
4.11 PowerPoint Emily Smith.
VIII Evaluation Conference ‘Methodological Developments and Challenges in UK Policy Evaluation’ Daniel Fujiwara Senior Economist Cabinet Office & London.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Who are the participants? Creating a Quality Sample 47:269: Research Methods I Dr. Leonard March 22, 2010.
Basic Logic of Experimentation The design of an Internally valid experimental procedure requires us to: Form Equivalent Groups Treat Groups Identically.
Making Impact Evaluations Happen World Bank Operational Experience 6 th European Conference on Evaluation of Cohesion Policy 30 November 2009 Warsaw Joost.
TOOLS OF POSITIVE ANALYSIS
Quantitative Research
PAI786: Urban Policy Class 2: Evaluating Social Programs.
Experiments and Observational Studies.  A study at a high school in California compared academic performance of music students with that of non-music.
Research and Evaluation Center Jeffrey A. Butts John Jay College of Criminal Justice City University of New York August 7, 2012 How Researchers Generate.
Randomised controlled trials Peter John. Causation in policy evaluation Outcome Intervention Other agency actions External environment.
Copyright © 2010 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Experiments and Observational Studies. Observational Studies In an observational study, researchers don’t assign choices; they simply observe them. look.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 13 Experiments and Observational Studies.
Measuring Impact: Experiments
Monitoring and Evaluation in MCH Programs and Projects MCH in Developing Countries Feb 10, 2011.
SEDA IMPACT EVALUATION WESTERN CAPE (SOUTH AFRICA) Varsha Harinath (the dti) Francisco Campos (World Bank) Finance and Private Sector Development IE Workshop.
Cultivating Demand Within USAID for Impact Evaluations of Democracy and Governance Assistance Mark Billera USAID Office of Democracy and Governance Perspectives.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Welfare Reform and Lone Parents Employment in the UK Paul Gregg and Susan Harkness.
1 The Need for Control: Learning what ESF achieves Robert Walker.
URBDP 591 I Lecture 3: Research Process Objectives What are the major steps in the research process? What is an operational definition of variables? What.
Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.
Beyond surveys: the research frontier moves to the use of administrative data to evaluate R&D grants Oliver Herrmann Ministry of Business, Innovation.
Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department.
CHAPTER 12 Descriptive, Program Evaluation, and Advanced Methods.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
0 Emerging Findings from the Employment Retention and Advancement (ERA) Evaluation Gayle Hamilton, MDRC Workforce Innovations 2005 Conference.
Developing Evidence on “What Works” in Moving TANF Recipients to Work through Job Search Assistance Karin Martinson, Abt Associates February,
Applying impact evaluation tools A hypothetical fertilizer project.
Impact Evaluation “Randomized Evaluations” Jim Berry Asst. Professor of Economics Cornell University.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Measuring Impact 1 Non-experimental methods 2 Experiments
CHAPTER 2 Research Methods in Industrial/Organizational Psychology
Monitoring and Evaluation in MCH Programs and Projects MCH in Developing Countries Feb 24, 2009.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Steps in Implementing an Impact Evaluation Nandini Krishnan.
Chapter 13 Repeated-Measures and Two-Factor Analysis of Variance
Getting Inside the “Black Box” – Capitalizing on Natural and Random Variation to Learn from the HPOG Impact Study Presenters: Alan Werner, co-Principal.
Evaluation Research Dr. Guerette. Introduction Evaluation Research – Evaluation Research – The purpose is to evaluate the impact of policies The purpose.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Social Experimentation & Randomized Evaluations Hélène Giacobino Director J-PAL Europe DG EMPLOI, Brussells,Nov 2011 World Bank Bratislawa December 2011.
EVALUATION RESEARCH To know if Social programs, training programs, medical treatments, or other interventions work, we have to evaluate the outcomes systematically.
What is Impact Evaluation … and How Do We Use It? Deon Filmer Development Research Group, The World Bank Evidence-Based Decision-Making in Education Workshop.
Research Design Quantitative. Quantitative Research Design Quantitative Research is the cornerstone of evidence-based practice It provides the knowledge.
Randomisation Bias and Post-Randomisation Selection Bias in RCTs: Barbara Sianesi Institute for Fiscal Studies September 14, 2006 The role of non-experimental.
Monitoring and evaluation Objectives of the Session  To Define Monitoring, impact assessment and Evaluation. (commonly know as M&E)  To know why Monitoring.
Copyright © 2015 Inter-American Development Bank. This work is licensed under a Creative Commons IGO 3.0 Attribution-Non Commercial-No Derivatives (CC-IGO.
[Presentation location] [Presentation date] (Confirm ABT logo) Building Bridges and Bonds (B3): An introduction.
Looking for statistical twins
Food and Agriculture Organization of the United Nations
Right-sized Evaluation
CHAPTER 2 Research Methods in Industrial/Organizational Psychology
Chapter 13- Experiments and Observational Studies
Development Impact Evaluation in Finance and Private Sector
Lesson Using Studies Wisely.
Class 2: Evaluating Social Programs
Class 2: Evaluating Social Programs
Research Design Quantitative.
Positive analysis in public finance
Estimating net impacts of the European Social Fund in England
Title Team Members.
Presentation transcript:

Designing a Random Assignment Social Experiment In the U.K.; The Employment Retention and Advancement Demonstration (ERA)

My Role in ERA A member of the 6-person team in the British Cabinet Office that: –Designed the ERA program –Designed the evaluation of ERA to determine how effective it is A member of the team evaluating ERA –I’m in charge of the cost-benefit analysis

My Talk Today How it was designed What the ERA program is The ERA evaluation and how experimental methods are being used to evaluated it Some early results Some lessons from designing ERA (if time permits)

Employment Retention and Advancement Demonstration (ERA) Design Work in Run as pilot program in in 6 sites Analysis conducted:

Why Was ERA Undertaken? To test a program that tries to keep low-wage workers employed after they find jobs and help them advance To promote the use of random assignment experiments in the U.K. Because it was expected the value of the information obtained about the program would exceed the cost of the obtaining the information (“the rational paradigm) –This requires that the information is actually used

Program Features Continued contact with Jobcentre Plus advisors (ASAs) after obtaining employment Retention bonus of £400 every 17 weeks if works full-time for 13 weeks Training bonus of £8 per hour of training Must contact ASA to receive bonus ASAs advise participants on job advancement

Unique Features of ERA Design of Program and Evaluation done simultaneously Design work done at British Cabinet Office Expectation that decision on national implementation of pilot would be based on evaluation findings Pilot run as large-scale random assignment experiment Focus is on what happens after a job is obtained

Evaluation Components Process or implementation study Impact analysis Cost study Cost-benefit analysis

Project Characteristics Long planning period –Absence of political pressure Developing the program and evaluation designs in tandem –Allowed for a good evaluation design Designing the project in the U.K. Cabinet Office –Permitted project team to focus on the project –Need to transfer project to DWP once designed Random assignment

How Does Random Assignment Work? Estimate the impact of a policy, a change in a program, or a intervention Provide evidence of whether the policy has led to (or caused) the change it was designed to - a causal link! The overall objective - to provide policymakers with evidence of whether their policy works

Measuring Impact Experiments measure the impact of policies/interventions in terms of their impact on outcomes; E.g. Does ERA increase earnings? The outcome measure is the earnings of the program group

Establishing Causality To establish that a policy or intervention has caused change to occur rather than some other factor –E.g. many factors will affect individuals earnings in addition to participation in ERA If we find that earnings have increased among the program group, how do we rule-out the influence of other factors? –E.g. most persons who entered ERA were not employed, but some would inevitably have found jobs without ERA. We do this by estimating what we call the ‘counterfactual’—what would have happened without the program

The Counterfactual Is what would have occurred in the absence of the policy or intervention –E.g. what would have happened to earnings over the same period of time for the same individuals had ERA not existed? This is unobservable or missing information We have to estimate the counterfactual - that is determine what would have happened in the absence of ERA

Estimating the Counterfactual Wide range of ways to do this These vary in complexity, rigor and the degree of control required by those planning the evaluation We will look at only one method: The simple two group random assignment experiment When feasible, random assignment is the best method for estimating a counterfactual

Random Assignment In theory the most rigorous way to assess the impact of a policy or intervention Provides unbiased estimates of intervention impacts How does it work? –You identify individuals or groups who are eligible for a new intervention or policy –You create two groups at random - intervention and control groups (essentially a computer flips a coin) –Intervention group receives the new service or intervention, control group does not.

Simple Randomised Experiment Intervention Eligible population R Intervention group Control group Outcome = O 1 Outcome = O 2 the counterfactual is simply ‘O 2 ’ policy impact is ‘O 1 ’- ‘O 2 ’

Random Assignment In ERA, half of those willing to be randomly assigned were randomly assigned to the program group and half to a control group Those assigned to the control group could not receive the services and financial incentives provided The baseline information collected for ERA at the point of random assignment indicates that the two groups are very similar in terms of all observable measures –randomisation worked

Random Assignment Continued After entering ERA, employment, earnings, and other outcomes of the program group are compared for several years to those of the control group during the same years Randomised evaluations of social programs used frequently in US (over 200 times) It is increasingly used in other countries, especially developing countries; but so-far it has been used far less often than in the U.S.

An Illustration At random assignment about 25% of one of the ERA program groups worked 24 months after random assignment, about 55% of the program group worked Was this increase due to ERA?

Illustration (continued) 24 months after random assignment, about 52% of the control group worked Thus, only about a 3 percentage point increase in employment (55% - 52%) is attributable to ERA

Advantages of Random Assignment As intervention and control groups were created randomly, they are statistically equivalent Equivalent in both what we can observed about them and what we cannot At follow-up, when me measure our outcome variables, the only difference between the two groups is the impact of the intervention In theory provides unambiguous results No need for complex statistics as with other methods - ease of interpretation Baseline information not essential, but helpful

Disadvantages of Random Assignment On its own, only provides a measure of average impact - policy makers may have other questions about the policy Can be expensive and complicated to implement Sometimes impractical to implement Possible lack of generalizability Can create political problems by denying services to controls (but if resources are limited some method must be used to deny some) In many cases, can take time for results to emerge Other evaluation designs may also be subject to these limitations-- depends on what is being evaluated

Conclusion Trade off -- rigour against difficulty in implementation Random assignment can be expensive - policy budget! Experimental methods require lots of data Only answers certain types of questions Still, when it is feasible, random assignment provides the most accurate estimates of impacts

Some Early Findings (* = statistically significant impact) NDDPWTC Program Group Control Group ImpactProgram Group Control Group Impact Ever worked Year 165.3%59.7%5.7*97.6%95.9%1.7* Year 267.5% *95.8%94.6%1.2 Months Worked Full time * * Part time * Earnings (£) Year 13,6122,764849*8,2557, Year 24,7814,108673*8,9628,458503*

Lessons from the ERA Design Work Developing ownership among those fielding and running a program important Randomized experiments are feasible, but circumstances must warrant their use Use of multi-disciplinary teams should be encouraged Designing programs and evaluations in tandem should be done whenever possible