Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department.

Slides:



Advertisements
Similar presentations
1 Chapter 4 The Designing Research Consumer. 2 High Quality Research: Evaluating Research Design High quality evaluation research uses the scientific.
Advertisements

Cross Sectional Designs
David Fairris Tarek Azzam
Sampling Design Corresponds to Chapter 2 in Triola
Designing Influential Evaluations Session 5 Quality of Evidence Uganda Evaluation Week - Pre-Conference Workshop 19 th and 20 th May 2014.
Experimental Design making causal inferences. Causal and Effect The IV precedes the DV in time The IV precedes the DV in time The IV and DV are correlated.
Reading the Dental Literature
Correlation AND EXPERIMENTAL DESIGN
Missing Data Issues in RCTs: What to Do When Data Are Missing? Analytic and Technical Support for Advancing Education Evaluations REL Directors Meeting.
Childhood Obesity Scenario: Quasi- Experiments and Natural Experiments Versus RCTs Steven Gortmaker, Ph.D. Harvard School of Public Health /Harvard Prevention.
Quasi-Experimental Designs
Chapter 51 Experiments, Good and Bad. Chapter 52 Experimentation u An experiment is the process of subjecting experimental units to treatments and observing.
Questions What is the best way to avoid order effects while doing within subjects design? We talked about people becoming more depressed during a treatment.
Questions I have had some professors who have a preference on APA style, is the library website a good source for APA format? Do you have a particular.
Experiments and Observational Studies.  A study at a high school in California compared academic performance of music students with that of non-music.
Principles of Research Writing & Design Educational Series Fundamentals of Study Design Lauren Duke, MA Program Coordinator Meharry-Vanderbilt Alliance.
Evaluating NSF Programs
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 AIM-CDD Using Randomized Evaluations.
Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Designing a Random Assignment Social Experiment In the U.K.; The Employment Retention and Advancement Demonstration (ERA)
S-005 Intervention research: True experiments and quasi- experiments.
SLWK – 609 Research Methods- online- M. Secret Podcast 5 BASIC CONCEPTS OF RESEARCH DESIGN
Brian Kelly '06 Chapter 13: Experiments. Observational Study n Observational Study: A type of study in which individuals are observed or certain outcomes.
Section 2.4: Simple Comparative Experiments. Experiment – A planned intervention undertaken to observe the effects of one or more explanatory variables,
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 4 Designing Studies 4.2Experiments.
Chapter 4 – Research Methods in Clinical Psych Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Study Session Experimental Design. 1. Which of the following is true regarding the difference between an observational study and and an experiment? a)
Chapter 3.1.  Observational Study: involves passive data collection (observe, record or measure but don’t interfere)  Experiment: ~Involves active data.
Lecture PowerPoint Slides Basic Practice of Statistics 7 th Edition.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
CHAPTER 9: Producing Data: Experiments. Chapter 9 Concepts 2  Observation vs. Experiment  Subjects, Factors, Treatments  How to Experiment Badly 
WWC Standards for Regression Discontinuity Study Designs June 2010 Presentation to the IES Research Conference John Deke ● Jill Constantine.
What is randomization and how does it solve the causality problem? 2.3.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Research Design ED 592A Fall Research Concepts 1. Quantitative vs. Qualitative & Mixed Methods 2. Sampling 3. Instrumentation 4. Validity and Reliability.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Steps in Implementing an Impact Evaluation Nandini Krishnan.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 4 Designing Studies 4.2Experiments.
Evaluation Designs Adrienne DiTommaso, MPA, CNCS Office of Research and Evaluation.
The Disability Employment Initiative (DEI): Impact Evaluation Design October 21, 2015 Sung-Woo Cho, Ph.D.
Elizabeth Spier, PhDJohannes Bos, PhD Principal ResearcherSenior Vice President FAST in Philadelphia SEPTEMBER 2015 Copyright © 2015 American Institutes.
Strategies for Effective Program Evaluations U.S. Department of Education The contents of this presentation were produced by the Coalition for Evidence-Based.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 4 Designing Studies 4.2Experiments.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Characteristics of Studies that might Meet the What Works Clearinghouse Standards: Tips on What to Look For 1.
Experimental and Ex Post Facto Designs
Social Experimentation & Randomized Evaluations Hélène Giacobino Director J-PAL Europe DG EMPLOI, Brussells,Nov 2011 World Bank Bratislawa December 2011.
Producing Data: Experiments BPS - 5th Ed. Chapter 9 1.
CHAPTER 9: Producing Data Experiments ESSENTIAL STATISTICS Second Edition David S. Moore, William I. Notz, and Michael A. Fligner Lecture Presentation.
CHAPTER 9: Producing Data Experiments ESSENTIAL STATISTICS Second Edition David S. Moore, William I. Notz, and Michael A. Fligner Lecture Presentation.
S-005 Intervention research: True experiments and quasi- experiments.
Experiments Textbook 4.2. Observational Study vs. Experiment Observational Studies observes individuals and measures variables of interest, but does not.
Patricia Gonzalez, OSEP June 14, The purpose of annual performance reporting is to demonstrate that IDEA funds are being used to improve or benefit.
Introduction to General Epidemiology (2) By: Dr. Khalid El Tohami.
Experimental Design Ragu, Nickola, Marina, & Shannon.
CHAPTER 4 Designing Studies
Quasi Experimental Methods I
Quasi Experimental Methods I
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
CHAPTER 9: Producing Data— Experiments
Impact Evaluation Methods: Difference in difference & Matching
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
Which Evaluation Designs Are Right for Your State?
CHAPTER 4 Designing Studies
Presentation transcript:

Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department of Labor Washington, DC

Abt Associates | pg 2 Objectives  Present fundamental concepts of quasi- experimental designs (QEDs), particularly matched comparison group designs  Discuss issues with clustered designs  Discuss issues with using a comparison group from an earlier time period  Discuss examples from the field

Abt Associates | pg 3 How Do You Measure Impact?  Often, the most important questions you want to answer are related to whether or not something is helping to improve student outcomes  Answering these questions requires more than just following a group of students or participants  An impact analysis is designed to help answer questions on the effectiveness of interventions

Abt Associates | pg 4 Randomized Controlled Trial (RCT)  Experimental design in which individuals are randomly assigned to the intervention (treatment group), while the other group is not (control group)  Often considered the “gold standard” of impact evaluations  Once you collect outcome data on both the treatment and control groups, you can measure the difference in mean outcomes to determine the effect of the intervention

Abt Associates | pg 5 If Randomization is Not Possible  Often, randomization is not feasible, for a variety of reasons –Difficulty in giving the intervention to just one group because the institution wants all eligible individuals to receive it –Costs/time associated with administration of randomization and follow-up  If running an RCT is not an option, you may be able to use a quasi-experimental design (QED) to determine the impact of an intervention

Abt Associates | pg 6 Quasi-Experimental Design (QED)  The basic idea is that you are matching a treatment group (students, for example) to a comparison group (of similar students) –Match students by using their characteristics (not their outcomes): gender, age, ethnicity, other demographic or academic characteristics  In the end, you have a treatment group and comparison group of students that look similar to one another on key characteristics – except that only the treatment group received the intervention

Abt Associates | pg 7 QED using a Matching Strategy 10 Treatment Students 15 Comparison Students

Abt Associates | pg 8 Match Based on Characteristics 10 Treatment Students 15 Comparison Students

Abt Associates | pg 9 5 Comparison Students are Not Matched (in red) 10 Treatment Students 15 Comparison Students

Abt Associates | pg 10 And They are Left Out of the Sample 10 Treatment Students 10 Comparison Students

Abt Associates | pg 11 Baseline Equivalence  Once you have a treatment group and comparison group that look similar to one another, measure their baseline (that is, pre- intervention) characteristics –Ex: Test scores or wages prior to the start of intervention  Demonstrate that the treatment and comparison groups are very similar at baseline  Helps convince audience that your outcomes are different due to the intervention’s impact on the treatment group

Abt Associates | pg 12 Clustered Designs  Sometimes, whether a person is in a treatment or comparison group depends on whether they attend a certain community college, or live in a certain district, county, etc.  In these situations, for example, the community college is a cluster – the treatment and comparison condition depends on which college you attend –As opposed to a situation where students can be in a treatment or comparison condition within a community college

Abt Associates | pg 13 Clusters and Power  A clustered design may make it easier to distinguish which people are in the treatment condition versus the comparison condition –Ex: A student in community college X received the treatment, but you know that a student in community college Y did not  However, a clustered design diminishes power – that is, your ability to detect an impact –Less variation between clusters than there are between individual students –A design where the treatment and comparison conditions are at the student level will have greater power

Abt Associates | pg 14 Clusters and “N of 1 Confounds”  In certain situations, a treatment or comparison group may consist of only one community college –Ex: One community college in the state runs the program, and you want to use three surrounding community colleges to use as comparison group colleges  In this situation, how would you know that your program is impacting outcomes, or that the characteristics of the community college overall is impacting outcomes?  We often call this an “N of 1 confound,” where we cannot disentangle the impact of the program vs. the characteristics of the cluster –TA guidance: Avoid N of 1 confounds!

Abt Associates | pg 15 Timing of Comparison Groups  One way that some evaluators have created comparison groups is to collect information on previous cohorts (pre-intervention) –Ex: If treatment starts Fall 2015, a comparison group may include a cohort that started in Fall 2013, prior to start of the program  However, what if there was a major change that occurred between the two appearance of the two cohorts? –Ex: A major policy change at the community college in Fall 2014 that had nothing to do with the program, and may have impacted student outcomes

Abt Associates | pg 16 Timing of Comparison Groups  In the previous case, one may argue that there is a time-related bias that may have impacted outcomes for one group and not the other  Having your treatment and comparison groups start at the same time would avoid this type of bias –Baseline test scores or wages would be measured right before the start of the program, for both groups

Abt Associates | pg 17 Concluding Remarks  Try to compare your treatment students’ outcomes against those of similar comparison students using a QED –Match students across the treatment and comparison groups, using the information that you have on students –Clustered QEDs have lower power, compared to non- clustered designs –Previous cohorts can be considered a comparison group, but keep time-related bias in mind

Abt Associates | pg 18 Additional questions? Sung-Woo Cho, Ph.D. Office: