WWC Standards for Regression Discontinuity Study Designs June 2010 Presentation to the IES Research Conference John Deke ● Jill Constantine.

Slides:



Advertisements
Similar presentations
Educational Research: Causal-Comparative Studies
Advertisements

Synthesizing the evidence on the relationship between education, health and social capital Dan Sherman, PhD American Institutes for Research 25 February,
Mywish K. Maredia Michigan State University
How Do We Know if a Charter School is Really Succeeding? – Various Approaches to Investigating School Effectiveness October 2012 Missouri Charter Public.
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
Experimental Research Designs
Regression Discontinuity Design Thanks to Sandi Cleveland and Marc Shure (class of 2011) for some of these slides.
PHSSR IG CyberSeminar Introductory Remarks Bryan Dowd Division of Health Policy and Management School of Public Health University of Minnesota.
Correlation AND EXPERIMENTAL DESIGN
How does the process work? Submissions in 2007 (n=13,043) Perspectives.
Study Design Data. Types of studies Design of study determines whether: –an inference to the population can be made –causality can be inferred random.
Regression Discontinuity (RD) Andrej Tusicisny, methodological reading group 2008.
Nasih Jaber Ali Scientific and disciplined inquiry is an orderly process, involving: problem Recognition and identification of a topic to.
Experimental Design The Gold Standard?.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar May 2014 Note: These slides are intended as guidance.
Overview of MSP Evaluation Rubric Gary Silverstein, Westat MSP Regional Conference San Francisco, February 13-15, 2008.
Research Design for Quantitative Studies
Research methods in clinical psychology: An introduction for students and practitioners Chris Barker, Nancy Pistrang, and Robert Elliott CHAPTER 8 Foundations.
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
Single-Case Research: Standards for Design and Analysis Thomas R. Kratochwill University of Wisconsin-Madison.
EVAL 6970: Cost Analysis for Evaluation Dr. Chris L. S. Coryn Nick Saxton Fall 2014.
Research Policies and Mechanisms: Key Points from the National Mathematics Advisory Panel Joan Ferrini-Mundy Director, Division of Research on Learning.
Systematic Review Module 7: Rating the Quality of Individual Studies Meera Viswanathan, PhD RTI-UNC EPC.
URBDP 591 A Lecture 8: Experimental and Quasi-Experimental Design Objectives Basic Design Elements Experimental Designs Comparing Experimental Design Example.
Assisting GPRA Report for MSP Xiaodong Zhang, Westat MSP Regional Conference Miami, January 7-9, 2008.
What Works Clearinghouse Susan Sanchez Institute of Education Sciences.
Preliminary Results – Not for Citation Strengthening Institutions Program Webinar on Competitive Priority on Evidence April 11, 2012 Note: These slides.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar April 25, 2012 Note: These slides are intended as.
For ABA Importance of Individual Subjects Enables applied behavior analysts to discover and refine effective interventions for socially significant behaviors.
Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department.
1 Copyright © 2011 by Saunders, an imprint of Elsevier Inc. Chapter 8 Clarifying Quantitative Research Designs.
Using Regression Discontinuity Analysis to Measure the Impacts of Reading First Howard S. Bloom
Classifying Designs of MSP Evaluations Lessons Learned and Recommendations Barbara E. Lovitts June 11, 2008.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
Evidence-based Education and the Culture of Special Education Chair: Jack States, Wing Institute Discussant: Teri Palmer, University of Oregon.
November 15, Regional Educational Laboratory - Southwest The Effects of Teacher Professional Development on Student Achievement: Finding from a Systematic.
Evaluating Ongoing Programs: A Chronological Perspective to Include Performance Measurement Summarized from Berk & Rossi’s Thinking About Program Evaluation,
Applying impact evaluation tools A hypothetical fertilizer project.
+ IDENTIFYING AND IMPLEMENTING EDUCATIONAL PRACTICES SUPPORTED BY RIGOROUS EVIDENCE: A USER FRIENDLY GUIDE Presented by Kristi Hunziker University of Utah.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Measuring Impact 1 Non-experimental methods 2 Experiments
Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.
An Expanded Model of Evidence-based Practice in Special Education Randy Keyworth Jack States Ronnie Detrich Wing Institute.
Finding, Evaluating, and Presenting Evidence Sharon E. Lock, PhD, ARNP NUR 603 Spring, 2001.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Randomized Assignment Difference-in-Differences
Design of Clinical Research Studies ASAP Session by: Robert McCarter, ScD Dir. Biostatistics and Informatics, CNMC
Developing an evaluation of professional development Webinar #4: Going Deeper into Analyzing Results 1.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Characteristics of Studies that might Meet the What Works Clearinghouse Standards: Tips on What to Look For 1.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Effectiveness of Selected Supplemental Reading Comprehension Interventions: Impacts on a First Cohort of Fifth-Grade Students June 8, 2009 IES Annual Research.
SOCI 4466 PROGRAM & POLICY EVALUATION LECTURE #8 1. Evaluation projects 2. Take-home final 3. Questions?
Regression Discontinuity Design Case Study : National Evaluation of Early Reading First Peter Z. Schochet Decision Information Resources, Inc.
IES Single-Case Research Institute: Training Visual Analysis Rob Horner University of Oregon
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar 2015 Update Note: These slides are intended as guidance.
Randomized Control Trials: What, Why, How, When, and Where Mark L. Davison MESI Conference March 11, 2016 Department of Educational Psychology.
Research design By Dr.Ali Almesrawi asst. professor Ph.D.
Evidence Review Methodological Framework: The Case of Teen Pregnancy Prevention Research April 2011 Presentation to the Inter-American Development Bank.
Evidence-Based Mental Health PSYC 377. Structure of the Presentation 1. Describe EBP issues 2. Categorize EBP issues 3. Assess the quality of ‘evidence’
Uses of Diagnostic Tests Screen (mammography for breast cancer) Diagnose (electrocardiogram for acute myocardial infarction) Grade (stage of cancer) Monitor.
V: Maryland’s High School Assessments (HSAs) & the Bridge Plan for Academic Validation Overview.
American Evaluation Association Annual Conference Friday October 17, 2014 Denver, CO Session #817.
June 25, Regional Educational Laboratory - Southwest Review of Evidence on the Effects of Teacher Professional Development on Student Achievement:
Gabe Avakian Orona, M.P.H. CAIR 2016 November 17, 2016
Don’t Be Fooled: Assessing the Quality of Evidence
Bryan G. Cook University of Hawaii
S1316 analysis details Garnet Anderson Katie Arnold
Presentation transcript:

WWC Standards for Regression Discontinuity Study Designs June 2010 Presentation to the IES Research Conference John Deke ● Jill Constantine

What is the What Works Clearinghouse (WWC)?  Initiative of the U.S. Department of Education's Institute of Education Sciences  Central and trusted source of scientific evidence for what works in education –Develops and implements standards for reviewing and synthesizing education research –Assesses the rigor of research evidence on the effectiveness of interventions 2

3

Current WWC Study Designs Standards  Randomized Control Trials (RCTs) –Well executed RCTs can obtain the highest level of causal validity and meet evidence standards  Quasi-experimental Designs (QEDs) –Some RCTs and matched comparison designs meet evidence standards with reservations  Regression Discontinuity Designs (RDDs)  Single-Case Study Designs 4

Components of a Research Design Standard  Establishes a level of rigor that informs causality  Operational –Follows a protocol –Replicable and reliable –Transparent  Can be consistently applied to completed studies 5

What is RDD?  Similar to RCT in that treatment and control groups formed by design, not by unobserved self-selection  Different from RCT in that groups are not formed randomly – they are formed purposefully  If groups formed using a cutoff on a continuous "forcing variable," then it is an RDD 6

Process for Developing Standards  Panel convenes  Standards drafted  Study Review Guide drafted  Standards pilot tested  Standards finalized and approved by IES 7

WWC RD Expert Panel  Panelists included Tom Cook, John Deke, Guido Imbens, J.R. Lockwood, Jack Porter, Peter Schochet, and Jeff Smith  The panel met once with follow-up conference calls over the course of several months  WWC staff drafted the RD standards and incorporated feedback from the panelists  The standards document is ready to be used by the WWC, but can be updated in the future 8

Key Considerations  Should the WWC allow RDD studies to join RCTs as studies that “meet standards without reservations”?  How can we distinguish between RDD studies that meet standards with versus without reservations? 9

Overview of RD Standards  Screener to identify studies as being RDD  Three possible designations: 1.Meets standards without reservations 2.Meets standards with reservations 3.Does not meet standards  Four individual standards: 1.Integrity of the forcing (assignment) variable 2.Attrition 3.Continuity of the outcome-forcing variable relationship 4.Functional form and bandwidth  Reporting requirement for standard errors 10

RDD Screener  Treatment assignment is based on a “forcing variable” – units on one side of a cutoff value are in the treatment group, units on the other side are in the comparison group  The forcing variable must be ordinal with at least 4 values above and 4 values below the cutoff value  There must be no factor confounded with the cutoff value of the forcing variable 11

1. Integrity of the Forcing Variable  Primary concern is “manipulation”  Two criteria: institutional integrity and statistical integrity  Both criteria must be met to pass this standard without reservations; one criterion must be met to pass with reservations 12

2. Attrition  Primary concern is nonresponse bias  Standard is almost the same as the WWC RCT standard on attrition  One difference: overall attrition and differential attrition can either be reported at the cutoff value of the forcing variable or for the full treatment and comparison samples 13

3. Continuity of Outcome-Forcing Variable Relationship  Primary concern is phantom impacts from lack of “smoothness” in the relationship between the outcome and the forcing variable  Two criteria: baseline equivalence on key covariates and no evidence of unexplained discontinuities away from the cutoff value  Both criteria must be met to pass this standard without reservations 14

4. Functional Form and Bandwidth  Primary concern is misspecification bias  Five criteria: A.An adjustment must be made for the forcing variable B.A graphical analysis must be included and should be consistent with bandwidth/functional form choice C.Statistical evidence must be provided of an appropriate bandwidth or functional form D.Model should “interact” forcing variable with treatment status or provide evidence that an interaction is not needed E.All of the above must be done for every combination of outcome, forcing variable, and cutoff value  All criteria must be met to pass standards without reservations; to meet with reservations, (A and D) plus (B or C) 15

Example of Graphical Analysis 16

Meeting Evidence Standards  To meet standards without reservations, must meet each individual standard without reservations  To meet standards with reservations, must meet individual standards 1, 2, and 4 with or without reservations  Fails to meet WWC evidence standards if individual standards 1, 2, or 4 are not met 17

Reporting Requirement for Standard Errors  Standard errors must reflect clustering at the unit of assignment  Lee and Card (2008) – clustering issue applies when forcing variable is not truly continuous (“random misspecification error”)  This does not affect whether a study meets standards, but it does affect how its findings can be used 18