Modeling “The Cause”: Assessing Implementation Fidelity and Achieved Relative Strength in RCTs David S. Cordray Vanderbilt University IES/NCER Summer Research.

Slides:



Advertisements
Similar presentations
Progress Towards Reading Success: The Reading First Evaluation Prepared by: Amy Kemp, Ph.D. Research Associate and Patricia A. Muller, Ph.D. Associate.
Advertisements

Using Student Data as a Basis for Feedback to Teachers Ronnie Detrich Wing Institute Cal-ABA, 2011.
Top ten non compliance findings from the Office for Exceptional Children from their Special Education Onsite Reviews.
Developing Indicators to Assess Program Outcomes Kathryn E. H. Race Race & Associates, Ltd. Panel Presentation at American Evaluation Association Meeting.
Chapter 2 Flashcards.
EdTPA: Task 1 Support Module Mike Vitale Mark L’Esperance College of Education East Carolina University Introduction edTPA INTERDISCIPLINARY MODULE SERIES.
Response to Intervention
1 Why is the Core important? To set high expectations – for all students – for educators To attend to the learning needs of students To break through the.
Evaluation is a professional and ethical responsibility and is a core part of PHN professional practice Commitment to evaluation helps build the PHN intelligence.
Analyzing Intervention Fidelity and Achieved Relative Strength David S. Cordray Vanderbilt University NCER/IES RCT Training Institute,2010.
1 Reading First Internal Evaluation Leadership Tuesday 2/3/03 Scott K. Baker Barbara Gunn Pacific Institutes for Research University of Oregon Portland,
1. 2 Dimensions of A Healthy System Districts Schools Grades Classrooms Groups.
Title I Needs Assessment and Program Evaluation
Reliability and Validity. Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? How do we judge.
Problem Solving Model Problem Solving Model NC DPI Summer Preparation Preparation & Implementation Implementation North Carolina.
Copyright © 2001 by The Psychological Corporation 1 The Academic Competence Evaluation Scales (ACES) Rating scale technology for identifying students with.
From Learning Goals to Assessment Plans University of Wisconsin Parkside January 20, 2012 Susan Hatfield Winona State University
Assessing Intervention Fidelity in RCTs: Concepts and Methods Panelists: David S. Cordray, PhD Chris Hulleman, PhD Joy Lesnick, PhD Vanderbilt University.
Evaluation Process/Requirements for CAPP Algebra Project.
One Voice – One Plan Office of Education Improvement and Innovation MI-CSI: Do Stage Implement Plan and Monitor Plan.
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
5-Step Process Clarification The 5-Step Process is for a unit, topic, or “chunk” of information. One form should be used for the unit, topic, etc. The.
Achieved Relative Intervention Strength: Models and Methods Chris S. Hulleman David S. Cordray Presentation for the SREE Research Conference Washington,
Specific Learning Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
Progressing Toward a Shared Set of Methods and Standards for Developing and Using Measures of Implementation Fidelity Discussant Comments Prepared by Carol.
Elementary & Middle School 2014 ELA MCAS Evaluation & Strategy.
MI draft of IDEIA 2004 (Nov 2009) WHAT HAS CHANGED? How LD is identified:  Discrepancy model strongly discouraged  Response To Instruction/Intervention.
Targeted Assistance Programs: Requirements and Implementation Spring Title I Statewide Conference May 15, 2014.
What Was Learned from a Second Year of Implementation IES Research Conference Washington, DC June 8, 2009 William Corrin, Senior Research Associate MDRC.
No Child Left Behind Act (NCLBA) Title IID Enhancing Education through Technology Initiative 2 American Recovery and Reinvestment Act (ARRA) Competitive.
INTERNATIONAL SOCIETY FOR TECHNOLOGY IN EDUCATION working together to improve education with technology Using Evidence for Educational Technology Success.
Comp 20 - Training & Instructional Design Unit 6 - Assessment This material was developed by Columbia University, funded by the Department of Health and.
 In Cluster, all teachers will write a clear goal for their IGP (Reflective Journal) that is aligned to the cluster and school goal.
1 PROJECT EVALUATION IT’S ALL ABOUT STUDENTS. 2 In partnership, we help America’s students stay in school and graduate by: Reducing gaps in college access.
TAP TAP Basics (Preparing for Success in a TAP School) [PSTS]
Measuring Complex Achievement
Conceptualizing Intervention Fidelity: Implications for Measurement, Design, and Analysis Implementation Research Methods Meeting September 20-21, 2010.
Committee on the Assessment of K-12 Science Proficiency Board on Testing and Assessment and Board on Science Education National Academy of Sciences.
Conceptualizing Intervention Fidelity: Implications for Measurement, Design, and Analysis Implementation: What to Consider At Different Stages in the Research.
Assessing Intervention Fidelity in RCTs: Models, Methods and Modes of Analysis David S. Cordray & Chris Hulleman Vanderbilt University Presentation for.
{ Principal Leadership Evaluation. The VAL-ED Vision… The construction of valid, reliable, unbiased, accurate, and useful reporting of results Summative.
Professional Learning Communities “The most promising strategy for sustained, substantial school improvement is developing the ability of school personnel.
PBIS Meeting for BCPS Team Leaders and Coaches March 14, 2008 Oregon Ridge.
Laying the Foundation for Scaling Up During Development.
Professional Learning and Development: Best Evidence Synthesis Helen Timperley, Aaron Wilson and Heather Barrar Learning Languages March 2008.
Educable Mental Retardation as a Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
MSRP Year 1 (Preliminary) Impact Research for Better Schools RMC Corporation.
Carol M. Trivette, Ph.D. Orelena Hawks Puckett Institute Asheville and Morganton, North Carolina Presentation prepared for Helping Extend Learning and.
SOCW 671 # 8 Single Subject/System Designs Intro to Sampling.
Tier III Implementation. Define the Problem  In general - Identify initial concern General description of problem Prioritize and select target behavior.
FASA Middle School Principal ’ s Leadership Academy Don Griesheimer Laura Hassler Lang July 22, 2007.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
© Crown copyright 2008 Subject Leaders’ Development Meeting Spring 2009.
Mathematics Performance Tasks Applying a Program Logic Model to a Professional Development Series California Educational Research Association December.
Part 2: Assisting Students Struggling with Reading: Multi-Tier System of Supports H325A
Progressing Toward a Shared Set of Methods and Standards for Developing and Using Measures of Implementation Fidelity Symposium Chair: Chris S. Hulleman,
Care Enough to Count: Measuring Teacher Performance Ronnie Detrich Wing Institute.
Effectiveness of Selected Supplemental Reading Comprehension Interventions: Impacts on a First Cohort of Fifth-Grade Students June 8, 2009 IES Annual Research.
Open Forum: Scaling Up and Sustaining Interventions Moderator: Carol O'Donnell, NCER
The Cause…or the “What” of What Works? David S. Cordray Vanderbilt University IES Research Conference Washington, DC June 16, 2006.
1 OBSERVATION CYCLE: CONNECTING DOMAINS 1, 2, AND 3.
Lesson 3 Measurement and Scaling. Case: “What is performance?” brandesign.co.za.
APR 2014 Report: Data, Analysis and Action Plan for Full Accreditation.
Specific Learning Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
Outcomes By the end of our sessions, participants will have…  an understanding of how VAL-ED is used as a data point in developing professional development.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Teacher Education Improvement Program- Practicum (TEIP-1)
David S. Cordray, PhD Vanderbilt University
H676 Week 5 - Plan for Today Review your project and coding to date
Analyzing Intervention Fidelity and Achieved Relative Strength
Presentation transcript:

Modeling “The Cause”: Assessing Implementation Fidelity and Achieved Relative Strength in RCTs David S. Cordray Vanderbilt University IES/NCER Summer Research Training Institute, 2010

Overview Define implementation fidelity and achieved relative strength A 4-step approach to assessment and analysis of implementation fidelity (IF) and achieved relative strength (ARS): –Model(s)-based –Quality Measures of Core Causal Components –Creating Indices –Integrating implementation assessments with models of effects (next week) Strategy: –Describe each step –Illustrate with an example –Planned Q&A segments

Caveat and Precondition Caveat –The black box (ITT model) is still #1 priority Implementation fidelity and achieve relative strength are supplemental to ITT-based results. Precondition –But, we consider implementation fidelity in RCTs that are conducted on mature (enough) interventions –That is, the intervention is stable enough to describe an underlying— Model/theory of change, and Operational (logic and context) models.

Dimensions Intervention Fidelity Operative definitions: –True Fidelity = Adherence or compliance: Program components are delivered/used/received, as prescribed With a stated criteria for success or full adherence The specification of these criteria is relatively rare –Intervention Exposure: Amount of program content, processes, activities delivered/received by all participants (aka: receipt, responsiveness) This notion is most prevalent –Intervention Differentiation: The unique features of the intervention are distinguishable from other programs, including the control condition A unique application within RCTs

Linking Intervention Fidelity Assessment to Contemporary Models of Causality Rubin’s Causal Model: –True causal effect of X is (Y i Tx – Y i C ) –In RCTs, the difference between outcomes, on average, is the causal effect Fidelity assessment within RCTs also entails examining the difference between causal components in the intervention and control conditions. Differencing causal conditions can be characterized as achieved relative strength of the contrast. –Achieved Relative Strength (ARS) = t Tx – t C –ARS is a default index of fidelity

Achieved Relative Strength =.15 Infidelity “Infidelity” (85)-(70) = 15 t C t tx T Tx TCTC Treatment Strength Expected Relative Strength = ( ) = Outcome

Why is this Important? Statistical conclusion validity Construct Validity: –Which is the cause? (T Tx - T C ) or (t Tx – t C ) Poor implementation: essential elements of the treatment are incompletely implemented. Contamination: The essential elements of the treatment group are found in the control condition (to varying degrees). Pre-existing similarities between T and C on intervention components. External validity – generalization is about (t Tx - t C ) –This difference needs to be known for proper generalization and future specification of the intervention components

So what is the cause? …The achieved relative difference in conditions across components Augmentation of Control Infidelity PD= Professional Development Asmt=Formative Assessment Diff Inst= Differentiated Instruction

Time-out for Questions

In Practice…. Step 1: Identify core components in the intervention group –e.g., via a Model of Change –Establish bench marks (if possible) for T TX and T C Step 2: Measure core components to derive t Tx and t C –e.g., via a “Logic model” based on Model of Change Step 3: Deriving indicators Step 4: Indicators of IF and ARSI Incorporated into the analysis of effects

Focused assessment is needed What are the options? (1) Essential or core components (activities, processes); (2) Necessary, but not unique, activities, processes and structures (supporting the essential components of T); and (3) Ordinary features of the setting (shared with the control group) Focus on 1 and 2.

Step 1: Specifying Intervention Models Simple version of the question: What was intended? Interventions are generally multi-component, sequences of actions Mature-enough interventions are specifiable as: –Conceptual model of change –Intervention-specific model –Context-specific model Start with a specific example  MAP RCT

Example –The Measuring Academic Progress (MAP) RCT The Northwest Evaluation Association (NWEA) developed the Measures of Academic Progress (MAP) program to enhance student achievement Used in school districts, 17,500 schools No evidence of efficacy or effectiveness

Measures of Academic Progress (MAP): Model of Change MAP Intervention: 4 days of training On-demand consultation Formative Testing Student Reports On-line resources Formative Assessment Differentiated Instruction Achievement Delivery – NWEA trainers Receipt – Teachers and School Leaders Enactment -- TeachersOutcomes -- Students Implementation Issues:

Logic Model for MAP Resources Testing System Multiple Assessment Reports NWEA Trainers NWEA Consultants On-line teaching resources Activities 4 training sessions Follow-up Consultation Access resources Outputs Use of Formative Assessment Differentiated Instruction Outcomes State tests MAP tests Impacts Improved Student Achievement Program-specific implementation fidelity assessments: MAP only Comparative implementation assessments: MAP and Non-MAP classes Focus of implementation fidelity and achieved relative strength

Context-Specific Model: MAP AcademicSchedule Fall Semester Spring Semester AugSeptOctNovDecJanFebMarAprMay Major MAP Program Components and Activities PD1 Con PD2 Con PD3 Con PD4 Con Data SysUse Data Diff Instr Change State Testing Implementation Two POINTS: 1. This tells us when assessments should be undertaken; and 2. Provides as basis for determining the length of the intervention study and the ultimate RCT design.

Step 2: Quality Measures of Core Components Measures of resources, activities, outputs Range from simple counts to sophisticated scaling of constructs Generally involves multiple methods Multiple indicators for each major component/activity Reliable scales (3-4 items per sub-scale)

Measuring Program-Specific Components MAP Resources Testing System Multiple Assessment Reports NWEA Trainers NWEA Consultants On-line teaching resources MAP Activities 4 training sessions Follow-up Consultation Access resources Criterion: Present or Absent Source or Method: MAP Records Criterion: Attendance Source or method: MAP records Criterion: Use Source or Method: Web- records MAP Outputs

Measuring Outputs: Both MAP and Non-MAP Conditions MAP Outputs Use of Formative Assessment Data Differentiated Instruction Method: End of year Teacher Survey Methods: End or year teacher survey Observations (3) Teacher Logs (10) Criterion: Achieved Relative Strength Indices: Difference in differentiated instruction (high v. low readiness students) Proportion of observations segments with any differentiated instruction

Fidelity and ARS Assessment Plan for the MAP Program

Step 3: Indexing Fidelity and Achieved Relative Strength True Fidelity – relative to a benchmark; Intervention Exposure – amount of sessions, time, frequency Achieved Relative Strength (ARS) Index Standardized difference in fidelity index across Tx and C Based on Hedges’ g (Hedges, 2007) Corrected for clustering in the classroom

Calculating ARSI When There Are Multiple Components Augmentation of Control Infidelity PD= Professional Development Asmt=Formative Assessment Diff Inst= Differentiated Instruction

Weighted Achieved Relative Strength

Time-out for Questions

Some Program-Specific Results

Achieved Relative Strength: Some Results

Achieved Relative Strength: Teacher Classroom Behavior

Preliminary Conclusions for the MAP Implementation Assessment The developer (NWEA) –Complete implementation of resources, training, and consultation Teachers: Program-specific implementation outcomes –Variable attendance at training and use of training sessions –Moderate use of data, differentiation activities services –Training extended through May, 2009 Teachers: Achieved Relative Strength –No between-group differences in enactment of differentiated instruction

Step 4: Indexing Cause-Effect Linkage Analysis Type 1: –Congruity of Cause-Effect in ITT analyses Effect = Average difference on outcomes  ES Cause = Average difference in causal components  ARS (Achieved Relative Strength) Descriptive reporting of each, separately Analysis Type 2: –Variation in implementation fidelity linked to variation in outcomes –Hierarchy of approaches (ITT  LATE/CACE  Regression  Descriptive) TO BE CONTINUED ……

Questions?