Achieved Relative Intervention Strength: Models and Methods Chris S. Hulleman David S. Cordray Presentation for the SREE Research Conference Washington,

Slides:



Advertisements
Similar presentations
Progress Towards Reading Success: The Reading First Evaluation Prepared by: Amy Kemp, Ph.D. Research Associate and Patricia A. Muller, Ph.D. Associate.
Advertisements

1 Oregon Reading First: Three-Year Report Preliminary Impact Evidence Oregon Reading First Center LLSSC Meeting, November 29, 2006.
LEE JOSEPH CRONBACH  Lee Joseph Cronbach was an American educational psychologist who made significant contributions to psychological testing and measurement.
Experimental Research Designs
Modeling “The Cause”: Assessing Implementation Fidelity and Achieved Relative Strength in RCTs David S. Cordray Vanderbilt University IES/NCER Summer Research.
Analyzing Intervention Fidelity and Achieved Relative Strength David S. Cordray Vanderbilt University NCER/IES RCT Training Institute,2010.
1 Reading First Internal Evaluation Leadership Tuesday 2/3/03 Scott K. Baker Barbara Gunn Pacific Institutes for Research University of Oregon Portland,
1 Oregon K-12 Literacy Framework and K-3 Statewide Outreach.
Using School Climate Surveys to Categorize Schools and Examine Relationships with School Achievement Christine DiStefano, Diane M. Monrad, R.J. May, Patricia.
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
National Center on Response to Intervention Developed by the National Center on Response to Intervention and RMC Research RTI Integrity Framework: A Tool.
Providing Leadership in Reading First Schools: Essential Elements Dr. Joseph K. Torgesen Florida Center for Reading Research Miami Reading First Principals,
The Targeted Reading Intervention: How Early Reading Intervention for Rural Kindergarten and First-Grade Students Affects Teachers’ Ratings of Students’
Copyright © 2001 by The Psychological Corporation 1 The Academic Competence Evaluation Scales (ACES) Rating scale technology for identifying students with.
Reliability and factorial structure of a Portuguese version of the Children’s Hope Scale José Tomás da Silva Maria Paula Paixão Catarina Carvalho dos Santos.
What We Know About Effective Professional Development: Implications for State MSPs Part 2 Iris R. Weiss June 11, 2008.
Aligning Observations and In-service Professional Development Prek-3 rd Jason Downer May 9, 2012.
DEVELOPING ALGEBRA-READY STUDENTS FOR MIDDLE SCHOOL: EXPLORING THE IMPACT OF EARLY ALGEBRA PRINCIPAL INVESTIGATORS:Maria L. Blanton, University of Massachusetts.
Assessing Intervention Fidelity in RCTs: Concepts and Methods Panelists: David S. Cordray, PhD Chris Hulleman, PhD Joy Lesnick, PhD Vanderbilt University.
Striving for Quality Using continuous improvement strategies to increase program quality, implementation fidelity and durability Steve Goodman Director.
NEXT GENERATION BALANCED ASSESSMENT SYSTEMS ALIGNED TO THE CCSS Stanley Rabinowitz, Ph.D. WestEd CORE Summer Design Institute June 19,
Dr. Bonnie J. Faddis & Dr. Margaret Beam RMC Research Fidelity of Implementation and Program Impact.
Leveling the Playing Field With Technology A New Approach to Differentiated Instruction.
KCCT Kentucky’s Commonwealth Accountability Testing System Overview of 2008 Regional KPR.
ERIKA HALL CENTER FOR ASSESSMENT PRESENTATION AT THE 2014 NATIONAL CONFERENCE ON STUDENT ASSESSMENT NEW ORLEANS JUNE 25, 2014 The Role of a Theory of Action.
Conference for Ocean Literacy Larry Snowhite Vice President, Government Relations Houghton Mifflin Company June 7, 2006 Conference.
Evaluating a Literacy Curriculum for Adolescents: Results from Three Sites of the First Year of Striving Readers Eastern Evaluation Research Society Conference.
Striving to Link Teacher and Student Outcomes: Results from an Analysis of Whole-school Interventions Kelly Feighan, Elena Kirtcheva, and Eric Kucharik.
Progressing Toward a Shared Set of Methods and Standards for Developing and Using Measures of Implementation Fidelity Discussant Comments Prepared by Carol.
What Was Learned from a Second Year of Implementation IES Research Conference Washington, DC June 8, 2009 William Corrin, Senior Research Associate MDRC.
Conceptualizing Intervention Fidelity: Implications for Measurement, Design, and Analysis Implementation Research Methods Meeting September 20-21, 2010.
Scientific Validation Of A Set Of Instruments Measuring Fidelity Of Implementation (FOI) Of Reform-based Science And Mathematics Instructional Materials.
Conceptualizing Intervention Fidelity: Implications for Measurement, Design, and Analysis Implementation: What to Consider At Different Stages in the Research.
Assessing Intervention Fidelity in RCTs: Models, Methods and Modes of Analysis David S. Cordray & Chris Hulleman Vanderbilt University Presentation for.
Slides to accompany Weathington, Cunningham & Pittenger (2010), Chapter 3: The Foundations of Research 1.
Laying the Foundation for Scaling Up During Development.
Assessing the Quality of Research
Reliability vs. Validity.  Reliability  the consistency of your measurement, or the degree to which an instrument measures the same way each time it.
Why must we do Data Teams? We know the implementation of DT benefit STUDENTS -helps teachers identify exceeding/proficient students and plan for targeted.
Illustration of a Validity Argument for Two Alternate Assessment Approaches Presentation at the OSEP Project Directors’ Conference Steve Ferrara American.
MSRP Year 1 (Preliminary) Impact Research for Better Schools RMC Corporation.
Issues in Validity and Reliability Conducting Educational Research Chapter 4 Presented by: Vanessa Colón.
Understanding and scaffolding inquiry: A tale of three teachers Lindsay B. Wheeler, Brooke A. Whitworth, & Jennifer L. Maeng Curry School of Education,
Review of Research Methods. Overview of the Research Process I. Develop a research question II. Develop a hypothesis III. Choose a research design IV.
Critical Issues in Formative Assessment NCME conference, April 2013, San Francisco, CA.
Using State Tests to Measure Student Achievement in Large-Scale Randomized Experiments IES Research Conference June 28 th, 2010 Marie-Andrée Somers (Presenter)
The Ohio STEM Learning Network: A Study of Factors Affecting Implementation Spread and Sustainability.
Part 2: Assisting Students Struggling with Reading: Multi-Tier System of Supports H325A
Progressing Toward a Shared Set of Methods and Standards for Developing and Using Measures of Implementation Fidelity Symposium Chair: Chris S. Hulleman,
Aligning Assessments to Monitor Growth in Math Achievement: A Validity Study Jack B. Monpas-Huber, Ph.D. Director of Assessment & Student Information Washington.
CAPS: COACHING TEACHERS Facilitator: Dr. Lynne Paradis BELIZE LITERACY PROGRAM June 2011.
Effectiveness of Selected Supplemental Reading Comprehension Interventions: Impacts on a First Cohort of Fifth-Grade Students June 8, 2009 IES Annual Research.
Open Forum: Scaling Up and Sustaining Interventions Moderator: Carol O'Donnell, NCER
A Multi-Level Framework to Understand Factors Influencing Program Implementation in Schools Celene E. Domitrovich, Ph.D. Penn State Prevention Research.
Evaluation Results MRI’s Evaluation Activities: Surveys Teacher Beliefs and Practices (pre/post) Annual Participant Questionnaire Data Collection.
The Cause…or the “What” of What Works? David S. Cordray Vanderbilt University IES Research Conference Washington, DC June 16, 2006.
AEMP Grade Level Collaborative Module 8 Office of Curriculum, Instruction, and School Support Language Acquisition Branch Academic English Mastery Program.
Randomized Control Trials: What, Why, How, When, and Where Mark L. Davison MESI Conference March 11, 2016 Department of Educational Psychology.
An Integrated Instructional Model for Accelerating Student Achievement in Science and Literacy in Grades 1-2 Nancy Romance, Florida Atlantic University.
Approaches to Linking Process and Outcome Data in a Cross-Site Evaluation Laura Elwyn, Ph.D. Kristin Stainbrook, Ph.D. American Evaluation Association.
New Survey Questionnaire Indicators in PISA and NAEP
Issues in Evaluating Educational Research
David S. Cordray, PhD Vanderbilt University
All readers by 3rd grade Guidance for the Use of Diagnostic
Model Demonstration Projects
Brian Gong Center for Assessment
Welcome to the NJTSS Early Reading Project!
Non-Experimental designs: Correlational & Quasi-experimental designs
Considering Fidelity as an Element Within Scale Up Initiatives Validation of a Multi-Phase Scale Up Design for a Knowledge-Based Intervention in Science.
Analyzing Intervention Fidelity and Achieved Relative Strength
Presentation transcript:

Achieved Relative Intervention Strength: Models and Methods Chris S. Hulleman David S. Cordray Presentation for the SREE Research Conference Washington, DC March 5, 2010

Overview Conceptual Framework –Definitions and Importance –Indexing Fidelity as Achieved Relative Strength (ARS) Three examples –Lab and Field Experiments –Reading First Practical Considerations and Challenges Questions and discussion

Definitions and Implications Fidelity –The extent to which the implemented Tx (t Tx ) was faithful to the intended Tx (T Tx ) –Measure core intervention components Achieved Relative Strength (ARS) –The difference between implemented causal components in the Tx and C –t Tx – t C –ARS is a default index of fidelity Implications –Infidelity reduces construct, external, and statistical conclusion validity

Achieved Relative Strength = 0.15 Infidelity “Infidelity” (85)-(70) = 15 t C t tx T Tx TCTC Treatment Strength Expected Relative Strength = T Tx - T C = ( ) = Outcome

Indexing Fidelity as Achieved Relative Strength Intervention Strength = Treatment – Control Achieved Relative Strength (ARS) Index Standardized difference in fidelity index across Tx and C Based on Hedges’ g (Hedges, 2007) Corrected for clustering in the classroom (ICC’s from.01 to.08) See Hulleman & Cordray (2009)

Indexing Fidelity Average –Mean levels of observed fidelity (t Tx ) Absolute –Compare observed fidelity (t Tx ) to absolute or maximum level of fidelity (T Tx ) Binary –Yes/No treatment receipt based on fidelity scores –Requires selection of cut-off value

Assessing Implementation Fidelity in the Lab and in Classrooms: The Case of a Motivation Intervention Examples 1 and 2

PERCEIVED UTILITY VALUE INTEREST PERFORMANCE MANIPULATED RELEVANCE Model Adapted from: Eccles et al. (1983); Hulleman et al. (2009) The Theory of Change Fidelity Measure: Quality of participant responsiveness (0 to 3 scale)

Achieved Relative Strength Indices Observed Fidelity Lab vs. Class Contrasts LabClassLab - Class AverageTx C g AbsoluteTx C g BinaryTx C0.00 g

Achieved Relative Strength = 1.32 Fidelity Infidelity T Tx TCTC Treatment Strength tCtC t tx Average ARS Index (0.74)-(0.04) = 0.70

Assessing Implementation Fidelity in a Large-Scale Policy Intervention: The Case of Reading First Example 3

In Education, Intervention Models are Multi-faceted (from Gamse et al., 2008) Use of research-based reading programs, instructional materials, and assessment, as articulated in the LEA/school application Teacher professional development in the use of materials and instructional approaches 1)Teacher use of instructional strategies and content based on five essential components of reading instruction 2) Use of assessments to diagnose student needs and measure progress 3) Classroom organization and supplemental services and materials that support five essential components

From Major Components to Indicators… Professional Development Reading Instruction Support for Struggling Readers Assessment Instructional Time Instructional Material Instructional Activities/Strategies Block Actual Time Scheduled block? Reported time Major Components Sub- components Facets Indicators

Reading First Implementation: Specifying Components and Operationalization ComponentsSub-componentsFacetsIndicators (I/F) Reading Instruction Instructional Time 2 2 (1) Instructional Materials 4 12 (3) Instructional Activities /Strategies 8 28 (3.5) Support for Struggling Readers (SR) Intervention Services 3 12 (4) Supports for Struggling Readers 2 16 (8) Supports for ELL/SPED 2 5 (2.5) AssessmentSelection/Interpretation 5 12 (2.4) Types of Assessment 3 9 (3) Use by Teachers 1 7 (7) Professional development Improved Reading Instruction (6.1) (4) Adapted from Moss et al. 2008

Reading First Implementation: Some Results ComponentsSub- components Performance Levels (% of Absolute Standard) Absolute Standard ARSI RFNon-RF Reading Instruction Daily (min.)105 (117%)87 (97%) Daily in 5 components (min.) Daily with High Quality practice Professional Development Hours of PD Five reading dimensions 4.3 (86%)3.7 (74%) Adapted from Gamse et al. (2008) and Moss et al. (2008)

Linking Fidelity to Outcomes

ARS: How Big is Big Enough? Effect Size StudyFidelity ARS Outcome Motivation – Lab Motivation – Field Reading First* *Averaged over 1 st, 2 nd, and 3 rd grades (Gamse et al., 2008).

What Do I Do With Fidelity Indices? Start with: –Scale construction, aggregation over model sub-components and components Use as: –Descriptive analyses –Causal analyses (Intent-to-Treat: ITT) –Explanatory (AKA exploratory) analyses E.g., LATE, Instrumental variables, TOT Except for descriptive analyses, most approaches are relative new and not fully tested

In Practice…. Identify core intervention components –e.g., via a Model of Change Establish bench marks for T TX and T C Measurement –Determine indicators of core components –Derive t Tx and t C –Develop scales –Convert to ARS Incorporate into intervention analyses –Multi-level analyses (Justice, Mashburn, Pence, & Wiggins, 2008)

Some Challenges Intervention models –Often unclear –Scripted vs. Unscripted Measurement –Novel constructs –Multiple levels –Aggregation (within and across levels) Analyses –Weighting of components –Uncertainty about psychometric properties –Functional form not always known

Summary of Key Points Identify and measure core components Fidelity assessment serves two roles: –Average causal difference between conditions –Using fidelity measures to assess the effects of variation in implementation on outcomes Post-experimental (re)specification of the intervention ARS: How much is enough? –Need more data!

Thank You Questions and Discussion