Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.

Slides:



Advertisements
Similar presentations
Ed-D 420 Inclusion of Exceptional Learners. CAT time Learner-Centered - Learner-centered techniques focus on strategies and approaches to improve learning.
Advertisements

Action Research Not traditional educational research often research tests theory not practical Teacher research in classrooms and/or schools/districts.
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
MSP Evaluation Rubric and Working Definitions Xiaodong Zhang, PhD, Westat Annual State Coordinators Meeting Washington, DC, June 10-12, 2008.
Impact and outcome evaluation involve measuring the effects of an intervention, investigating the direction and degree of change Impact evaluation assesses.
performance INDICATORs performance APPRAISAL RUBRIC
Evaluation of Math-Science Partnership Projects (or how to find out if you’re really getting your money’s worth)
Proposal Writing.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar May 2014 Note: These slides are intended as guidance.
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
Overview of MSP Evaluation Rubric Gary Silverstein, Westat MSP Regional Conference San Francisco, February 13-15, 2008.
Quasi-Experimental Designs For Evaluating MSP Projects: Processes & Some Results Dr. George N. Bratton Project Evaluator in Arkansas.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
Specific Learning Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
Stronge Teacher Effectiveness Performance Evaluation System
Classroom Assessments Checklists, Rating Scales, and Rubrics
INTERNATIONAL SOCIETY FOR TECHNOLOGY IN EDUCATION working together to improve education with technology Using Evidence for Educational Technology Success.
Evaluating a Research Report
Assisting GPRA Report for MSP Xiaodong Zhang, Westat MSP Regional Conference Miami, January 7-9, 2008.
TWS Aids for Student Teachers & Interns Overview of TWS.
Mathematics and Science Partnerships: Summary of the Performance Period 2008 Annual Reports U.S. Department of Education.
Educable Mental Retardation as a Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
Understanding Alaska Measures of Progress Results: Reports 1 ASA Fall Meeting 9/25/2015 Alaska Department of Education & Early Development Margaret MacKinnon,
Classifying Designs of MSP Evaluations Lessons Learned and Recommendations Barbara E. Lovitts June 11, 2008.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
Mathematics and Science Partnerships: Summary of the FY2006 Annual Reports U.S. Department of Education.
Primary Purposes of the Evaluation System
Copyright © Allyn & Bacon 2008 Intelligent Consumer Chapter 14 This multimedia product and its contents are protected under copyright law. The following.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Ch 9 Internal and External Validity. Validity  The quality of the instruments used in the research study  Will the reader believe what they are readying.
Mathematics and Science Partnerships: Summary of the Performance Period 2008 Annual Reports U.S. Department of Education.
Changes in Professional licensure Teacher evaluation system Training at Coastal Carolina University.
Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.
Securing External Federal Funding Janice F. Almasi, Ph.D. Carol Lee Robertson Endowed Professor of Literacy University of Kentucky
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Onsite Quarterly Meeting SIPP PIPs June 13, 2012 Presenter: Christy Hormann, LMSW, CPHQ Project Leader-PIP Team.
1 DEMONSTRATION PROJECTS TO ENSURE STUDENTS WITH DISABILITIES RECEIVE A QUALITY HIGHER EDUCATION PROGRAM Performance Measurement, Program and Project Evaluation.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar 2015 Update Note: These slides are intended as guidance.
Assistant Instructor Nian K. Ghafoor Feb Definition of Proposal Proposal is a plan for master’s thesis or doctoral dissertation which provides the.
CONNECT WITH CAEP | | Measures of Teacher Impact on P-12 Students Stevie Chepko, Sr. VP for Accreditation.
Specific Learning Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
STEPS IN RESEARCH PROCESS 1. Identification of Research Problems This involves Identifying existing problems in an area of study (e.g. Home Economics),
Data Conventions and Analysis: Focus on the CAEP Self-Study
EVALUATING EPP-CREATED ASSESSMENTS
Classroom Assessments Checklists, Rating Scales, and Rubrics
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Elayne Colón and Tom Dana
Presenter: Christi Melendez, RN, CPHQ
Melanie Taylor Horizon Research, Inc.
Understanding Results
Research Designs, Threats to Validity and the Hierarchy of Evidence and Appraisal of Limitations (HEAL) Grading System.
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Classroom Assessments Checklists, Rating Scales, and Rubrics
Critical Reading of Clinical Study Results
Measuring Project Performance: Tips and Tools to Showcase Your Results
IB Environmental Systems and Societies
Overview of the Teacher Work Sample (TWS)
Finding Answers through Data Collection
Chapter Eight: Quantitative Methods
2018 OSEP Project Directors’ Conference
Performance Improvement Projects: PIP Library
Understanding and Using Standardized Tests
Assessing Academic Programs at IPFW
Deconstructing Standard 2a Dr. Julie Reffel Valdosta State University
Developing a Rubric for Assessment
Presenter: Kate Bell, MA PIP Reviewer
CCSSO National Conference on Student Assessment June 21, 2010
Biological Science Applications in Agriculture
Educational Testing Service
Presentation transcript:

Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011

Evaluation is a requirement of all ED MSPs According to the statute: Each MSP project is required to develop an evaluation plan that includes rigorous objectives that measure the impact of activities. However, the type of evaluation design is not specified. The MSP Program only requires projects to report on two aspects of evaluation findings: Teacher gains in content knowledge based on pre- and post-testing; and Proficiency levels on state-level assessments of students of teachers who received professional development. Rigorous evaluations demonstrating that impacts are caused by the program are encouraged, but they may not be appropriate for small projects in their early stages. If your project prepares an evaluation report, this should be attached to the APR.

According to the 2007 Report of the Academic Competitive Council “Successful, large-scale interventions to improve STEM education are unlikely to arise without serious study and trial and error. There is a critical pathway for the development of successful educational interventions and activities, starting generally with small-scale studies to test new ideas and generate hypotheses, leading to increasingly larger and more rigorous studies to test the effect of a given intervention or activity on a variety of students in a variety of settings. Different research methodologies are used along the development pathway, and corresponding evaluation strategies must be used to assess their progress.”

Hierarchy of Study Designs for Evaluating Effectiveness

Overview Criteria for Classifying Designs of MSP Evaluations (“the rubric”) created through the Data Quality Initiative (DQI) Rubric’s key criteria for a rigorous design Common issues with evaluation reports Recommendations for better reporting Preliminary framework for review standards for all evaluation reports

Evaluations Reviewed Using the Rubric All final year evaluations that report using an experimental or quasi-experimental design are considered for review Evaluations need to include a comparison group to ultimately be reviewed with the rubric Within each project, we review evaluations of teacher content knowledge, classroom practices, and student achievement

6 criteria used in rubric Rubric comprises six criterion: Equivalence of groups at baseline Adequate sample size Use of valid & reliable measurement instruments Use of consistent data collection methods Sufficient response and retention rates Reporting of relevant statistics

Criterion 1 – Baseline Equivalence Requirement Study demonstrates no significant differences in key characteristics between treatment and comparison groups at baseline (for the analytic sample) OR Adequate steps were taken to address the lack of baseline equivalence in the statistical analysis Purpose – Helps rule out alternative explanations for differences between groups

Criterion 2 – Sample Size Requirement Sample size is adequate to detect a difference, based on a power analysis using: Significance level = 0.05, Power = 0.8 Minimum detectable effect informed by the literature or otherwise justified Alternatively, meets or exceeds “rule of thumb” sample sizes: School/district-level interventions: 12 schools Teacher-level interventions: 60 teachers (teacher outcomes) or 18 teachers (student outcomes) Purpose – Increases the likelihood of finding an impact

Criterion 3 – Measurement Instruments Requirement – Data collection instruments used were shown to be valid and reliable to measure key outcomes Use existing instruments that have already been deemed valid and reliable Refer to TCK instrument database developed by MSP Knowledge Management and Dissemination Project at http://mspkmd.net/ OR Create new instruments that have either been: Sufficiently tested with subjects comparable to the study sample and found to be valid and reliable, OR Created using scales and items from pre-existing data collection instruments that have been validated and found to be reliable Resulting instrument needs to include at least 10 items, and at least 70 percent of the items are from the validated and reliable instrument(s) Purpose – Ensure that instruments used accurately capture the intended outcomes

Criterion 4 – Data Collection Methods Requirement - Methods, procedures, and timeframes used to collect the key outcome data from treatment and comparison groups are comparable Purpose – Limits possibility that observed differences can be attributed to factors besides the program, such as passage of time and differences in testing conditions

Criterion 5 – Attrition Requirement Need to measure key outcomes for at least 70% of original sample (both treatment and control groups), or evidence that attrition is unrelated to treatment If the attrition rates between groups equal or exceed 15 percentage points, difference should be accounted for in the statistical analysis Purpose – Helps ensure that sample attrition does not bias results as participants/control group members who drop out may systematically differ from those who remain

Criterion 6 – Relevant Statistics Reported Requirement Include treatment and comparison group post-test means and tests of significance for key outcomes OR, Provide sufficient information for calculation of statistical significance (e.g., mean, sample size, standard deviation/standard error) Purpose – Provides context for interpreting results, indicating where observed differences between groups are most likely larger than what chance alone might cause

Common Issues Found in Evaluation Reports Information critical for complete assessment of all criteria is often not reported, inconsistently reported, or only reported for the treatment group Pre & post sample sizes for both groups, means, standard deviations/ errors are frequently missing – these are needed for statistical testing and to calculate attrition rates Varying sample sizes throughout report without explanations for changes Validity and reliability testing not reported for locally – developed instruments or cited for pre-existing instruments Data collection methods are not discussed

Key Recommendation – Report the Details Describe the intervention in detail, including such information as: the format of the training, the background of the trainer(s), the material covered, the frequency of classes, and the numbers of participants. Report pre & post sample sizes for both groups and explain changes in samples sizes; if reporting sub-groups, indicate their sample sizes as well Report key characteristics associated with outcomes at baseline (e.g., pretest scores, teaching experience) Document and describe the data collection procedures Report means, standard deviations/errors, for both groups on key outcomes; if using a regression model, describe it Report results from appropriate significance testing of differences observed between groups (e.g., t-statistics or p-values)

Preliminary Framework for Review Standards for Evaluation Reports Preliminary framework based on generally acceptable standards for high quality research studies. General Framework and Background Data Sources and Methods Presentation Of Findings And Placing Them In Context Report Summary

Mathematics and Science Partnership (MSP) Programs U.S. Department of Education San Francisco Regional Meeting March 21-23, 2011 17