Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.

Slides:



Advertisements
Similar presentations
Brian A. Harris-Kojetin, Ph.D. Statistical and Science Policy
Advertisements

Student Learning Targets (SLT) You Can Do This! Getting Ready for the School Year.
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
Experimental Research Designs
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
Assessing Program Impact Chapter 8. Impact assessments answer… Does a program really work? Does a program produce desired effects over and above what.
MSP Evaluation Rubric and Working Definitions Xiaodong Zhang, PhD, Westat Annual State Coordinators Meeting Washington, DC, June 10-12, 2008.
Mathematics and Science Partnership Grant Title IIB Information Session April 10, 2006.
Questions What is the best way to avoid order effects while doing within subjects design? We talked about people becoming more depressed during a treatment.
Carl D. Perkins Career and Technical Education Improvement Act of 2006.
Chapter 7 Correlational Research Gay, Mills, and Airasian
PISA Partnership to Improve Student Achievement through Real World Learning in Engineering, Science, Mathematics and Technology.
Assessing Financial Education: A Practitioner’s Guide December 2010.
Evaluation of Math-Science Partnership Projects (or how to find out if you’re really getting your money’s worth)
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar May 2014 Note: These slides are intended as guidance.
What We Know About Effective Professional Development: Implications for State MSPs Part 2 Iris R. Weiss June 11, 2008.
Mathematics/Science Partnerships U.S. Department of Education: New Program Grantees.
Overview of MSP Evaluation Rubric Gary Silverstein, Westat MSP Regional Conference San Francisco, February 13-15, 2008.
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
The Evaluation of Mathematics and Science Partnership Program A Quasi Experimental Design Study Abdallah Bendada, MSP Director
Measuring Changes in Teachers’ Mathematics Content Knowledge Dr. Amy Germuth Compass Consulting Group, LLC.
Measuring Changes in Teachers’ Science Content Knowledge Dr. Anne D’Agostino Compass Consulting Group, LLC.
Quasi-Experimental Designs For Evaluating MSP Projects: Processes & Some Results Dr. George N. Bratton Project Evaluator in Arkansas.
Evaluating a Research Report
Accessing and Reporting State Student Achievement Data for GPRA Purposes Amy A. Germuth, Ph.D. Compass Consulting Group, LLC.
Research Indicators for Sustaining and Institutionalizing Change CaMSP Network Meeting April 4 & 5, 2011 Sacramento, CA Mikala L. Rahn, PhD Public Works,
Mathematics and Science Education U.S. Department of Education.
U.S. Department of Education Mathematics and Science Partnerships: FY 2005 Summary.
Assisting GPRA Report for MSP Xiaodong Zhang, Westat MSP Regional Conference Miami, January 7-9, 2008.
Mathematics and Science Partnerships: An Introduction for New State Coordinators February /2013.
Mathematics and Science Partnerships, Title II, Part B, NCLB.
Mathematics and Science Partnerships: Summary of the Performance Period 2008 Annual Reports U.S. Department of Education.
1 Copyright © 2011 by Saunders, an imprint of Elsevier Inc. Chapter 8 Clarifying Quantitative Research Designs.
MSRP Year 1 (Preliminary) Impact Research for Better Schools RMC Corporation.
Classifying Designs of MSP Evaluations Lessons Learned and Recommendations Barbara E. Lovitts June 11, 2008.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
Mathematics and Science Partnerships program U.S. Department of Education Regional Conferences February - March, 2006.
Research and Evaluation Team Lines of Work Andy Porter, Director Building a Partnership – Susan Millar District Case Studies – William Clune Targeted Studies.
Mathematics and Science Partnerships: Summary of the FY2006 Annual Reports U.S. Department of Education.
Precision Gains from Publically Available School Proficiency Measures Compared to Study-Collected Test Scores in Education Cluster-Randomized Trials June.
Evaluation of the Noyce Teacher Scholarship Program 2010 NSF Noyce Conference Abt Associates Inc. July 9, 2010.
The Evaluation of Mathematics and Science Partnerships Program A Quasi Experimental Design Study Abdallah Bendada, Title II Director
Welcome to the San Francisco Mathematics and Science Partnerships Regional Meeting March 21-23, 2011.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Research Design ED 592A Fall Research Concepts 1. Quantitative vs. Qualitative & Mixed Methods 2. Sampling 3. Instrumentation 4. Validity and Reliability.
U.S. Department of Education Mathematics and Science Program State Coordinators’ Meeting.
Mathematics and Science Partnerships: Summary of the Performance Period 2008 Annual Reports U.S. Department of Education.
U.S. Department of Education Mathematics and Science Program State Coordinators Meeting.
Evaluation Designs Adrienne DiTommaso, MPA, CNCS Office of Research and Evaluation.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Mathematics and Science Partnership APR Updates apr.ed-msp.net.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Characteristics of Studies that might Meet the What Works Clearinghouse Standards: Tips on What to Look For 1.
Experimental and Ex Post Facto Designs
1 NO CHILD LEFT BEHIND The reauthorized elementary and secondary education act.
C R E S S T / CU University of Colorado at Boulder National Center for Research on Evaluation, Standards, and Student Testing Measuring Adequate Yearly.
1 DEMONSTRATION PROJECTS TO ENSURE STUDENTS WITH DISABILITIES RECEIVE A QUALITY HIGHER EDUCATION PROGRAM Performance Measurement, Program and Project Evaluation.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar 2015 Update Note: These slides are intended as guidance.
CONNECT WITH CAEP | | Measures of Teacher Impact on P-12 Students Stevie Chepko, Sr. VP for Accreditation.
CaMSP Science Assessment Webinar Public Works, Inc. Sharing Lessons Learned in the Development and Use of Science Assessments for CaMSP Teachers and Students.
June 25, Regional Educational Laboratory - Southwest Review of Evidence on the Effects of Teacher Professional Development on Student Achievement:
EVALUATING A MIDDLE SCHOOL MATH M.ED. PROFESSIONAL DEVELOPMENT PROGRAM. Knowledge, Pedagogy, Practice or Student Achievement:
MSP Summary of First Year Annual Report FY 2004 Projects.
Research and Evaluation
“Excuse Me Sir, Here’s Your Change”
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Evaluation of An Urban Natural Science Initiative
Clinical Studies Continuum
Research Designs, Threats to Validity and the Hierarchy of Evidence and Appraisal of Limitations (HEAL) Grading System.
Quantitative Research
Presentation transcript:

Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions

2 Overview Purpose of MSP Program GPRA Indicators – Teacher Content Knowledge – Student Achievement – Evaluation Design Application of Rubric to Determine Rigor of Evaluation Design – Key Criteria for a Rigorous Design – Common Issues and Potential Solutions

3 Purpose of MSP Program The MSP Program supports partnerships between STEM faculties of institutions of higher education (IHE) and teachers in high-need school districts. These partnerships focus on: 1. Facilitating professional development activities for teachers that focus on improving teacher content knowledge and instruction, 2.Improving classroom instruction, and 3.Improving student achievement. These are linked to indicators that the MSP Program needs to report on annually.

4 GPRA Indicators for MSP Program Under the Government Performance and Results Act (GPRA), all federal agencies are required to develop indicators in order to report to the U.S. Congress on federal program impacts and outcomes. For the MSP Program, the following indicators have been developed that look at the effects of the program on teacher and student outcomes: Teacher Knowledge The percentage of MSP teachers who significantly increase their content knowledge as reflected in project-level pre- and post-assessments. Student Achievement The percentage of students in classrooms of MSP teachers who score at the basic/proficient level or above in State assessments of mathematics or science. Note: The information necessary to report on these indicators is taken directly from the APR.

5 GPRA Indicators for MSP Program (continued) In order to provide information about the impact of the MSP intervention on teacher and student outcomes, a rigorous evaluation design is necessary. The following indicator gets at design issues. Evaluation Design The percentage of MSP projects that use an experimental or quasi- experimental design for their evaluations that are conducted successfully and that yield scientifically valid results.

6 Measuring GRPA Indicators – Evaluation Design Criteria for Evaluating Designs: We apply the Criteria for Classifying Designs of MSP Evaluations (hereafter, referred to as “the rubric”) to projects to determine which projects had rigorous evaluations. The rubric sets the minimum criteria for an MSP evaluation to be considered rigorous. There are seven criteria included. An evaluation has to meet each of the seven criteria in order to meet the GPRA indicator. Based on our previous experience, we have found that one of the most common issues in meeting all of the criteria is missing data. Therefore, throughout this presentation, we will let you know the information we need to apply the rubric. Information Sources: We apply the rubric to final year projects only. We primarily use the information contained in the final evaluation reports, but we compare it to the evaluation data contained in the APRs, and the data do not always agree. It is important to ensure the information contained in all sources is consistent, and that information contained in the final evaluation report is complete.

7 Rubric Criteria 1.Type of design – needs to be experimental or quasi-experimental with comparison group 2.Equivalence of groups at baseline – for quasi-experimental designs, groups should be matched at baseline on variables related to key outcomes 3.Sufficient sample size – to detect a real impact rather than chance findings 4.Quality of measurement instruments – need to be valid and reliable 5.Quality of data collection methods – methods, procedures, and timeframes used to collect the key outcome data need to be comparable for both groups 6.Attrition rates – no more than 70% overall up to 15% differential attrition between groups 7.Relevant statistics reported – treatment and comparison group post- test means, and tests of statistical significance for key outcomes

8 Applying the Rubric – Type of Design 1. Type of Design To determine impact on teacher and student outcomes, need to use an experimental or quasi-experimental design with a comparison group. Common Issues: Many projects used one-group only pre-post studies. These do not account for differences that would have naturally occurred in the absence of the intervention. Potential Solutions: Using a comparison group will help to make a much more rigorous study.

9 Applying the Rubric – Baseline Equivalence 2. Baseline Equivalence of Groups (Quasi- Experimental Only) Demonstration of no significant differences between treatment and comparison at baseline on variables related to the study’s key outcomes. – Pre-test scores should be provided for treatment and comparison groups. – A statistical test of differences should be applied to the treatment and comparison groups.

10 Applying the Rubric – Baseline Equivalence Common Issues: No pre-test information on outcome-related measures. Pre-test results given for the treatment and comparison groups, but no tests of between groups differences. Potential Solutions: Administer pre-test to both groups and test for differences between groups. Alternatively, provide means, standard deviations, and sample sizes of pretest scores for both groups, so differences can be tested. If there were differences at baseline, control for the differences between groups in statistical analyses.

11 Applying the Rubric – Sample Size 3. Sample Size Sample size is adequate – Based on a power analysis with recommended: significance level = 0.05 power = 0.8 minimum detectable effect informed by the literature or otherwise justified

12 Applying the Rubric – Sample Size Common Issues: Power analyses rarely conducted. Different sample sizes given throughout the APR and Evaluation Report. Sample sizes and subgroup sizes not reported for all teacher and student outcomes or are reported inconsistently. Potential Solutions: Conduct power analyses. Provide sample sizes for all groups and subgroups.

13 Applying the Rubric – Measurement Instruments 4. Quality of the Measurement Instruments The study used existing data collection instruments that had already been deemed valid and reliable to measure key outcomes; or Data collection instruments developed specifically for the study were sufficiently pre-tested with subjects who were comparable to the study sample, and instruments were found to be valid and reliable.

14 Applying the Rubric – Measurement Instruments Common Issues: Locally developed instruments not tested for validity or reliability. Instrument identified as “not tested for validity or reliability” in APR, but instruments were pre-existing instruments that had already been tested for validity and reliability. Use many instruments, but do not report validity or reliability for all of them. Assessments aligned with the intervention (this provides an unfair advantage to treatment participants). Potential Solutions: Report on validity and reliability on all instruments. If the instrument was designed for the study, conduct a validity and reliability study. If using pre-existing instrument, cite the validity and reliability of instrument. If using part of existing instruments, consider using full subscales rather than selecting a limited number of items. Do not use instruments which may provide an “unfair advantage” to a particular group.

15 Applying the Rubric – Data Collection Methods 5. Quality of the Data Collection Methods The methods, procedures, and timeframes used to collect the key outcome data from treatment and comparison groups were comparable. Common Issues: Little to no information is provided in general about data collection or only provided for treatment group. Timing of the tests were not comparable for treatment and comparison groups. Potential Solutions: It is important to provide the names and timing of all assessments given to both groups.

16 Applying the Rubric – Attrition 6. Attrition Need to retain at least 70% of original sample; AND Show that if there is differential attrition of more than 15% between groups, it is accounted for in the statistical model. Common Issues: Attrition information is typically not reported, or is reported for treatment groups only. Sample and subsample sizes are not reported for all groups or are reported inconsistently. Potential Solutions: Provide initial and final sample sizes for all groups and subgroups.

17 Applying the Rubric – Statistics Reported 7. Relevant Statistics Reported Include treatment and comparison group post-test means and tests of significance for key outcomes; OR Provides sufficient information for calculation of statistical significance (e.g., mean, sample size, standard deviation/standard error).

18 Applying the Rubric – Statistics Reported Common Issues: Projects report that the results were significant or non-significant but do not provide supporting data. Projects provide p-values but do not provide means or standard deviations. Projects report gain scores for the treatment and comparison groups but do not provide between-group tests of significance. Potential Solutions: Provide full data (means, sample sizes, and standard deviations/errors) for treatment and comparison groups on all key outcomes. Provide complete information about statistical tests that were performed for both groups.

19 Projects with Rigorous Designs Projects that meet all of the rubric criteria will be able to make a more accurate determination of impact of their program on teacher and student outcomes.

20