Assisting GPRA Report for MSP Xiaodong Zhang, Westat MSP Regional Conference Miami, January 7-9, 2008.

Slides:



Advertisements
Similar presentations
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
Advertisements

Experimental Research Designs
Quasi-Experimental Design
ESTEEMS (ESTablishing Excellence in Education of Mathematics and Science) Project Overview and Evaluation Dr. Deborah H. Cook, Director, NJ SSI MSP Regional.
MSP Evaluation Rubric and Working Definitions Xiaodong Zhang, PhD, Westat Annual State Coordinators Meeting Washington, DC, June 10-12, 2008.
Educational Research by John W. Creswell. Copyright © 2002 by Pearson Education. All rights reserved. Slide 1 Chapter 11 Experimental and Quasi-experimental.
Quantitative Research  Adequate experimental control  Lack of artificiality  Basis for comparison  Adequate information from the data  Uncontaminated.
PISA Partnership to Improve Student Achievement through Real World Learning in Engineering, Science, Mathematics and Technology.
Experimental Research
Nasih Jaber Ali Scientific and disciplined inquiry is an orderly process, involving: problem Recognition and identification of a topic to.
Evaluation of Math-Science Partnership Projects (or how to find out if you’re really getting your money’s worth)
Experimental Design The Gold Standard?.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar May 2014 Note: These slides are intended as guidance.
I want to test a wound treatment or educational program in my clinical setting with patient groups that are convenient or that already exist, How do I.
Overview of MSP Evaluation Rubric Gary Silverstein, Westat MSP Regional Conference San Francisco, February 13-15, 2008.
Research Design for Quantitative Studies
Day 6: Non-Experimental & Experimental Design
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
Quasi-Experimental Designs Manipulated Treatment Variable but Groups Not Equated.
The Evaluation of Mathematics and Science Partnership Program A Quasi Experimental Design Study Abdallah Bendada, MSP Director
Measuring Changes in Teachers’ Mathematics Content Knowledge Dr. Amy Germuth Compass Consulting Group, LLC.
Measuring Changes in Teachers’ Science Content Knowledge Dr. Anne D’Agostino Compass Consulting Group, LLC.
Quasi-Experimental Designs For Evaluating MSP Projects: Processes & Some Results Dr. George N. Bratton Project Evaluator in Arkansas.
Power Point Slides by Ronald J. Shope in collaboration with John W. Creswell Chapter 11 Experimental Designs.
The Genetics Concept Assessment: a new concept inventory for genetics Michelle K. Smith, William B. Wood, and Jennifer K. Knight Science Education Initiative.
Mathematics and Science Education U.S. Department of Education.
Chapter 1: Introduction to Statistics
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
U.S. Department of Education Mathematics and Science Partnerships: FY 2005 Summary.
Chapter Four Experimental & Quasi-experimental Designs.
Mathematics and Science Partnerships: Summary of the Performance Period 2008 Annual Reports U.S. Department of Education.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar April 25, 2012 Note: These slides are intended as.
MSP Annual Performance Report: Online Instrument MSP Regional Conferences November, 2006 – February, 2007.
Classifying Designs of MSP Evaluations Lessons Learned and Recommendations Barbara E. Lovitts June 11, 2008.
Tim Brower Professor & Chair Manufacturing & Mechanical Engr. Oregon Institute of Technology MSP Regional Meeting, San Francisco, February 14 & 15, 2008.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
WWC Standards for Regression Discontinuity Study Designs June 2010 Presentation to the IES Research Conference John Deke ● Jill Constantine.
EDCI 696 Dr. D. Brown Presented by: Kim Bassa. Targeted Topics Analysis of dependent variables and different types of data Selecting the appropriate statistic.
Mathematics and Science Partnerships: Summary of the FY2006 Annual Reports U.S. Department of Education.
Second Language Classroom Research (Nunan, D. 1990) Assoc. Prof. Dr. Sehnaz Sahinkarakas.
The Evaluation of Mathematics and Science Partnerships Program A Quasi Experimental Design Study Abdallah Bendada, Title II Director
Welcome to the San Francisco Mathematics and Science Partnerships Regional Meeting March 21-23, 2011.
Experiments. The essential feature of the strategy of experimental research is that you… Compare two or more situations (e.g., schools) that are as similar.
Chapter 10 Common Experimental Research Designs
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Chapter 10 Finding Relationships Among Variables: Non-Experimental Research.
U.S. Department of Education Mathematics and Science Program State Coordinators’ Meeting.
Mathematics and Science Partnerships: Summary of the Performance Period 2008 Annual Reports U.S. Department of Education.
U.S. Department of Education Mathematics and Science Program State Coordinators Meeting.
Evaluation Designs Adrienne DiTommaso, MPA, CNCS Office of Research and Evaluation.
Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Mathematics and Science Partnership APR Updates apr.ed-msp.net.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Characteristics of Studies that might Meet the What Works Clearinghouse Standards: Tips on What to Look For 1.
Effectiveness of Selected Supplemental Reading Comprehension Interventions: Impacts on a First Cohort of Fifth-Grade Students June 8, 2009 IES Annual Research.
MSP Regional Meeting February 13-15, 2008 Calli Holaway-Johnson, Ph.D. Charles Stegman, Ph.D. National Office for Research on Measurement and Evaluation.
How do you know your product “works”? And what does it mean for a product to “work”?
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar 2015 Update Note: These slides are intended as guidance.
1 Teacher Evaluation Institute July 23, 2013 Roanoke Virginia Department of Education Division of Teacher Education and Licensure.
June 25, Regional Educational Laboratory - Southwest Review of Evidence on the Effects of Teacher Professional Development on Student Achievement:
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
EXPERIMENTAL RESEARCH
Evaluation of An Urban Natural Science Initiative
Research Designs, Threats to Validity and the Hierarchy of Evidence and Appraisal of Limitations (HEAL) Grading System.
Classroom Assessment Validity And Bias in Assessment.
Chapter Eight: Quantitative Methods
Designing an Experiment
Quantitative Research
Reminder for next week CUELT Conference.
Presentation transcript:

Assisting GPRA Report for MSP Xiaodong Zhang, Westat MSP Regional Conference Miami, January 7-9, 2008

GPRA GPRA measures GPRA Measure—the percentage of MSP teachers who significantly increase their content knowledge, as reflected in project-level pre- and post-assessments GPRA Measure—the percentage of MSP projects that use an experimental or quasi- experimental design for their evaluations that are conducted successfully and that yield scientifically valid results

DQI and GPRA Data Quality Initiative: technical assistance to ED grant programs to improve the quality of information on program performance GPRA Measure: MSP Teacher Content Knowledge (MSPTCK) software GPRA Measure: Evaluation rubric

MSPTCK objectives Designed to assist grantees to determine the number of teachers who have made significant gains in content knowledge (APR Q8b-f) Mathematics b) Number of teachers with both pretest and posttest in math content knowledge b) Number of teachers with both pretest and posttest in math content knowledge c) Number of teachers who showed significant gains in math content knowledge c) Number of teachers who showed significant gains in math content knowledgeScience e) Number of teachers with both pretest and posttest in science content knowledge e) Number of teachers with both pretest and posttest in science content knowledge f)Number of teachers who showed significant gains in science content knowledge f)Number of teachers who showed significant gains in science content knowledge

MSPTCK process Two-step process Step 1. Determine if overall posttest scores are significantly higher than pretest (paired-samples t-test) Step 2. Determine the number of teachers in the tested group who made significant gains If more than one test used for different groups, run application separately for each test If more than one test used for different groups, run application separately for each test If more than one test used for the same group, select the most relevant one If more than one test used for the same group, select the most relevant one

MSPTCK tutorial Access the spreadsheet online

MSPTCK tutorial

Enter data here tab Blank rows not allowed between data Blank rows not allowed between data Import data from another Excel file Import data from another Excel file Perform range check Perform range check

MSPTCK tutorial

MSP tutorial View results here tab

MSPTCK tutorial Information tab

Evaluation rubric The criteria on the rubric are the minimum criteria that need to be met for an evaluation to have been successfully conducted and yield valid data An evaluation has to meet each criterion in order to meet the GPRA measure

Rubric: designs Experimental study—the study measures the intervention’s effect by randomly assigning individuals (or other units, such as classrooms or schools) to a group that participated in the intervention, or to a control group that did not, and then compares post-intervention outcomes for the two groups Quasi-experimental study—the study measures the intervention’s effect by comparing post-intervention outcomes for treatment participants with outcomes for a comparison group (that was not exposed to the intervention), chosen through methods other than random assignment. For example: Comparison-group study with equating—a study in which statistical controls and/or matching techniques are used to make the treatment and comparison groups similar in their pre-intervention characteristics Comparison-group study with equating—a study in which statistical controls and/or matching techniques are used to make the treatment and comparison groups similar in their pre-intervention characteristics

Rubric: designs Regression-discontinuity study—a study in which individuals (or other units, such as classrooms or schools) are assigned to treatment or comparison groups on the basis of a “cutoff” score on a pre-intervention non-dichotomous measure Regression-discontinuity study—a study in which individuals (or other units, such as classrooms or schools) are assigned to treatment or comparison groups on the basis of a “cutoff” score on a pre-intervention non-dichotomous measure Other study—The study uses a design other than a randomized controlled trial, comparison-group study with equating, or regression-discontinuity study, including pre-post studies, which measure the intervention’s effect based on the pre-test to post-test differences of a single group, and comparison-group studies without equating, or non-experimental studies that compare outcomes of groups that vary with respect to implementation fidelity or program dosage

Rubric: aspects For Experimental Designs: Sample Size Quality of Measurement Instruments Quality of Data Collection Methods Attrition Rates, Response Rates Relevant Statistics Reported For Quasi-experimental Designs: All of the above, plus Baseline Equivalence of Groups

Rubric: sample size Met the criterion—sample size was adequate (i.e., based on power analysis with recommended significance level=0.05, power=0.8, and a minimum detectable effect informed by the literature or otherwise justified) Did not meet the criterion —the sample size was too small Did not address the criterion

Rubric: quality of measurement instruments Met the criterion—the study used existing data collection instruments that had already been deemed valid and reliable to measure key outcomes, or data collection instruments developed specifically for the study were sufficiently pre-tested with subjects who were comparable to the study sample Did not meet the criterion—the key data collection instruments used in the evaluation lacked evidence of validity and reliability Did not address the criterion

Rubric: quality of data collection methods Met the criterion—the methods, procedures, and timeframes used to collect the key outcome data from treatment and control groups were the same Did not meet the criterion— instruments/assessments were administered differently in manner and/or at different times to treatment and control group participants

Rubric: data reduction rates Met the criterion—(1) the study measured the key outcome variable(s) in the post-tests for at least 70% of the original study sample (treatment and control groups combined), or there is evidence that the high rates of data reduction were unrelated to the intervention, AND (2) the proportion of the original study sample that was retained in follow-up data collection activities (e.g., post-intervention surveys) and/or for whom post-intervention data were provided (e.g., test scores) was similar for both the treatment and control groups (i.e., less or equal to a 15-percent difference), or the proportion of the original study sample that was retained in the follow-up data collection was different for the treatment and control groups, but sufficient steps were taken to address this differential attrition in the statistical analysis

Rubric: data reduction rates Did not meet the criterion—(1) the study failed to measure the key outcome variable(s) in the post-tests for 30% or more of the original study sample (treatment and control groups combined), and there is no evidence that the high rates of data reduction were unrelated to the intervention; OR (2) the proportion of study participants who participated in follow-up data collection activities (e.g., post-intervention surveys) and/or for whom post- intervention data were provided (e.g., test scores) was significantly different for the treatment and control groups (i.e., more than a 15-percent difference) and sufficient steps to address differential attrition were not taken in the statistical analysis Did not address the criterion

Rubric: relevant statistics reported Met the criterion—the final report includes treatment and control group post-test means and tests of statistical significance for key outcomes or provides sufficient information for calculation of statistical significance (e.g., mean, sample size, standard deviation/standard error) Did not meet the criterion—the final report does not include treatment and control group post-test means and/or tests of statistical significance for key outcomes or provide sufficient information for calculation of statistical significance (e.g., mean, sample size, standard deviation/standard error) Did not address the criterion

Rubric: baseline equivalence of groups for QED Met the criterion—there were no significant pre- intervention differences between treatment and comparison group participants on variables related to the study’s key outcomes, or adequate steps were taken to address the lack of baseline equivalence in the statistical analysis Did not meet the criterion—there were statistically significant pre-intervention differences between treatment and comparison group participants on variables related to the study’s key outcomes, and no steps were taken to address lack of baseline equivalence in the statistical analysis Did not address the criterion

Disclaimer The instructional practices and assessments discussed or shown in this presentation are not intended as an endorsement by the U.S. Department of Education.