1 Maximizing Predictive Accuracy of District Benchmarks Illuminate Education, Inc. User’s Conference Aliso Viejo, California June 4&5, 2012.

Slides:



Advertisements
Similar presentations
Comparing State Reading and Math Performance Standards Using NAEP Don McLaughlin Victor Bandeira de Mello National Conference on Large-Scale Assessment.
Advertisements

Rules and Legislation Regarding A-F Report Cards June 2013 Jennifer Stegman, Program Manager CTB.
Jack Buckley Commissioner National Center for Education Statistics May 10, 2012.
Bloomington Public Schools K-12 Pathways Based in part on NWEA (2012) study of RIT scales and ACT College Readiness Benchmarks, TIES equating with MCA.
Current legislation requires the phase-out of high school TAKS and replaces it with 12 EOC assessments in  English I, English II, English III  Algebra.
State of Texas Assessment of Academic Readiness--EOC Tim Walker Nathaniel Session.
1May 6, 2010 Florida Department of Education.  Science Updates  FCAT and FCAT 2.0  End-of-Course Science Assessments (EOC)  Interpretive Products.
Assessment Dashboard March 2013 Page Unified School District Curriculum AssessmentInstruction Data.
Advanced Topics in Standard Setting. Methodology Implementation Validity of standard setting.
Combining Test Data MANA 4328 Dr. Jeanne Michalski
Cutoff Points Patrick Traynor, Ph.D., Director of Assessment and Evaluation Colton Joint Unified School District
1 Effective Use of Benchmark Test and Item Statistics and Considerations When Setting Performance Levels California Educational Research Association Anaheim,
1 Module 2 Using DIBELS Next Data: Identifying and Validating Need for Support.
DETERMINING ELIGIBILITY IN MA EARLY INTERVENTION A General Overview to Scoring.
An Update from the Iowa Testing Programs Mississippi Bend AEA October 7, 2010 David Henkhaus.
Evaluating Pretest to Posttest Score Differences in CAP Science and Social Studies Assessments: How Much Growth is Enough? February 2014 Dale Whittington,
STAR 2010 September 10, Agenda New in 2010 Interpreting reports Comparing results Appendixes A-G 2.
STAR Basics.
Understanding Quick Scores & This Year’s ChangeUnderstanding Quick Scores & This Year’s Change Dr. Nakia TownsDr. Nakia Towns Assistant Commissioner for.
Assessment Dashboard January 2013 Page Unified School District Curriculum AssessmentInstruction Data.
Introduction to GREAT for ELs Office of Student Assessment Wisconsin Department of Public Instruction (608)
12 Ways MAP Data Can Be Used in a School. 12 Ways To Use MAP Data Monitor Academic Growth Using National Norms Identify Individual Reading Pathway using.
Valentine Elementary School San Marino Unified School District Standardized Testing and Reporting (STAR) Spring 2009 California Standards Test.
Analyzing Access For ELL Scores Tracy M. Klingbiel Nash Rocky Mount School District October 11, 2010.
Interpreting DIBELS reports LaVerne Snowden Terri Metcalf
Update on the State Testing Program November 14, 2011.
Validation of the Assessment and Comparability to the PISA Framework Hao Ren and Joanna Tomkowicz McGraw-Hill Education CTB.
Interpreting Assessment Results using Benchmarks Program Information & Improvement Service Mohawk Regional Information Center Madison-Oneida BOCES.
Department of Research and Evaluation Santa Ana Unified School District 2011 CST API and AYP Elementary Presentation Version: Elementary.
FCAT 2.0 and End-of-Course Assessments 1 Kris Ellington Deputy Commissioner Division of Accountability, Research and Measurement 850/
1 Paul Tuss, Ph.D., Program Manager Sacramento Co. Office of Education August 17, 2009 California’s Integrated Accountability System.
Setting Performance Standards for the Hawaii State Alternate Assessments: Reading, Mathematics, and Science Presentation for the Hawaii State Board of.
Reclassification of English Learner Students in California
Review and Validation of ISAT Performance Levels for 2006 and Beyond MetriTech, Inc. Champaign, IL MetriTech, Inc. Champaign, IL.
Points in Distributions n Up to now describing distributions n Comparing scores from different distributions l Need to make equivalent comparisons l z.
Unpublished Work © 2005 by Educational Testing Service Growth Options for California County and District Evaluators’ Meetings May 10 and 19, 2005.
Assessment Training Nebo School District. Assessment Literacy.
Introduction to GREAT for ELs Office of Student Assessment Wisconsin Department of Public Instruction (608)
Spring 2012 Testing Results. GRANT API HISTORY
Method Participants. Approximately 64 second (43.8%) and third (56.3%) grade students. 51.6% female and 48.6% male Measures Oral Reading Fluency (ORF)
35th Annual National Conference on Large-Scale Assessment June 18, 2005 How to compare NAEP and State Assessment Results NAEP State Analysis Project Don.
Combining Test Data MANA 4328 Dr. Jeanne Michalski
Jackson County School District 2012 Preliminary Test Results Mississippi Curriculum Test, 2 nd Edition (Grades 3 – 8, Language Arts and Mathematics) Subject.
Santa Ana Unified School District 2011 CST Enter School Name Version: Intermediate.
Scaling and Equating Joe Willhoft Assistant Superintendent of Assessment and Student Information Yoonsun Lee Director of Assessment and Psychometrics Office.
Future Ready Schools National Assessment of Educational Progress (NAEP) in North Carolina Wednesday, February 13, 2008 Auditorium III 8:30 – 9:30 a.m.
Assessing Learners with Special Needs: An Applied Approach, 6e © 2009 Pearson Education, Inc. All rights reserved. Chapter 5: Introduction to Norm- Referenced.
OPERATION PROFICIENT Spring 2013 USC HYBRID HIGH SCHOOL.
Welcome to MMS MAP DATA INFO NIGHT 2015.
Creating Your School Improvement Plan in ASSIST. Click on the “Goals & Plans” Tab.
Department of Research and Evaluation Santa Ana Unified School District 2011 CST High School.
Chapter 3 Percentiles. Standard Scores A standard score is a score derived from raw data and has a known basis for comparison. A standard score is a score.
The Normal Distribution and Norm-Referenced Testing Norm-referenced tests compare students with their age or grade peers. Scores on these tests are compared.
A Review of the MAP/K-PREP Linking Study and College Readiness (ACT) Benchmarks Nate Jensen, Ph.D. Senior Research Scientist Northwest Evaluation Association.
LISA A. KELLER UNIVERSITY OF MASSACHUSETTS AMHERST Statistical Issues in Growth Modeling.
ESSENTIAL SKILLS ACHIEVEMENT STANDARD ON SMARTER BALANCED Cristen McLean Policy Analyst Derek Brown Director of Assessment.
Student Achievement Data Mount Olive Township Public Schools Winter 2016 RESULTS.
TAKS Release Plan  In 2007 SB 1031 changed the release of tests to every three years  In 2009 HB 3 changed the release of tests to exclude retests 2.
Gilchrist County School District Teacher-Developed Assessments.
STAR Reading. Purpose Periodic progress monitoring assessment Quick and accurate estimates of reading comprehension Assessment of reading relative to.
Standardized Test Reporting
Department of Research and Evaluation
Cedar Falls Board of Education October 2017
How can we use this data to help your child succeed?
NWEA Measures of Academic Progress (MAP)
California Educational Research Association
MANA 4328 Dr. Jeanne Michalski
EVAAS Overview.
Data Tables Packet #19.
12.4 – Measures of Position In some cases, the analysis of certain individual items in the data set is of more interest rather than the entire set. It.
Presentation transcript:

1 Maximizing Predictive Accuracy of District Benchmarks Illuminate Education, Inc. User’s Conference Aliso Viejo, California June 4&5, 2012

2 Maximizing Predictive Accuracy of District Benchmarks Objective 1. Better understand how performance level setting is key to predictive validity. 2. Better understand how to create performance level bands based on equipercentile equating

3 Comparing District Benchmarks to CST Results Common Methods for Setting Cutoffs on District Benchmarks:  Use default settings on assessment platform (e.g. 20%, 40%, 60%, 80%)  Ask curriculum experts for their opinion of where cutoffs should be set  Determine percent correct corresponding to performance levels on CSTs and apply to benchmarks

4 Comparing District Benchmarks to CST Results There is a better way!

5 Comparing District Benchmarks to CST Results “Two scores, one on form X and the other on form Y, may be considered equivalent if their corresponding percentile ranks in any given group are equal.” (Educational Measurement-Second Edition, p. 563)

6 Comparing District Benchmarks to CST Results  Equipercentile Method of Equating at the Performance Level Cut-points  Establishes cutoffs for benchmarks at equivalent local percentile ranks as cutoffs for CSTs  By applying same local percentile cutoffs to each trimester benchmark, comparisons across trimesters within a grade level are more defensible

7 Equipercentile Equating Method Step 1-Identify CST SS Cut-points

8 Equipercentile Equating Method Step 2 - Establish Local Percentiles at CST Performance Level Cutoffs (from scaled score frequency distribution)

9 Equipercentile Equating Method Step 3 – Locate Benchmark Raw Scores Corresponding to the CST Cutoff Percentiles (from benchmark raw score frequency distribution)

10 Equipercentile Equating Method Step 4 – Validate Classification Accuracy – Old Cutoffs

11 Equipercentile Equating Method Step 4 – Validate Classification Accuracy – Old Cutoffs

12 Equipercentile Equating Method Step 4 – Validate Classification Accuracy – Old Cutoffs

13 Equipercentile Equating Method Step 4 – Validate Classification Accuracy – New Cutoffs

14 Equipercentile Equating Method Step 4 – Validate Classification Accuracy – New Cutoffs

15 Equipercentile Equating Method Step 4 – Validate Classification Accuracy – New Cutoffs

16 Example: Classification Accuracy Biology OldNew 2 nd Semester Proficient or Advanced42%77% Each Level38%55% 1 st Semester Proficient or Advanced30%77% Each Level31%50%

17 Example: Classification Accuracy Biology OldNew 1 st Quarter Proficient or Advanced53%71% Each Level41%46%

18 Example: Classification Accuracy Chemistry OldNew 2 nd Semester: Prof. & Adv. 63%79% 2 nd Semester: Each Level 47%52% 1 st Semester: Prof. & Adv. 74% 1 st Semester: Each Level 49%50% 1 st Quarter: Prof. & Adv. 83%76% 1 st Quarter: Each Level 48%47%

19 Example: Classification Accuracy Earth Science OldNew 2 nd Semester: Prof. & Adv. 48%68% 2 nd Semester: Each Level 43%52% 1 st Semester: Prof. & Adv. 33%66% 1 st Semester: Each Level 38%47% 1 st Quarter: Prof. & Adv. 42%56% 1 st Quarter: Each Level 34%41%

20 Example: Classification Accuracy Physics OldNew 2 nd Semester: Prof. & Adv. 57%87% 2 nd Semester: Each Level 37%57% 1 st Semester: Prof. & Adv. 60%88% 1 st Semester: Each Level 42%50% 1 st Quarter: Prof. & Adv. 65%87% 1 st Quarter: Each Level 47%45%

21 Things to Consider Prior to Establishing the Benchmark Cutoffs  Will there be changes to the benchmarks after CST percentile cutoffs are established?  If NO then raw score benchmark cutoffs can be established by linking CST to same year benchmark administration (i.e. spring 2011 CST matched to benchmark raw scores)  If YES then wait until new benchmark is administered and then establish raw score cutoffs on benchmark  How many cases are available for establishing the CST percentiles? (too few cases could lead to unstable percentile distributions)

22 Things to Consider Prior to Establishing the Benchmark Cutoffs (Continued)  How many items comprise the benchmarks to be equated? (as test gets shorter it becomes more difficult to match the percentile cutpoints established on the CST’s)

23 Summary Equipercentile Equating Method  Method generally establishes a closer correspondence between the CST and Benchmarks  Comparisons between benchmark and CST performance can be made more confidently  Comparisons between benchmarks within the school year can be made more confidently

24 Coming Soon from Illuminate Education, Inc.! Reports using the equipercentile methodology are being programmed to: (1) establish benchmark cutoffs for performance bands (2) create validation tables showing improved classification accuracy based on the method

Contact: Tom Barrett, Ph.D. President, Barrett Enterprises, LLC (office) (cell) 25