Teacher Value-Added Reports State Board of Education January 14, 2014

Slides:



Advertisements
Similar presentations
Measuring Growth Using the Normal Curve Equivalent
Advertisements

NYC Teacher Data Initiative: An introduction for Teachers ESO Focus on Professional Development December 2008.
NYC Teacher Data Initiative: An introduction for Principals ESO Focus on Professional Development October 2008.
Statistics Part II Math 416. Game Plan Creating Quintile Creating Quintile Decipher Quintile Decipher Quintile Per Centile Creation Per Centile Creation.
Copyright © 2014 by McGraw-Hill Higher Education. All rights reserved.
STATISTICS INTERVAL ESTIMATION Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering National Taiwan University.
Using Growth Models to improve quality of school accountability systems October 22, 2010.
Summary of NCDPI Staff Development 12/4/12
/4/2010 Box and Whisker Plots Objective: Learn how to read and draw box and whisker plots Starter: Order these numbers.
Alaska Accountability Adequate Yearly Progress January 2008, Updated.
Alaska Accountability Adequate Yearly Progress February 2007, Updated.
Lecture Slides Elementary Statistics Tenth Edition
Quality Liaisons November 6th, Responsibilities of the Quality Liaison: Main communication channel between the District and school/department Serve.
Variation, uncertainties and models Marian Scott School of Mathematics and Statistics, University of Glasgow June 2012.
Lecture 7 THE NORMAL AND STANDARD NORMAL DISTRIBUTIONS
School Performance Framework Sponsored by The Colorado Department of Education Summer 2010 Version 1.3.
Kara Hanges, Martha Slamer, Maribeth Pennekamp, Debbie Dye, Rhonda Parker, and Kim Davidson.
The SCPS Professional Growth System
The basics for simulations
Chapter 4: Basic Estimation Techniques
Pennsylvania Value-Added Assessment System (PVAAS) High Growth, High Achieving Schools: Is It Possible? Fall, 2011 PVAAS Webinar.
Data Analysis 53 Given the two histograms below, which of the following statements are true?
No Child Left Behind The reauthorization of the Elementary and Secondary Education Act (ESEA), also known as the “No Child Left Behind Act,” will have.
Copyright © 2010, SAS Institute Inc. All rights reserved. Making Data Work for Kids: EVAAS Teacher Reports October 2012 SAS ® EVAAS ® for K-12.
Before Between After.
2011 FRANKLIN COMMUNITY SURVEY YOUTH RISK BEHAVIOR GRADES 9-12 STUDENTS=332.
Data Driven Decisions Moving from 3D to D 3. Data Driven Decisions Moving from 3D to D 3 Malcolm Thomas Director, Evaluation Services Escambia School.
1  Janet Hensley  Pam Lange  Barb Rowenhorst Meade School District.
PPT Presentation Template: This PPT includes all slides to present a district-level overview of PVAAS. This was used with a district-wide elementary faculty.
12-2 Conditional Probability Obj: To be able to find conditional probabilities and use formulas and tree diagrams.
Basic Statistics Measures of Central Tendency.
Tutorial: Understanding the normal curve. Gauss Next mouse click.
Copyright © 2013 Pearson Education, Inc. All rights reserved Chapter 11 Simple Linear Regression.
SMART GOALS APS TEACHER EVALUATION. AGENDA Purpose Balancing Realism and Rigor Progress Based Goals Three Types of Goals Avoiding Averages Goal.
Introduction to Teacher Evaluation August 20, 2014 Elizabeth M. Osga, Ph.D.
Teacher Practice in  In 2012, the New Jersey Legislature unanimously passed the TEACHNJ Act, which mandates implementation of a new teacher.
Title One Program Evaluation Report to the CCSD Board of Education June 17, 2013 Bill Poock, Title One Coordinator Leslie Titler, Title One Teacher.
VALUE – ADDED 101 Ken Bernacki and Denise Brewster.
Upper Darby School District Growth Data
Enquiring mines wanna no.... Who is it? Coleman Report “[S]chools bring little influence to bear upon a child’s achievement that is independent of.
12.3 – Measures of Dispersion
Classroom Assessment A Practical Guide for Educators by Craig A
OCTEO INTRODUCTION TO VALUE-ADDED ANALYSIS October 25, 2012.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 12 Describing Data.
Descriptive Statistics Used to describe the basic features of the data in any quantitative study. Both graphical displays and descriptive summary statistics.
DLT September 28, State Indicators and Rating for OFCS (have) Key Factors and Points to Keep in Mind (have) This power point presentation (will.
Jim Lloyd_2007 Educational Value Added Assessment System (EVAAS) Olmsted Falls City Schools Initial Presentation of 4 th Grade Students.
Copyright ©2006. Battelle for Kids. Understanding & Using Value-Added Analysis.
PPT Presentation Template: This PPT includes all slides to present a district or building level overview of PVAAS. This was used with a district- wide.
© 2006 McGraw-Hill Higher Education. All rights reserved. Numbers Numbers mean different things in different situations. Consider three answers that appear.
Jim Lloyd_2007 Educational Value Added Assessment System (EVAAS) Phase 2 Presentation.
© 2006 McGraw-Hill Higher Education. All rights reserved. Numbers Numbers mean different things in different situations. Consider three answers that appear.
The Power of Two: Achievement and Progress. The Achievement Lens Provides a measure of what students know and are able to do relative to the Ohio standards,
Copyright © 2010, SAS Institute Inc. All rights reserved. EVAAS Concepts: NCEs and Standard Errors Sandy Horn January 2013 SAS ® EVAAS ® for K-12.
MATRIX OF ACHIEVEMENT AND PROGRESS (MAAP) A New Interactive Data Tool for Ohio Districts.
Jim Lloyd_2007 Educational Value Added Assessment System (EVAAS) Olmsted Falls City Schools Initial Presentation of 4 th Grade Students.
Gifted Presentation Mike Nicholson, Senior Director of Research and Evaluation.
Copyright © 2010, SAS Institute Inc. All rights reserved. How Do They Do That? EVAAS and the New Tests October 2013 SAS ® EVAAS ® for K-12.
DVAS Training Find out how Battelle for Kids can help Presentation Outcomes Learn rationale for value-added progress measures Receive conceptual.
Using EVAAS to Improve Student Performance Donna Albaugh Rachel McBroom Heather Stewart Region 4 PD Leads NCDPI.
Copyright © 2010, SAS Institute Inc. All rights reserved. How Do They Do That? EVAAS and the New Tests October 2013 SAS ® EVAAS ® for K-12.
1 Getting Up to Speed on Value-Added - An Accountability Perspective Presentation by the Ohio Department of Education.
EVAAS Proactive and Teacher Reports: Assessing Students’ Academic Needs and Using Teacher Reports to Improve Student Progress Becky Pearson and Joyce Gardner.
CTE Directors April 11, 2013 Understanding EVAAS: Teacher Effectiveness Reporting.
Chapter 6: Descriptive Statistics. Learning Objectives Describe statistical measures used in descriptive statistics Compute measures of central tendency.
Copyright © 2014 American Institutes for Research and Cleveland Metropolitan School District. All rights reserved. March 2014 Interpreting Vendor Assessment.
Making Data Work for Kids: EVAAS Teacher Reports October 2012 SAS ® EVAAS ® for K-12.
What is Value Added?.
EVAAS EVALUATION EFFECTIVENESS …AND WHAT DOES IT SAY???
EVAAS Overview.
Presentation transcript:

Teacher Value-Added Reports State Board of Education January 14, 2014 Connect for Success Conference 2013 Teacher Value-Added Reports State Board of Education January 14, 2014 Jamie Meade Managing Director, Strategic Measures Battelle for Kids jmeade@battelleforkids.org pnyquist@battelleforkids.org

Session Objectives Provide value-added information relevant to the practitioner’s perspective. What should teachers and administrators know and understand about value-added reporting? How may educators use value-added information to improve professional practice and impact student academic achievement and progress?

Value-Added in Ohio For over 10 years, Battelle for Kids has provided support for professional learning and fostered collegial dialogue for understanding and using value-added measures. Battelle for Kids continues to advocate for the use of value-added measures, in combination with other educational measures, to improve practice and accelerate student academic progress.

Ohio’s Value-Added History 2002: Battelle for Kids’ SOAR Collaborative School and District Value-Added Reports 2006: Battelle for Kids’ TCAP Project Teacher Value-Added Reports 2007: ODE Value-Added on Local Report Card 2009: Battelle for Kids’ Ohio VA High Schools 2011: RttT: 30% Teachers, VA Reports 2012: RttT: 60% Teachers, VA Reports 2013: 4-8 Reading & Math Teachers, VA Reports 80 RttT Districts, K-3 R & M Teachers, VA Reports

Value-Added in Ohio Educational Value Added Assessment System (EVAAS) SAS Analytics Customers in 135 countries More than 65,000 business, government and university sites SAS customers or their affiliates represent 90 of the top 100 companies on the 2012 FORTUNE Global 500® list.

Value-Added Information in Practice: Building Awareness Connect for Success Conference 2013 Value-Added Information in Practice: Building Awareness Understanding the Difference Achievement Measures Progress / Growth Measures jmeade@battelleforkids.org pnyquist@battelleforkids.org

Achievement & Growth: Understanding the Difference Connect for Success Conference 2013 Achievement & Growth: Understanding the Difference Growth Achievement The Achievement Bar (blue bar) represents our expectations for what we expect students to know, understand, and be able to do at a specific point in time. Achievement is measured at a single point in time, usually with a single assessment. Perhaps, referenced as “passing” a test. As students enter our classrooms each year, we note that students enter the room with various levels of prior knowledge, skills, and understandings. (Here, click through the animation for the blue dots to represent students at their varying entry points). Then, discuss how the school’s or the teacher’s effect/impact on learning through a single achievement assessment cannot be determined by an achievement measure at a single point in time since students enter at various levels. Measures of student growth take into consideration where a student begins, and his/her historical assessment data (as available). Thus, multiple data points are used. Growth measures a student from point A to point B. Important Note: Value Added and other growth measures are NOT measuring whether or not a student passes a test, this is a common misunderstanding about Value Added and growth measures, in general. Remember, it’s not about “passing” a test. jmeade@battelleforkids.org pnyquist@battelleforkids.org

Measuring Growth is Important for ALL Students 3rd 4th 5th 6th 7th 8th Jacob Proficient Adam Value-added provides a picture of student growth regardless of students’ achievement levels. Value-added can help us understand whether high-achieving students are making enough progress to sustain or even improve their achievement levels. Value-added can help us understand whether low-achieving students are making enough progress to close the proficiency gap. Talk about this slide from a perspective of a school with Jacob-like students or Adam-like students since value-added is a group effect versus individual student results. In this slide, Student A is currently above the proficiency bar, but is losing ground relative to proficiency. In this slide, Student B is not yet proficient, but is closing the gap on the proficiency bar. Grade

ES/MS Data Team Session 1 2010-2011 “Why can’t we simply compare OAA scaled scores from one year to the next to measure growth?” OAA Math Scaled Score Ranges May 2013 Level 3rd 4th 5th 6th 7th 8th Advanced 447-521 452-551 439-548 448-575 458-574 459-537 Accelerated 429-446 432-451 424-438 429-447 436-457 432-458 Proficient 400-428 400-431 400-423 400-435 Basic 378-399 377-399 382-399 379-399 Limited 240-377 239-376 245-381 235-377 268-377 288-378 Notice the maximum scaled score and minimum scaled score are different at each grade level. Additionally, the maximum and minimum scaled score on each grade level test will vary from year to year. This means, measuring progress through the scaled score is not possible because the scales across grade level tests are not vertically linked. Although a common practice among Ohio educators (and some parents), a student’s increase in scaled score from one school year to the next does not indicate student progress. Likewise, a decrease in the scaled score does not indicate a lack of progress. For example, a score of 596 in grade 6 and then a score of 578 in grade 7 does not mean the student did not progress. Simply looking at scaled scores to detect student growth is quite deceiving and simply inaccurate. at scaled scores to detect student growth is quite deceiving and simply inaccurate. There are ___ points up for grabs. What is the maximum PI score…. gogden@battelleforkids.org; pnyquist@battelleforkids.org

Questions?

Value-Added Information in Practice: SAS EVAAS MRM Model Connect for Success Conference 2013 Value-Added Information in Practice: SAS EVAAS MRM Model Mutlivariate Response Model (MRM) Grades 4–8 Reading and Math VA Reports Uses Ohio Achievement Assessments Reminder: This session is not about OTES/eTPES; this session is about exploring the “anatomy” of the Teacher Value Added Report so that you will be prepared to interpret the Report. jmeade@battelleforkids.org pnyquist@battelleforkids.org

SAS EVAAS Value-Added MRM Model* Connect for Success Conference 2013 SAS EVAAS Value-Added MRM Model* Uses Grades 3–8 Reading and Math OAAs. Compares the average growth of students in the most recent year to the average growth of students in 2010 (state’s baseline year)* Growth expectation is defined as maintaining placement in the distribution of NCE scores from one year to the next* *conceptual definition jmeade@battelleforkids.org pnyquist@battelleforkids.org

Raw Score Scaled Score NCE Connect for Success Conference 2013 Raw Score Scaled Score NCE Sample Raw Score Range Scaled Score Normal Curve Equivalent NCE 52 551 247 99 1 Here, animations can be used to keep the participants focused on the message as you explain the sequential process for deriving NCEs from a student’s raw score. First, a raw score represents the points received from questions answered correctly. (MC = 1 pt. SA = 2 pts. ER = 4 pts.) The raw score is then converted to a 3-digit scaled score. The scaled scores on this sample, range from the minimum scaled score of 247 to the maximum SS of 551. These scaled scores are then converted to an NCE. On Value Added reports, NCE are illustrated on a range from 1-99; it may be important to note that NCEs can range higher than 99, and is often the case. For simplicity, the EVAAS reports will convey a maximum NCE at 99. NCE = Normal Curve Equivalent; let’s discuss that more on the next slide. jmeade@battelleforkids.org pnyquist@battelleforkids.org

Scaled Scores Converted to NCEs in State’s 2010 Baseline Year DCS Value Added 11-12 Scaled Scores Converted to NCEs in State’s 2010 Baseline Year Conversion Values Are Fixed / Frozen Example 2010 Scaled Scores Rank Ordered 551 247 Normal Curve Equivalent (NCE) 99 1 Data Team Session 2

Value-Added Terminology DCS Value Added 11-12 Value-Added Terminology Normal Curve Equivalent (NCE) The NCE is similar to a percentile rank in that scores are derived from scaled scores and ranked based upon performance. A significant difference between percentile rank and NCE is that an NCE scale is an equal interval scale. Data Team Session 2

Normal Curve Equivalent (NCE) Distribution of Scores Percentile Equivalents 1 5 10 20 30 40 50 60 70 80 90 95 99 A normal curve equivalent (NCE), indicates a student's rank compared to other students on the same test (similar to Percentile): NCEs run from 1 to 99 with 50 at the center of the base year distribution. BUT: Normal curve equivalents convert scaled scores to an equal-interval scale Since NCEs are represented on an equal-interval scale, scores can be averaged and compared longitudinally The NCE scale enables longitudinal data connections and the definition of a growth standard that does not change from year to year NCEs represent where a student’s score would place that student relative to student performance in the state’s base year 2010 for OH Normal Curve Equivalents 1 10 20 30 40 50 60 70 80 90 99

Value-Added Terminology Connect for Success Conference 2013 Value-Added Terminology Baseline Score Group of students’ prior year mean NCE Example: Spring 2012 OAA mean NCE Observed Score Group of students’ new/most recent mean NCE Example: Spring 2013 OAA mean NCE Before we use the terms on the next slide, a brief explanation of these terms may be helpful. jmeade@battelleforkids.org pnyquist@battelleforkids.org

SAS EVAAS MRM Model Basic, Conceptual Example Connect for Success Conference 2013 SAS EVAAS MRM Model Basic, Conceptual Example Baseline OAA 2012 99th NCE Observed OAA 2013 99th NCE Expected Growth Maintain Placement in Distribution of Scores For example, a student at the 20th NCE must “at least” stay at the 20th NCE Student 3 92 NCE Student 3 89 NCE (-3) Student 2 67 NCE (+12) Student 2 55 NCE Student 1 20 NCE Student 1 20 NCE (0) jmeade@battelleforkids.org pnyquist@battelleforkids.org

Basic, Conceptual Example Scaled scores are converted to NCEs Grade 6 Baseline Grade 7 Observed 394 = 46 Student 1 430 = 59 402 = 50 Student 2 417 = 54 384 = 42 Student 3 400 = 49 394 = 46 Student 4 390 = 44 410 = 52 Student 5 425 = 57 This is an example for teaching purposes only. EVAAS calculations are more statistically sophisticated to ensure that all students are included in the analysis and that confidence intervals reflect the entire history of student testing. The EVAAS methodology also allows future data to refine past data estimates for more accuracy. NCEs are Normal Curve Equivalents. The NCE scale enables longitudinal data connections and the definition of a growth standard that does not change from year-to-year. This model links student data from one year to the next. Mean Baseline = 47.2 Mean Observed = 52.6 Growth = Mean Observed – Mean Baseline Growth = 52.6 – 47.2 = 5.4 (Mean NCE Gain) A basic measure of the growth for this group is 5.4 NCEs

Questions?

Levels of Value-Added Effects Connect for Success Conference 2013 Levels of Value-Added Effects Students are making substantially more progress than the state growth standard. Students are making more progress than the state growth standard. Students are making about the same amount of progress as the state growth standard. Students are making less progress than the state growth standard. Students are making substantially less progress than the state growth standard. jmeade@battelleforkids.org pnyquist@battelleforkids.org

DCS Value Added 11-12 This slide illustrates the basic calculations of the mean gain model. For example, note the estimated district mean NCE for grade 3 in 2010 was 60.0 (Baseline Mean Score). In 2011, the same cohort of students had an estimated mean NCE of 61.2 (Observed Mean Score). The difference between the two (give / take the standard error doubled of .8) is the Value Added Mean Gain in the top half of the report. Data Team Session 2

Connect for Success Conference 2013 Standard Error A measure of the uncertainty All measures of student learning contain error. In the EVAAS teacher value-added report, the size of the standard error is influenced by N size (size of the student group). Missing scores. jmeade@battelleforkids.org pnyquist@battelleforkids.org

Connect for Success Conference 2013 Year 1 Estimate = 1.5 True value lies somewhere within the range of the standard error 4.0 4 3 2 1 -1 -2 Estimate is Most likely value -1.0 Multi-yr. Ave. Y1 Y2 Y3 jmeade@battelleforkids.org pnyquist@battelleforkids.org

Teacher Value-Added Report Connect for Success Conference 2013 Teacher Value-Added Report Note: Battelle for Kids is utilizing visual representations of copyrighted EVAAS® Web reporting software from SAS in this presentation for instructional purposes. Reminder: This session is not about OTES/eTPES; this session is about exploring the “anatomy” of the Teacher Value Added Report so that you will be prepared to interpret the Report. jmeade@battelleforkids.org pnyquist@battelleforkids.org

Value-Added Terminology ODE Connecting the Dots Conference Value-Added Terminology Growth Index Since the size of the standard error (degree of certainty) will vary across teachers, their estimated gain must be standardized to include both the estimate and the degree of certainty (standard error). Divides a teacher’s estimated gain by the associated standard error. Growth Index appears on the Teacher Report, but is not on the School or District Report. For this reason, this may be a new term for LEAs who have not yet received a Teacher VA report. Take the time to discuss this term before using it in the upcoming slides. jmeade@battelleforkids.org

The Teacher Value-Added Report Garilee Snapshot of the whole report – breakdown to follow.

The Teacher Progress Table

Levels of Teacher Value-Added Effects Connect for Success Conference 2013 Levels of Teacher Value-Added Effects Most Effective Teacher's index: 2 or greater Students are making substantially more progress than the state growth standard. Above Average Teacher's index: equal to or greater than 1, but less than 2 Students are making more progress than the state growth standard. Average Teacher's index: equal to or greater than -1, but less than 1 Students are making about the same amount of progress as the state growth standard. Approaching Average Teacher's index: equal to or greater than -2, but less than -1 Students are making less progress than the state growth standard. Least Effective Teacher's index: less than -2 Students are making substantially less progress than the state growth standard. jmeade@battelleforkids.org pnyquist@battelleforkids.org

Foundational Statistics GR Ohio’s Academic Content Standards Mathematics: Statistical Methods 10 Interpret the relationship between two variables using multiple graphical displays and statistical measures (e.g., box-and-whisker plots and measures of center and spread). 8 Describe how the relative size of a sample compared to the target population affects the validity of predictions. Explain the mean’s sensitivity to extremes… 7 …describe how the inclusion and exclusion of outliers affect those measures. 6 Understand the different information provided by measures of center (mean, mode, median) and measures of spread (range). 1 Describe the likelihood of simple events as possible/impossible and more likely/less likely.

Using Value-Added to Inform Practice Key Considerations: Systemic Programs, Delivery Models, Structures, Services, etc. Professional Learning Curriculum Alignment What I teach Assessment How I measure/monitor learning along the way Instruction How I teach

Combining Measures to Inform Practice Connect for Success Conference 2013 Combining Measures to Inform Practice All measures of student learning contain error. No single measure can capture the complexity of learning and teaching. There’s an important distinction between a flawed measure and a flawed assumption based upon a single measure. jmeade@battelleforkids.org pnyquist@battelleforkids.org

Questions? Thank You! BattelleforKids.org