Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Student Growth in Non-Tested Subjects and for At-Risk Students.

Slides:



Advertisements
Similar presentations
Understanding Student Learning Objectives (S.L.O.s)
Advertisements

Rhode Island Model for Educator Evaluation Systems August 2010.
Mississippi Statewide Teacher Appraisal Rubric (M-STAR)
Measuring Teacher Impact on Student Learning PEAC Discussion Document| August 20, 2010.
Science Subject Leader Training
1 When DAP Meets GAP Promoting Peaceful Coexistence between Developmentally Appropriate Practice & the Need to Address the Achievement Gap International.
RIDE – Office of Special Populations
1 DPAS II Process and Procedures for Teachers Developed by: Delaware Department of Education.
August 8, 2013 Texas Education Agency | Office of Assessment and Accountability Division of Performance Reporting Shannon Housson, Director Overview of.
Response to Intervention (RtI) in Primary Grades
APS Teacher Evaluation
Teacher Keys Effectiveness System
Discuss the charge of the Michigan Council for Educator Effectiveness (MCEE) Summarize the MCEE Interim Report Provide an Overview of the Pilot.
The SCPS Professional Growth System
Trustee Gough Ward 3 Forum TDSB/Provincial Report Card
Pennsylvania Value-Added Assessment System (PVAAS) High Growth, High Achieving Schools: Is It Possible? Fall, 2011 PVAAS Webinar.
Field Testing Testing the Test March PARCC Consortium 2 Governed by the education chiefs in the states.
Evaluation Orientation Meeting Teacher Evaluation System
Wethersfield Teacher Evaluation and Support Plan
North Carolina Educator Evaluation System. Future-Ready Students For the 21st Century The guiding mission of the North Carolina State Board of Education.
Introduction to Creating a Balanced Assessment System Presented by: Illinois State Board of Education.
The Rubric Reality Cobb Keys Classroom Teacher Evaluation System.
 Teacher Evaluation and Effectiveness laws are now in place  Legislature has passed a law that student performance can now be a part of teacher evaluation.
RTI Implementer Webinar Series: Establishing a Screening Process
The Design and Implementation of Educator Evaluation Systems, Variability of Systems and the Role of a Theory of Action Rhode Island Lisa Foehr Rhode Island.
Deana Holinka, MA, CRC, Administrative Coordinator,
Student Growth Measures in Teacher Evaluation Module 3: Using Data to Inform Growth Targets and Submitting Your SLO 1.
Summative Assessment Kansas State Department of Education ASSESSMENT LITERACY PROJECT1.
LOUISIANA DEPARTMENT OF EDUCATION
1 Phase III: Planning Action Developing Improvement Plans.
By the end of this session we will have an understanding of the following:  A model for teacher evaluation based on current research  The FEAPs as a.
SLG Goals, Summative Evaluations, and Assessment Guidance Training LCSD#7 10/10/14.
Southern Regional Education Board WELCOME Strategy Work Session For What Should the Tech Center of the Future Look Like? Nancy Headrick, Director State.
Compass: Module 3 Student Growth.
Annual UMES Summer Institute “Making the Adjustment” Student Learning Objectives :
Virginia Teacher Performance Evaluation System 0 August 2012.
Introduction to Teacher Evaluation August 20, 2014 Elizabeth M. Osga, Ph.D.
Teacher Evaluation New Teacher Orientation August 15, 2013.
Support Professionals Evaluation Model Webinar Spring 2013.
Teacher Evaluation System LSKD Site Administrator Training August 6, 2014.
Educator Evaluation: A Protocol for Developing S.M.A.R.T. Goal Statements.
1 Literacy Leadership Teams December 2004 Common High-Quality Differentiated Instruction for Achievement for All within The Cleveland Literacy System Module.
Professional Learning
Connecting the Process to: -Current Practice -CEP -CIITS/EDS 1.
Student Learning Objectives Overview. Defining SLOS A vital component of the Teacher Keys and Leader Keys Effectiveness System is Student Growth and Academic.
Data, Now What? Skills for Analyzing and Interpreting Data
Student Growth Developing Quality Growth Goals II
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Measuring Teacher Effectiveness in Untested Subjects and Grades.
Teacher Evaluation Systems: Opportunities and Challenges An Overview of State Trends Laura Goe, Ph.D. Research Scientist, ETS Sr. Research and Technical.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Measuring Teacher Effectiveness in Untested Subjects and Grades.
2012 Secondary Curriculum Teacher In-Service
Student Learning Objectives (SLOs) Measuring Teacher Effectiveness Through the Use of Student Data Overview of the SLO Process April 7,
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Models of Comprehensive Teacher Evaluation Systems Laura Goe,
Student Learning Objectives: Approval Criteria and Data Tracking September 9, 2013 This presentation contains copyrighted material used under the educational.
Compass: Module 2 Compass Requirements: Teachers’ Overall Evaluation Rating Student Growth Student Learning Targets (SLTs) Value-added Score (VAM) where.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Multiple Measures of Teacher Effectiveness Laura Goe, Ph.D. Tennessee.
Student Learning Objectives: Approval Criteria and Data Tracking September 17, 2013 This presentation contains copyrighted material used under the educational.
Hastings Public Schools PLC Staff Development Planning & Reporting Guide.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Evaluating Teacher Effectiveness: Some Models to Consider Laura.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Using Student Growth as Part of a Multiple Measures Teacher Evaluation.
1 National Center on Educational Outcomes What’s so Difficult About Including Special Education Teachers and Their Students in Growth Models Used to Evaluate.
Changes in Professional licensure Teacher evaluation system Training at Coastal Carolina University.
Teacher Evaluation Systems 2.0: What Have We Learned? EdWeek Webinar March 14, 2013 Laura Goe, Ph.D. Research Scientist, ETS Sr. Research and Technical.
Student Learning Objectives NYS District-Wide Growth Goal Setting Process December 1, 2011 EVOLVING.
Models for Evaluating Teacher Effectiveness Laura Goe, Ph.D. CCSSO National Summit on Educator Effectiveness April 29, 2011  Washington, DC.
Forum on Evaluating Educator Effectiveness: Critical Considerations for Including Students with Disabilities Lynn Holdheide Vanderbilt University, National.
Supporting the Development of Student Learning Objectives How to Create an SLO.
Phyllis Lynch, PhD Director, Instruction, Assessment and Curriculum
Teacher Evaluation “SLO 101”
Presentation transcript:

Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Student Growth in Non-Tested Subjects and for At-Risk Students Laura Goe, Ph.D. Measuring Educator Effectiveness: A Workshop for the Southwest States Denver, Colorado  October 5, 2010

Federal priorities (August 2010)  From “Race to the Top” and reiterated in the August 5, 2010 Federal Register (Vol. 75, No. 150) “Secretary’s Priorities for Discretionary Grant Programs” Teachers should be evaluated using state standardized tests where possible For non-tested subjects, other measures (including pre- and post-tests) can be used but must be “rigorous and comparable across classrooms” and must be “between two points in time” Multiple measures should be used, such as multiple classroom evaluations 2

Austin Independent School District Student Learning Objectives:  Teachers determine two SLOs for the semester/year  One SLO must address all students, other may be targeted  Use broad array of assessments  Assess student needs more directly  Align classroom, campus, and district expectations  Aligned to state standards/campus improvement plans  Based on multiple sources of student data  Assessed with pre and post assessment  Targets of student growth  Peer collaboration 3

4 Rubric for student learning objectives

5 Rubric for student learning objectives (cont’d)

SLO Model Strengths/Weaknesses  Strengths  Teachers take an active role in determining student learning goals  Good professional growth opportunity for teachers  If objectives are of high-quality and teachers plan instruction to meet them, students should benefit  Weaknesses  Heavily dependent on administrator understanding and time commitment to supervision  Not clear how or if “rigor” could be determined  Not “comparable across classrooms” because teachers set the objectives and they will vary widely  Not clear how students’ beginning point is determined 6

“Rhode Island Model” is another example of an SLO Model  Under consideration, not yet implemented Teachers measure student growth by setting student academic goals aligned to standards Principals, during the goal setting process, will confer with teachers to establish each goal’s degree of ambition and select the appropriate assessments for measuring progress against the goals Teacher evaluation will be based on students’ progress on the established goals, as determined by an end-of- the-year principal review of the pre-determined assessments and their results 7

The “Rhode Island Model”  The Rhode Island Model (RI Model) 1.Impact on student learning 2.Professional Practice (including content knowledge) 3.Professional Responsibilities  “…each teacher’s Student Learning (SL) rating will be determined by a combination of state-wide standardized tests, district-selected standardized tests, and local school-based measures of student learning whenever possible.” 8

RIDE Model: Impact on Student Learning  Category 1: Student growth on state standardized tests that are developed and/or scored by RIDE  Category 2: Student performance (as measured by growth) on standardized district-wide tests that are developed and/or scored by either the district or by an external party but not by RIDE (e.g., NWEA, AP exams, Stanford-10, ACCESS, etc.)  Category 3: Other, more subjective measures of student performance (growth measures and others, as appropriate) that would likely be developed and/or scored at the district- or school- level (e.g., student performance on school- or teacher-selected assessments, administrator review of student work, attainment of student learning goals that are developed and approved by both teacher and evaluator, etc.) 9

Rhode Island DOE Model: Framework for Applying Multiple Measures of Student Learning Category 1: Student growth on state standardized tests (e.g., NECAP, PARCC) Student learning rating Professional practice rating Professional responsibilities rating + + Final evaluation rating Category 2: Student growth on standardized district-wide tests (e.g., NWEA, AP exams, Stanford- 10, ACCESS, etc.) Category 3: Other local school-, administrator-, or teacher- selected measures of student performance The student learning rating is determined by a combination of different sources of evidence of student learning. These sources fall into three categories: 10

“‘Rhode Island Model”: Student Learning Group Guiding Principles “Not all teachers’ impact on student learning will be measured by the same mix of assessments, and the mix of assessments used for any given teacher group may vary from year to year.” Teacher A (5 th grade) Teacher B (11 th grade English) Teacher C (middle school art) This teacher may use several category 3 assessments Category 1 (growth on NECAP) Category 2 (e.g., growth on NWEA) Category 3 (e.g., principal review of student work over a six month span) Teacher A’s student learning rating + += Category 2 (e.g., AP English exam) Category 3 (e.g., joint review of critical essay portfolio) Teacher B’s student learning rating += 11

“Rhode Island Model” Strengths and Weaknesses  Strengths Includes teachers in evaluation of student learning (outside of standardized tests) Teachers will benefit from having assessment of student learning at the classroom level  Weaknesses Heavily administrator/evaluator driven process Teachers can weigh in on assessments, but do not determine student growth 12

Teacher Advancement Program (TAP) Model  TAP requires that teachers in tested subjects be evaluated with value-added models  All teachers are observed in their classrooms (using a Charlotte Danielson type instrument) at least three times per year by different observers (usually one administrator and two teachers who have been appointed to the role)  Teacher effectiveness (for performance awards) determined by combination of value-added and observations  Teachers in non-tested subjects are given the school- wide average for their value-added component, which is combined with their observation scores 13

TAP strengths/weaknesses  Strengths Value-added becomes everyone’s responsibility, which should encourage efforts from teachers in non-tested subjects to support teachers in tested subjects Multiple yearly observations should be more informative and produce more reliable information about practice Professional development aligned with results is required  Weaknesses Concerns about “fairness” when only a few teachers’ student achievement and progress toward learning goals “counts” Tells you nothing about how teachers in other subjects are performing in terms of student learning growth (grades are not always good indicators) 14

IMPACT sorts teachers into groups that are evaluated differently  Group 1: general ed teachers for whom value-added data can be generated  Group 2: general ed teachers for whom value-added data cannot be generated  Group 3: special education teachers  Group 4: non-itinerant English Language Learner (ELL) teachers and bilingual teachers  Group 5: itinerant ELL teachers  Etc… 15

IMPACT components for Group 1  Individual Value-Added (IVA) = 50% of score  Teaching and Learning Framework (TLF) (measure of instructional expertise) = 40% of score  Commitment to the School Community (CSC) (measure of the extent to which you support your colleagues and your school’s local initiatives) = 5% of score 16

IMPACT components for Group 2  Teaching and Learning Framework (TLF) (measure of instructional expertise) = 80% of score  Non-Value-Added Student Achievement Growth (NVA) = 10%  Commitment to the School Community (CSC) (measure of the extent to which you support your colleagues and your school’s local initiatives) = 5%  School-wide value-added = 5% 17

Explanation for 10% for test scores for Group 2 and others  “As a school system, we recognize that we still have a significant amount of work to do to establish norms for student achievement growth outside of the DC CAS grades and subjects. In recognition of this fact, we have decided to limit this component to just 10% of the overall assessment. As we develop clearer norms, we plan to increase this percentage.” 18

Group 2 assessment rubric  3 “cycles” of data collected & averaged/year  Highest level of rubric: “Teacher has at least 1 high-quality source of evidence (i.e., one that is rigorous and reliable) demonstrating that approximately 90% or more of her/his students are on track to make significant learning growth (i.e., at least a year’s worth) towards mastery of the DCPS content standards over the course of the year.” 19

Non-VAM tests (accepted under Washington, DC’s IMPACT evaluation system)  DC Benchmark Assessment System (DC BAS)  Dynamic Indicators of Basic Early Literacy Skills (DIBELS)  Developmental Reading Assessment (DRA)  Curriculum-based assessments (e.g., Everyday Mathematics)  Unit tests from DCPS-approved textbooks  Off-the-shelf standardized assessments that are aligned to the DCPS Content Standards  Rigorous teacher-created assessments that are aligned to the DCPS Content Standards  Rigorous portfolios of student work that are aligned to the DCPS Content Standards 20

DC IMPACT Strengths & Weaknesses  Strengths Uses multiples measures to assess effectiveness Permits the use of many types of assessment for students in non-tested subjects and grades Includes what is important in the system (in order to encourage specific teacher behaviors)  Weaknesses No multiple measures of student learning growth for teachers in tested subjects and grades Huge differences in how teachers are measured 21

22 Georgia KEYS

23 Georgia KEYS for Non-tested subjects

Georgia KEYS Strengths & Weaknesses  Strengths Rubric for measuring teacher contribution is easy to understand Includes examples of multiple measures of student learning for all teachers, including those in tested grades and subjects  Weaknesses Rubric (including observation and other information) is about 100 pages long Might be a challenge to implement 24

Delaware Model  Standardized test will be used as part of teachers’ scores in some grades/subjects  “Group alike” teachers, meeting with facilitators, determine which assessments, rubrics, processes can be used in their subjects/grades (multiple measures)  Assessments must focus on standards, be given in a “standardized” way, i.e., giving pre-test on same day, for same length of time, with same preparation  Teachers recommend assessments to the state for approval  Teachers/groups of teachers take primary responsibility for determining student growth  State will monitor how assessments are “working” 25

Delaware Model: Strengths & Weaknesses  Strengths Teacher-driven process (assumes teachers are the experts in assessing their students’ learning growth) Great professional growth opportunity as teachers work together across schools to determine assessments, score student work, etc.  Weaknesses Validity issues (how the assessments are given and scored, teacher training to score, etc.) Time must be built in for teachers to work together on scoring (particularly for rubric-based assessments) 26

Challenges for Special Education Teachers and ELL Specialists  Challenges for SWD and ELLs Smaller student numbers create large margins of error How to measure multiple teachers’ contributions Teacher’s contribution to social and behavioral growth may not be factored into results  For teachers of students with disabilities, IEP progress on those factors could be considered  May need an additional rubric-based measure for ELLs Threats to validity of student growth results  Teachers working with students on alternate standards  Little research exploring whether growth rates are comparable  Little research on the use of testing accommodations & their impact related to teacher effects 27

Questions to ask about models  Are they “rigorous and comparable across classrooms”?  Do they show student learning growth “between two points in time”?  Are they based on grade level and subject standards?  Do they allow teachers from all subjects to be evaluated with evidence of student learning growth? 28

29 Questions?

30 Laura Goe, Ph.D.