Evaluating Teacher Effectiveness Laura Goe, Ph.D. Presentation to National Conference of State Legislatures July 14, 2011  Webinar.

Slides:



Advertisements
Similar presentations
Briefing: NYU Education Policy Breakfast on Teacher Quality November 4, 2011 Dennis M. Walcott Chancellor NYC Department of Education.
Advertisements

VALUE – ADDED 101 Ken Bernacki and Denise Brewster.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Evaluating Teacher Effectiveness: Where Do We Go From Here? Laura.
1 Monthly Webinar Statewide Teacher Evaluation Systems January 11, 2011 State Consortium on Educator Effectiveness (SCEE)
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Evaluating Teacher Effectiveness Laura Goe, Ph.D. Teacher Evaluation:
Michigan Council for Educator Effectiveness Toward an Improvement-Focused System of Educator Evaluation Jennifer Hammond OCTE Meeting November 7, 2013.
Implementing Virginia’s Growth Measure: A Practical Perspective Deborah L. Jonas, Ph.D. Executive Director, Research and Strategic Planning Virginia Department.
MEASURING TEACHING PRACTICE Tony Milanowski & Allan Odden SMHC District Reform Network March 2009.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Measuring Teacher Effectiveness in Untested Subjects and Grades.
Evaluating Teacher Effectiveness Laura Goe, Ph.D. Presentation to the Hawaii Department of Education July 20, 2011  Honolulu, HI.
Teacher Evaluation Models: A National Perspective Laura Goe, Ph.D. Research Scientist, ETS Principal Investigator for Research and Dissemination,The National.
How can we measure teachers’ contributions to student learning growth in the “non-tested” subjects and grades? Laura Goe, Ph.D. Research Scientist, ETS,
Using Student Learning Growth as a Measure of Effectiveness Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive.
Deliberate Practice Technical Assistance Day
Teacher Evaluation Systems: Opportunities and Challenges An Overview of State Trends Laura Goe, Ph.D. Research Scientist, ETS Sr. Research and Technical.
Seattle’s Teacher Evaluation Reform in Context. Two questions How does PG&E’s design compare to evaluation systems nationwide? What can SPS learn about.
Silas Deane Middle School Steven J. Cook, Principal Cynthia Fries, Assistant Principal October 22, 2013 Wethersfield Board of Education.
Next Generation Professionals Opportunities, Practice & Outcomes Opportunities, Practice & Outcomes Interim Joint Committee on Education July 12, 2010.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Measuring Teacher Effectiveness in Untested Subjects and Grades.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Measuring Teacher Effectiveness in Untested Subjects and Grades.
Presentation for Teachers and Administrators in the New Canaan Public Schools, New Canaan, CT What is differentiated supervision? Why is it necessary?
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Evaluating Teacher/Leader Effectiveness Laura Goe, Ph.D. Webinar.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Evaluating Teacher Effectiveness: Decisions for Districts Laura.
Student Learning Objectives (SLOs) Measuring Teacher Effectiveness Through the Use of Student Data Overview of the SLO Process April 7,
1 Orientation to Teacher Evaluation /15/2015.
Examining Value-Added Models to Measure Teacher Effectiveness Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive.
GTEP Resource Manual Training 2 The Education Trust Study (1998) Katie Haycock “However important demographic variables may appear in their association.
Measuring Teacher Effectiveness: Challenges and Opportunities
Teacher Evaluation in Rural Schools Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive Center for Teacher.
Compass: Module 2 Compass Requirements: Teachers’ Overall Evaluation Rating Student Growth Student Learning Targets (SLTs) Value-added Score (VAM) where.
Evaluating teacher effectiveness with multiple measures Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Evaluating Teacher/Leader Effectiveness Laura Goe, Ph.D. Webinar.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Multiple Measures of Teacher Effectiveness Laura Goe, Ph.D. Tennessee.
Evaluation Team Progress Collaboration Grant 252.
Models for Evaluating Teacher/Leader Effectiveness Laura Goe, Ph.D. Eastern Regional SIG Conference Washington, DC  April 5, 2010.
Alaska Staff Development Network – Follow-Up Webinar Emerging Trends and issues in Teacher Evaluation: Implications for Alaska April 17, :45 – 5:15.
Teacher Effectiveness Pilot II Presented by PDE. Project Development - Goal  To develop a teacher effectiveness model that will reform the way we evaluate.
Comprehensive Educator Effectiveness: New Guidance and Models Presentation for the Special Education Advisory Committee Virginia Department of Education.
Measuring teacher effectiveness using multiple measures Ellen Sullivan Assistant in Educational Services, Research and Education Services, NYSUT AFT Workshop.
What Are We Measuring? The Use of Formative and Summative Assessments Laura Goe, Ph.D. Research Scientist, ETS Principal Investigator for Research and.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Teacher/Leader Effectiveness Laura Goe, Ph.D. Research Scientist,
1 Evaluating Teacher Effectiveness Presented at the March 11, 2010 Meeting of Professional Standards Commission for Teachers (PSCT) Rachele DiMeglio Michigan.
Sales Training Master PPT00 SIOP Research. Sales Training Master PPT1 SIOP Research.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Evaluating Teacher Effectiveness: Some Models to Consider Laura.
Evaluating Teacher Effectiveness: Selecting Measures Laura Goe, Ph.D. SIG Schools Webinar August 12, 2011.
Models for Evaluating Teacher Effectiveness Laura Goe, Ph.D. California Labor Management Conference May 5, 2011  Los Angeles, CA.
ESEA, TAP, and Charter handouts-- 3 per page with notes and cover of one page.
Changes in Professional licensure Teacher evaluation system Training at Coastal Carolina University.
Teacher Evaluation Systems 2.0: What Have We Learned? EdWeek Webinar March 14, 2013 Laura Goe, Ph.D. Research Scientist, ETS Sr. Research and Technical.
Teacher evaluation: A multiple- measures system Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive Center.
Models for Evaluating Teacher Effectiveness Laura Goe, Ph.D. CCSSO National Summit on Educator Effectiveness April 29, 2011  Washington, DC.
Teacher Quality/Effectiveness: Defining, Developing, and Assessing Policies and Practices Part III: Setting Policies around Teacher Quality/Effectiveness.
Weighting components of teacher evaluation models Laura Goe, Ph.D. Research Scientist, ETS Principal Investigator for Research and Dissemination, The National.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Evaluating Teacher and Principal Effectiveness Laura Goe, Ph.D.
Forum on Evaluating Educator Effectiveness: Critical Considerations for Including Students with Disabilities Lynn Holdheide Vanderbilt University, National.
UPDATE ON EDUCATOR EVALUATIONS IN MICHIGAN Directors and Representatives of Teacher Education Programs April 22, 2016.
1 Update on Teacher Effectiveness July 25, 2011 Dr. Rebecca Garland Chief Academic Officer.
Using Student Growth in Teacher Evaluation and Development Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive.
If you do not have speakers or a headset, dial For technical assistance, Thursday, February 3,
Evaluating Teacher Effectiveness Laura Goe, Ph.D. Presentation to the Hawaii Department of Education July 20, 2011  Honolulu, HI.
Measuring Principals’ Effectiveness:
Growing Leader Practice: Observing Data Discussions
Phyllis Lynch, PhD Director, Instruction, Assessment and Curriculum
Evaluating Teacher Effectiveness: Where Do We Go From Here?
Evaluating Teacher Effectiveness: An Overview Laura Goe, Ph.D.
The School Turnaround Group
State Consortium on Educator Effectiveness (SCEE)
Teacher Evaluation Models: A National Perspective
Teacher Evaluation: The Non-tested Subjects and Grades
California Education Research Association
Presentation transcript:

Evaluating Teacher Effectiveness Laura Goe, Ph.D. Presentation to National Conference of State Legislatures July 14, 2011  Webinar

2 Trends in teacher evaluation Policy is way ahead of the research in teacher evaluation measures and models  Though we don’t yet know which model and combination of measures will identify effective teachers, many states and districts are compelled to move forward at a rapid pace Inclusion of student achievement growth data represents a huge “culture shift” in evaluation  Communication and teacher/administrator participation and buy-in are crucial to ensure change The implementation challenges are enormous  Few models exist for states and districts to adopt or adapt  Many districts have limited capacity to implement comprehensive systems, and states have limited resources to help them

3 Research behind the push for new evaluation measures and systems Value-added research shows that teachers vary greatly in their contributions to student achievement (Rivkin, Hanushek, & Kain, 2005). The Widget Effect report (Weisberg et al., 2009) “…examines our pervasive and longstanding failure to recognize and respond to variations in the effectiveness of our teachers.” (from Executive Summary)

4 Teacher observations: strengths and weaknesses Strengths  Great for teacher formative evaluation (if observation is followed by opportunity to discuss)  Helps evaluator (principals or others) understand teachers’ needs across school or across district Weaknesses  Only as good as the instruments and the observers  Considered “less objective”  Expensive to conduct (personnel time, training, calibrating)  Validity of observation results may vary with who is doing them, depending on how well trained and calibrated they are

5 Research on observations: Danielson Framework Lots of research on Danielson Framework (1996) and whether its scores correlate with student achievement growth  Goe (2007) reviews many studies, most finding weak correlation (likely due to lack of rater training)  Kane et al. (2010) describes research linking observation scores with value-added scores (found relatively small but significant correlations)  Sartain et al. (2010) describes challenges in implementation, differences in researcher/principal ratings - Principals tend to see the evidence the same as researchers, but score teachers higher

6 Research on observations: CLASS Considerable research, mostly conducted by creators of CLASS  Howes et al. (2008): children’s relationship with teachers, not teachers’ qualifications, mattered  Pianta et al. (2007): “Children from nonpoor families and who scored high on achievement at 54 months were most likely to experience classrooms high in positive emotional or instructional climate throughout elementary school. Poor children were highly unlikely (only 10%) to experience classrooms with high instructional climate across multiple grades.”

7 Teacher behaviors & practices that correlate with achievement High ratings on learning environment (classroom observations) (Kane et al., 2010) Positive student/teacher relationships (Howes et al., 2008) Parent engagement efforts by teachers and schools (Redding et al., 2004) Teachers’ participation in intensive professional development with follow-up (Yoon et al., 2007) IN MANY CURRENT TEACHER EVALUATION MODELS, THESE ARE NEVER MEASURED.

8 8 8 Value-added models Many variations on value-added models  TVAAS (Sander’s original model) typically uses 3+ years of prior test scores to predict the next score for a student - Used since the 1990’s for teachers in Tennessee, but not for high-stakes evaluation purposes - Most states and districts that currently use VAMs use the Sanders’ model, also called EVAAS  There are other models that use less student data to make predictions  Considerable variation in “controls” used

9 Promises of value-added… Value-added can provide useful evidence of teacher’s contribution to student growth  “It is not a perfect system of measurement, but it can complement observational measures, parent feedback, and personal reflections on teaching far better than any available alternative.” Glazerman et al. (2010) pg 4  “…value-added estimates should best be used in combination with other indicators…when there are no serious consequences for individual teachers, administrators, or students.” Braun et al. (2010) p 54

10 …and perils of value-added “Any practical application of value-added measures should make use of confidence intervals in order to avoid false precision, and should include multiple years of value-added data in combination with other sources of information to increase reliability and validity.” Glazerman et al. (2010) pg 5 “Type I and II error rates for comparing a teacher’s performance to the average are likely to be about 25 percent with three years of data and 35 percent with one year of data.” Schochet & Chiang (2010) abstract

11 Technical considerations in using value-added for teacher evaluation Koedel & Betts (2009) suggest using multiple years of data for teacher evaluation to mitigate sorting bias; novice teachers cannot be evaluated under this system McCaffrey et al. (2009) “…there are significant gains in the stability [of teachers’ value-added scores] obtained by using two-year average performance measures rather than singe-year estimates”

12 Most popular growth models: Colorado Growth Model Colorado Growth model  Focuses on “growth to proficiency”  Measures students against “academic peers”  Also called criterion ‐ referenced growth ‐ to ‐ standard models The student growth percentile is “descriptive” whereas value-added seeks to determine the contribution of a school or teacher to student achievement (Betebenner 2008)

13 What nearly all state and district models have in common Value-added or Colorado Growth Model will be used for those teachers in tested grades and subjects (4-8 ELA & Math in most states) States want to increase the number of tested subjects and grades so that more teachers can be evaluated with growth models States are generally at a loss when it comes to measuring teachers’ contribution to student growth in non-tested subjects and grades

14 Measuring teachers’ contributions to student learning growth: A summary of current models ModelDescription Student learning objectives Teachers assess students at beginning of year and set objectives then assesses again at end of year; principal or designee works with teacher, determines success Subject & grade alike team models (“Ask a Teacher”) Teachers meet in grade-specific and/or subject-specific teams to consider and agree on appropriate measures that they will all use to determine their individual contributions to student learning growth Pre-and post-tests model Identify or create pre- and post-tests for every grade and subject School-wide value- added Teachers in tested subjects & grades receive their own value-added score; all other teachers get the school- wide average

15 Considerations Consider whether human resources and capacity are sufficient to ensure fidelity of implementation  Poor implementation threatens validity of results Establish a plan to evaluate measures to determine if they can effectively differentiate among teacher performance  Need to identify potential “widget effects” in measures  If measure is not differentiating among teachers, may be faulty training or poor implementation, not the measure itself Examine correlations among results from measures Evaluate processes and data each year and make needed adjustments

16 Evaluation System Models Austin (Student learning objectives with pay-for-performance, group and individual SLOs assess with comprehensive rubric) Delaware Model (Teacher participation in identifying grade/subject measures which then must be approved by state) Georgia CLASS Keys (Comprehensive rubric, includes student achievement— see last few pages) System: Rubric: pdf?p=6CC6799F8C1371F6B59CF81E4ECD54E63F615CF1D9441A9 2E28BFA2A0AB27E3E&Type=D pdf?p=6CC6799F8C1371F6B59CF81E4ECD54E63F615CF1D9441A9 2E28BFA2A0AB27E3E&Type=D Hillsborough, Florida (Creating assessments/tests for all subjects)

17 Evaluation System Models (cont’d) New Haven, CT (SLO model with strong teacher development component and matrix scoring; see Teacher Evaluation & Development System) Rhode Island DOE Model (Student learning objectives combined with teacher observations and professionalism) snt_Sup_August_24_rev.ppt Teacher Advancement Program (TAP) (Value-added for tested grades only, no info on other subjects/grades, multiple observations for all teachers) Washington DC IMPACT Guidebooks (Variation in how groups of teachers are measured—50% standardized tests for some groups, 10% other assessments for non-tested subjects and grades) CT+(Performance+Assessment)/IMPACT+Guidebooks

18 References Betebenner, D. W. (2008). A primer on student growth percentiles. Dover, NH: National Center for the Improvement of Educational Assessment (NCIEA). Braun, H., Chudowsky, N., & Koenig, J. A. (2010). Getting value out of value-added: Report of a workshop. Washington, DC: National Academies Press. Finn, Chester. (July 12, 2010). Blog response to topic “Defining Effective Teachers.” National Journal Expert Blogs: Education. Glazerman, S., Goldhaber, D., Loeb, S., Raudenbush, S., Staiger, D. O., & Whitehurst, G. J. (2011). Passing muster: Evaluating evaluation systems. Washington, DC: Brown Center on Education Policy at Brookings. Glazerman, S., Goldhaber, D., Loeb, S., Raudenbush, S., Staiger, D. O., & Whitehurst, G. J. (2010). Evaluating teachers: The important role of value-added. Washington, DC: Brown Center on Education Policy at Brookings.

19 References (continued) Goe, L. (2007). The link between teacher quality and student outcomes: A research synthesis. Washington, DC: National Comprehensive Center for Teacher Quality. Goe, L., Bell, C., & Little, O. (2008). Approaches to evaluating teacher effectiveness: A research synthesis. Washington, DC: National Comprehensive Center for Teacher Quality. Hassel, B. (Oct 30, 2009). How should states define teacher effectiveness? Presentation at the Center for American Progress, Washington, DC. how-should-states-define-teacher-effectiveness Howes, C., Burchinal, M., Pianta, R., Bryant, D., Early, D., Clifford, R., et al. (2008). Ready to learn? Children's pre-academic achievement in pre-kindergarten programs. Early Childhood Research Quarterly, 23(1), Kane, T. J., Taylor, E. S., Tyler, J. H., & Wooten, A. L. (2010). Identifying effective classroom practices using student achievement data. Cambridge, MA: National Bureau of Economic Research

20 References (continued) Koedel, C., & Betts, J. R. (2009). Does student sorting invalidate value-added models of teacher effectiveness? An extended analysis of the Rothstein critique. Cambridge, MA: National Bureau of Economic Research. McCaffrey, D., Sass, T. R., Lockwood, J. R., & Mihaly, K. (2009). The intertemporal stability of teacher effect estimates. Education Finance and Policy, 4(4), Pianta, R. C., Belsky, J., Houts, R., & Morrison, F. (2007). Opportunities to learn in America’s elementary classrooms. [Education Forum]. Science, 315, Prince, C. D., Schuermann, P. J., Guthrie, J. W., Witham, P. J., Milanowski, A. T., & Thorn, C. A. (2006). The other 69 percent: Fairly rewarding the performance of teachers of non-tested subjects and grades. Washington, DC: U.S. Department of Education, Office of Elementary and Secondary Education. Race to the Top Application Rivkin, S. G., Hanushek, E. A., & Kain, J. F. (2005). Teachers, schools, and academic achievement. Econometrica, 73(2),

21 References (continued) Sartain, L., Stoelinga, S. R., & Krone, E. (2010). Rethinking teacher evaluation: Findings from the first year of the Excellence in Teacher Project in Chicago public schools. Chicago, IL: Consortium on Chicago Public Schools Research at the University of Chicago. Schochet, P. Z., & Chiang, H. S. (2010). Error rates in measuring teacher and school performance based on student test score gains. Washington, DC: National Center for Education Evaluation and Regional Assistance, Institute of Education Sciences, U.S. Department of Education. Redding, S., Langdon, J., Meyer, J., & Sheley, P. (2004). The effects of comprehensive parent engagement on student learning outcomes. Paper presented at the American Educational Research Association Weisberg, D., Sexton, S., Mulhern, J., & Keeling, D. (2009). The widget effect: Our national failure to acknowledge and act on differences in teacher effectiveness. Brooklyn, NY: The New Teacher Project.

22 References (continued) Yoon, K. S., Duncan, T., Lee, S. W.-Y., Scarloss, B., & Shapley, K. L. (2007). Reviewing the evidence on how teacher professional development affects student achievement (No. REL 2007-No. 033). Washington, D.C.: U.S. Department of Education, Institute of Education Sciences, National Center for Education Evaluation and Regional Assistance, Regional Educational Laboratory Southwest.

23 Laura Goe, Ph.D National Comprehensive Center for Teacher Quality th Street NW, Suite 500 Washington, DC >