Evaluating teacher effectiveness with multiple measures Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive.

Slides:



Advertisements
Similar presentations
Assessment Adapted from text Effective Teaching Methods Research-Based Practices by Gary D. Borich and How to Differentiate Instruction in Mixed Ability.
Advertisements

PD Plan Agenda August 26, 2008 PBTE Indicators Track
Teacher Evaluation New Teacher Orientation August 15, 2013.
April 6, 2011 DRAFT Educator Evaluation Project. Teacher Education and Licensure DRAFT The ultimate goal of all educator evaluation should be… TO IMPROVE.
MEASURING TEACHING PRACTICE Tony Milanowski & Allan Odden SMHC District Reform Network March 2009.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Measuring Teacher Effectiveness in Untested Subjects and Grades.
Evaluating Teacher Effectiveness Laura Goe, Ph.D. Presentation to the Hawaii Department of Education July 20, 2011  Honolulu, HI.
Teacher Evaluation Models: A National Perspective Laura Goe, Ph.D. Research Scientist, ETS Principal Investigator for Research and Dissemination,The National.
How can we measure teachers’ contributions to student learning growth in the “non-tested” subjects and grades? Laura Goe, Ph.D. Research Scientist, ETS,
What should be the basis of
performance INDICATORs performance APPRAISAL RUBRIC
Matt Moxham EDUC 290. The Idaho Core Teacher Standards are ten standards set by the State of Idaho that teachers are expected to uphold. This is because.
Using Student Learning Growth as a Measure of Effectiveness Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive.
Linking Teacher Evaluation and Professional Growth IEL Washington Policy Seminar April 22, 2013  Washington, D.C. Laura Goe, Ph.D. Research Scientist,
TIMELESS LEARNING POLICY & PRACTICE. JD HOYE President National Academy Foundation.
Teacher Evaluation Systems: Opportunities and Challenges An Overview of State Trends Laura Goe, Ph.D. Research Scientist, ETS Sr. Research and Technical.
Session Materials  Wiki
Principal Evaluation in Massachusetts: Where we are now National Summit on Educator Effectiveness Principal Evaluation Breakout Session #2 Claudia Bach,
Meeting SB 290 District Evaluation Requirements
Aligning Academic Review and Performance Evaluation (AARPE)
Horizon Middle School June 2013 Balanced Scorecard In a safe, collaborative environment we provide educational opportunities that empower all students.
Interstate New Teacher Assessment and Support Consortium (INTASC)
Washington State Teacher and Principal Evaluation 1.
Measures of Teachers’ Contributions to Student Learning Growth Laura Goe, Ph.D. Research Scientist, Understanding Teaching Quality Research Group, ETS.
Student Learning Objectives (SLOs) Measuring Teacher Effectiveness Through the Use of Student Data Overview of the SLO Process April 7,
Measuring Teacher Effectiveness: Challenges and Opportunities
FewSomeAll. Multi-Tiered System of Supports A Comprehensive Framework for Implementing the California Common Core State Standards Professional Learning.
Stronge Teacher Effectiveness Performance Evaluation System
Teacher Evaluation in Rural Schools Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive Center for Teacher.
Evaluating the “Caseload” Educators Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive Center for Teacher.
Measures of Teachers’ Contributions to Student Learning Growth August 4, 2014  Burlington, Vermont Laura Goe, Ph.D. Senior Research and Technical Assistance.
NCATE Standard 3: Field Experiences & Clinical Practice Monica Y. Minor, NCATE Jeri A. Carroll, BOE Chair Professor, Wichita State University.
THE DANIELSON FRAMEWORK. LEARNING TARGET I will be be able to identify to others the value of the classroom teacher, the Domains of the Danielson framework.
What Are We Measuring? The Use of Formative and Summative Assessments Laura Goe, Ph.D. Research Scientist, ETS Principal Investigator for Research and.
South Western School District Differentiated Supervision Plan DRAFT 2010.
PERSONNEL EVALUATION SYSTEMS How We Help Our Staff Become More Effective Margie Simineo – June, 2010.
Hastings Public Schools PLC Staff Development Planning & Reporting Guide.
Washington State Teacher and Principal Evaluation Project Update 11/29/12.
March Madness Professional Development Goals/Data Workshop.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Evaluating Teacher Effectiveness: Some Models to Consider Laura.
Evaluating Teacher Effectiveness: Selecting Measures Laura Goe, Ph.D. SIG Schools Webinar August 12, 2011.
PGES: The Final 10% i21: Navigating the 21 st Century Highway to Top Ten.
Models for Evaluating Teacher Effectiveness Laura Goe, Ph.D. California Labor Management Conference May 5, 2011  Los Angeles, CA.
Welcome to todays session!  Please take a moment to check your connection and audio settings.  If this is your first time using LYNC please see the resources.
What Are the Characteristics of an Effective Portfolio? By Jay Barrett.
Evaluating Teacher Effectiveness Laura Goe, Ph.D. Presentation to National Conference of State Legislatures July 14, 2011  Webinar.
Changes in Professional licensure Teacher evaluation system Training at Coastal Carolina University.
Candidate Assessment of Performance CAP The Evidence Binder.
Teacher Evaluation Systems 2.0: What Have We Learned? EdWeek Webinar March 14, 2013 Laura Goe, Ph.D. Research Scientist, ETS Sr. Research and Technical.
Candidate Assessment of Performance CAP The Evidence Binder.
Weighting components of teacher evaluation models Laura Goe, Ph.D. Research Scientist, ETS Principal Investigator for Research and Dissemination, The National.
Measures of Teachers’ Contributions to Student Learning Growth Laura Goe, Ph.D. Research Scientist, Understanding Teaching Quality Research Group, ETS.
1 Far West Teacher Center Network - NYS Teaching Standards: Your Path to Highly Effective Teaching 2013 Far West Teacher Center Network Teaching is the.
Forum on Evaluating Educator Effectiveness: Critical Considerations for Including Students with Disabilities Lynn Holdheide Vanderbilt University, National.
UPDATE ON EDUCATOR EVALUATIONS IN MICHIGAN Directors and Representatives of Teacher Education Programs April 22, 2016.
Purpose of Teacher Evaluation and Observation Minnesota Teacher Evaluation Requirements Develop, improve and support qualified teachers and effective.
OEA Leadership Academy 2011 Michele Winship, Ph.D.
Implementing the Professional Growth Process Session 3 Observing Teaching and Professional Conversations American International School-Riyadh Saturday,
OEA Leadership Academy 2011 Michele Winship, Ph.D.
Using Student Growth in Teacher Evaluation and Development Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive.
Evaluating Teacher Effectiveness Laura Goe, Ph.D. Presentation to the Hawaii Department of Education July 20, 2011  Honolulu, HI.
Educator Recruitment and Development Office of Professional Development The NC Teacher Evaluation Process 1.
Clinical Practice evaluations and Performance Review
Phyllis Lynch, PhD Director, Instruction, Assessment and Curriculum
Evaluating Teacher Effectiveness: An Overview Laura Goe, Ph.D.
FEAPs (Florida Educator Accomplished Practices)
Teacher Evaluation: The Non-tested Subjects and Grades
Leveraging Performance Management to Support School Priorities
Aligning Academic Review and Performance Evaluation (AARPE)
Presentation transcript:

Evaluating teacher effectiveness with multiple measures Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive Center for Teacher Quality Performance Evaluation Advisory Committee Hartford, CT  March 20,

2 Laura Goe, Ph.D. Former teacher in rural & urban schools  Special education (7 th & 8 th grade, Tunica, MS)  Language arts (7 th grade, Memphis, TN) Graduate of UC Berkeley’s Policy, Organizations, Measurement & Evaluation doctoral program Principal Investigator for the National Comprehensive Center for Teacher Quality Research Scientist in the Performance Research Group at ETS 2

3 The National Comprehensive Center for Teacher Quality A federally-funded partnership whose mission is to help states carry out the teacher quality mandates of ESEA Vanderbilt University Learning Point Associates, an affiliate of American Institutes for Research Educational Testing Service

4 Today’s presentation available online To download a copy of this presentation or look at it on your iPad, smart phone or laptop now, go to  Go to Publications and Presentations page  Today’s presentation is at the bottom of the page 4

5 The goal of teacher evaluation

6 Agenda What we value in teachers and teaching The role of teaching standards An overview of teacher evaluation measures Observations of practice Indicators of professional responsibility Peer, student and parent feedback Multiple indicators of student learning An overview of teacher evaluation models Teacher professional development Weighting, performance levels, exceptions Final thoughts

7 Goe, Bell, & Little (2008) definition of teacher effectiveness 1.Have high expectations for all students and help students learn, as measured by value-added or alternative measures. 2.Contribute to positive academic, attitudinal, and social outcomes for students, such as regular attendance, on-time promotion to the next grade, on-time graduation, self-efficacy, and cooperative behavior. 3.Use diverse resources to plan and structure engaging learning opportunities; monitor student progress formatively, adapting instruction as needed; and evaluate learning using multiple sources of evidence. 4.Contribute to the development of classrooms and schools that value diversity and civic-mindedness. 5.Collaborate with other teachers, administrators, parents, and education professionals to ensure student success, particularly the success of students with special needs and those at high risk for failure.

8 Teaching standards A set of practices teachers should aspire to A teaching tool in teacher preparation programs A guiding document with which to align:  Measurement tools and processes for teacher evaluation, such as classroom observations, surveys, portfolios/evidence binders, student outcomes, etc.  Teacher professional growth opportunities, based on evaluation of performance on standards A tool for coaching and mentoring teachers:  Teachers analyze and reflect on their strengths and challenges and discuss with consulting teachers

9 Measures and models: Definitions Measures are the instruments, assessments, protocols, rubrics, and tools that are used in determining teacher effectiveness Models are the state or district systems of teacher evaluation including all of the inputs and decision points (measures, instruments, processes, training, and scoring, etc.) that result in determinations about individual teachers’ effectiveness

10 Multiple measures of teacher effectiveness Evidence of growth in student learning and competency  Standardized tests, pre/post tests in untested subjects  Student performance (art, music, etc.)  Curriculum-based tests given in a standardized manner  Classroom-based tests such as DIBELS Evidence of instructional quality  Classroom observations  Lesson plans, assignments, and student work  Student surveys such as Harvard’s Tripod  Evidence binder (next generation of portfolio) Evidence of professional responsibility  Administrator/supervisor reports, parent surveys  Teacher reflection and self-reports, records of contributions

11 Teacher observations: strengths and weaknesses Strengths  Great for teacher professional growth - If observation is followed by opportunity to discuss results - If support is provided for those who need it  Helps evaluator (principals or others) understand teachers’ needs across school or across district Weaknesses  Essential to have alignment between teaching standards and observation instrument  Resource intensive (personnel time, training, calibrating)  Validity of observation results may vary with who is doing them, depending on how well trained and calibrated they are

12 Example: University of Virginia’s CLASS observation tool Emotional Support Classroom Organization Instructional Support Pre-K and K-3 Positive Climate Negative Climate Teacher Sensitivity Regard for Student (Adolescent) Perspectives Behavior Management Productivity Instructional Learning Formats Concept Development Quality of Feedback Language Modeling Upper Elementary/ Secondary Content Understanding Analysis and Problem Solving Quality of Feedback

13 Example: Charlotte Danielson’s Framework for Teaching

14 Validity of classroom observations is highly dependent on training A teacher should get the same score no matter who observes him  This requires that all observers be trained on the instruments and processes  Occasional “calibrating” should be done; more often if there are discrepancies or new observers  Who the evaluators are matters less than the fact that they are trained to recognize evidence and score it consistently Teachers should also be trained on the observation forms and processes so they can participate actively and fully in the process

15 Risk management vs. one-size-fits-all in teacher observations Conducting high-quality observations is a resource-intensive process  A more efficient use of resources is for teachers who have not yet demonstrated competence to be on a more intensive observation schedule - New teachers - Teachers who have changed teaching assignments or schools Other measures are less resource intensive and can be used routinely (surveys, student outcomes, portfolios)

16 Reliability results when using different combinations of raters and lessons Figure 2. Errors and Imprecision: the reliability of different combinations of raters and lessons. From Hill et al., 2012 (see references list). Used with permission of author.

17 Formal vs. informal observations Formal observations are likely to be  Announced and scheduled in advance according to a pre-determined yearly schedule  Include pre- and post-conferences with review of lesson plans and artifacts  Last an entire class period  Result in a set of scores on multiple indicators Informal observations are likely to be  Unannounced, drop-in  Last less than an entire class period  Result in informal verbal or written feedback to the teacher, perhaps on only one indicator

18 Questions to ask about observations How many observations per year?  Vary by new vs. experience?  Vary by demonstrated competence?  Combination of formal and informal? Who should conduct the observations? Will multiple observers be required? How will they be trained?  Workshops? Online (video-based)? Will they need to be certified?

19 Valuing other professional contributions, other student outcomes Professional contributions  Working closely with parents and community  Participating in Response to Intervention teams  Curriculum teams (within & across grades)  Leadership in grade/subject/school Student outcomes  Successful outcomes for special populations may include social and behavioral outcomes  Improvements in attendance, behavior, participation in class, engagement, etc.

20 Measuring Professional Contributions Some observation instruments include this category (Charlotte Danielson’s, TAP, Marzano) while others don’t (CLASS) If professional contributions are not included, may want to consider teacher- constructed portfolios  Specific types of documents according to guidelines, not just a bunch of “kudos”  Document level of participation and contributions, not just “participation”

21 Peer feedback Research on using peer feedback in K-12 teaching is scarce (maybe non-existent) but can be found in higher ed for formative and summative (tenure) purposes Done in pairs or teams Two key areas  Classroom observation with feedback  Feedback on classroom artifacts (lesson plans, teaching resources, assessments, etc.)

22 Parent feedback Very little research on the use of parent surveys and feedback in teacher evaluation A study by Peterson et al (2003) suggests parents took the anonymous survey seriously (answers were not random)  Positive: Teachers liked being able to select the survey as one of the measures for their evaluation (from a menu of measures)  Negative: Principals were worried about sampling (could be a problem in elementary or whenever you have small classes)

23 Parent survey (from Peterson et al, 2003) Did you ask the teacher for and did the teacher give you (yes/no responses): 1. An overview of class content & goals? 2. Description of student’s progress? 3. Ideas for home support of learning? Answer Yes Somewhat No ( ) 4. Did your child know what was expected in this class? 5. Was the classroom work the right difficulty for your child? 6. Did the teacher treat your child with respect, care, and knowledge of child’s needs? 7. Were you satisfied with your child’s overall school experience as provided by this teacher? Do you have any comments for the teacher?

24 Tripod Student Survey (1) Harvard’s Tripod Survey – the 7 C’s – Caring about students (nurturing productive relationships); – Controlling behavior (promoting cooperation and peer support); – Clarifying ideas and lessons (making success seem feasible); – Challenging students to work hard and think hard (pressing for effort and rigor); – Captivating students (making learning interesting and relevant); – Conferring (eliciting students’ feedback and respecting their ideas); – Consolidating (connecting and integrating ideas to support learning)

25 Tripod Student Survey (2) Improved student performance depends on strengthening three legs of teaching practice: content, pedagogy, and relationships There are multiple versions: k-2, 3-5, 6-12 Measures:  student engagement  school climate  home learning conditions  teaching effectiveness  youth culture  family demographics Takes min There are English and Spanish versions Comes in paper form or in online version

26 Tripod Student Survey (3) Control is the strongest correlate of value added gains However, it is important to keep in mind that a good teacher achieves control by being good on the other dimensions

27 Tripod Student Survey (4) Different combinations of the 7 C's predict different outcomes (student learning is one outcome) Using the data, you can determine what a teacher needs to focus on to improve important outcomes Besides student learning, other important outcomes include:  happiness  good behavior  healthy responses to social pressures  self-consciousness  engagement/effort  satisfaction

28 Surveys as evaluation measures Relative to other measures, they’re very inexpensive (though there is a cost for Tripod) They can provide useful, actionable information to teachers and to principals about aspects of teachers’ performance not captured in other measures The survey can be repeated over time to show improvement in problem areas

29 Rhode Island’s SLO language “Student Learning Objectives are not set by educators in isolation; rather, they are developed by teams of administrators, grade- level teams or groups of content-alike teachers and, are aligned to district and school priorities, wherever possible.” (pg 12) From Rhode Island’s “Guide to Measures of Student Learning for Administrators and Teachers ” evaluation/Docs/GuideSLO.pdf evaluation/Docs/GuideSLO.pdf

30 Evidence of growth in student learning Evidence is strongest when it is  Standardized, meaning that all teachers used the assessment in exactly the same way - Gave the assessment on the same day - Gave students a specific amount of time to complete the test - Used the same preparation/instructions prior to the test - Recorded/reported results accurately  Valid, meaning that it measures what is intended - Items (questions) accurately capture students’ understanding and knowledge - Progress towards proficiency in a subject is captured because there are sufficient items to measure students at all levels  Recorded, meaning that student progress can be compared across classrooms and schools

31 Collect evidence in a standardized way (to the extent possible) Evidence of student learning growth  Locate or develop rubrics with explicit instructions and clear indicators of proficiency for each level of the rubric  Establish time for teachers to collectively examine student work and come to a consensus on performance at each level - Identify “anchor” papers or examples  Provide training for teachers to determine how and when assessments should be given, and how to record results in specific formats

32 The 4 Ps (Projects, Performances, Products, Portfolios) Yes, they can be used to demonstrate teachers’ contributions to student learning growth Here’s the basic approach  Use a high-quality rubric to judge initial knowledge and skills required for mastery of the standard(s)  Use the same rubric to judge knowledge and skills at the end of a specific time period (unit, grading period, semester, year, etc.)

4 types of musical behaviors: Types of assessment 1.Responding 2.Creating 3.Performing 4.Listening 1.Rubrics 2.Playing tests 3.Written tests 4.Practice sheets 5.Teacher Observation 6.Portfolios 7.Peer and Self- Assessment Assessing Musical Behaviors: The type of assessment must match the knowledge or skill Slide used with permission of authors Carla Maltas, Ph.D. and Steve Williams, M.Ed. See reference list for details.

34 The “caseload” educators For nurses, counselors, librarians and other professionals who do not have their own classroom, what counts for you is your “caseload”  May be all the students in the school  May be a specific set of students  May be other teachers  May be all of the above!

35 New Haven “matrix” Asterisks indicate a mismatch—teacher is very high on one area (practice or growth) and very low on the other area.

36 Considerations for choosing and implementing measures Consider whether human resources and capacity are sufficient to ensure fidelity of implementation Conserve resources by encouraging districts to join forces with other districts or regional groups Establish a plan to evaluate measures to determine if they can effectively differentiate among teacher performance Examine correlations among measures Evaluate processes and data each year and make needed adjustments

37 Measuring teachers’ contributions to student learning growth: A summary of current models ModelDescription Student learning objectives Teachers assess students at beginning of year and set objectives then assesses again at end of year; principal or designee works with teacher, determines success Subject & grade alike team models (“Ask a Teacher”) Teachers meet in grade-specific and/or subject-specific teams to consider and agree on appropriate measures that they will all use to determine their individual contributions to student learning growth Content CollaborativesContent experts (external) identify measures and groups of content teachers consider the measures from the perspective of classroom use; may not include pre- and post measures Pre-and post-tests modelIdentify or create pre- and post-tests for every grade and subject School-wide value-addedTeachers in tested subjects & grades receive their own value-added score; all other teachers get the school-wide average

38 Interpreting results for alignment with teacher professional learning options Different approach; not looking at “absolute gains” Requires ability to determine and/or link student outcomes to what likely happened instructionally Requires ability to “diagnose” instruction and recommend/and or provide appropriate professional growth opportunities  Individual coaching/feedback on instruction  Observing “master teachers”  Group professional development (when several teachers have similar needs)

39 Memphis professional development system: An aligned system Teaching and Learning Academy began April ‘96 Nationally commended program intended to  “…provide a collegial place for teachers, teacher leaders and administrators to meet, study, and discuss application and implementation of learning…to impact student growth and development” Practitioners propose and develop courses  Responsive to school/district evaluation results  Offerings must be aligned with NSDC standards  ~300+ On-line and in-person courses, many topics

40 Growth opportunities for all teachers Duke & Stiggins, 1986, p. 15

41 Measures that help teachers grow Measures that motivate teachers to examine their own practice against specific standards Measures that allow teachers to participate in or co-construct the evaluation (such as “evidence binders”) Measures that give teachers opportunities to discuss the results with evaluators, administrators, colleagues, teacher learning communities, mentors, coaches, etc. Measures that are aligned with professional development offerings Measures which include protocols and processes that teachers can examine and comprehend

42 Results inform professional growth opportunities Are evaluation results discussed with individual teachers? Do teachers collaborate with instructional managers to develop a plan for improvement and/or professional growth?  All teachers (even high-scoring ones) have areas where they can grow and learn Are effective teachers provided with opportunities to develop their leadership potential? Are struggling teachers provided with coaches and given opportunities to observe/be observed?

43 Effectiveness can be improved! Most teachers are doing the best they can  Help them do better with feedback, support, coaching, and a focus on classroom environment and relationships with students Teachers who are discouraged may need to see successful teachers with similar kids Teachers who are consistently effective should be encouraged to model and teach specific practices to less effective teachers Classroom learning environment is key: helping teachers create and maintain a better classroom learning environment improves student oucomes

44 Weights and measures There are no “rules” here; weights are likely to be determined by local priorities and beliefs Need to decide whether a high score on one measure/component can make up for a low score on another (“compensatory”) Need to decide whether to have a minimum score  High score on another component will not compensate The specific “mix” of measures may be locally determined within state guidelines  The mix should be evaluated year-to-year to see how the set of measures and weights are working

45 Teacher evaluation in isolated and/or low-capacity districts External evaluators may need to be brought in for very small, isolated districts  For example, a district where the superintendent is also principal, history teacher, and bus driver  May also be needed when evaluators’ objectivity is impacted by factors such as fear of losing teachers or damaging long-term relationships in the community Evaluators could be “exchanged” across districts within a specific region (“you evaluate mine, and I’ll evaluate yours”) or regional evaluators could serve a set of districts

46 Before you implement teacher and principal evaluation systems, ask yourself… How will this component of the teacher and principal evaluation system impact teaching and learning in classrooms and schools? How will this component look different in low- capacity vs. high-capacity schools? How will reporting on this component be done (to provide actionable information to teachers, principals, schools, districts, teacher preparation programs, and the state)? How will we know if this component is working as we intended?

47 Final thoughts The limitations:  There are no perfect measures  There are no perfect models  Changing the culture of evaluation is hard work The opportunities:  Evidence can be used to trigger support for struggling teachers and acknowledge effective ones  Multiple sources of evidence can provide powerful information to improve teaching and learning  Evidence is more valid than “judgment” and provides better information for teachers to improve practice

48 Resources and links Memphis Professional Development System  Main site:  PD Catalog: lr.pdf lr.pdf  Individualized Professional Development Resource Book: 11.pdf 11.pdf Charlotte Danielson’s Framework for Teaching CLASS Peer Review of Teaching (Higher Ed) learning.utas.edu.au/__data/assets/pdf_file/0010/1054/Peer_revi ew_of_teaching.pdfhttp:// learning.utas.edu.au/__data/assets/pdf_file/0010/1054/Peer_revi ew_of_teaching.pdf 48

49 Resources and links (cont’d) Harvard’s Tripod Survey National Response to Intervention Center Progress Monitoring Tools toolschart.htm toolschart.htm Colorado Content Collaboratives New York State approved teacher and principal practice rubrics Rhode Island Department of Education Teacher Evaluation – Student Learning Objectives Tennessee Teacher Evaluation

50 References Anderson, L. (1991). Increasing teacher effectiveness. Paris: UNESCO, International Institute for Educational Planning. Betebenner, D. W. (2008). A primer on student growth percentiles. Dover, NH: National Center for the Improvement of Educational Assessment (NCIEA). Braun, H., Chudowsky, N., & Koenig, J. A. (2010). Getting value out of value-added: Report of a workshop. Washington, DC: National Academies Press. Duke, DL; Stiggins, RJ. (1986.) Teacher Evaluation: Five Keys to Growth. West Haven, CT: National Education Association. ERIC #ED (full text, pg 15) Ellerson, N. M. (2009). Exploring the possibility and potential for pay for performance in America’s public schools. Washington, DC: American Association of School Administrators. Finn, Chester. (July 12, 2010). Blog response to topic “Defining Effective Teachers.” National Journal Expert Blogs: Education. Fuller, E., & Young, M. D. (2009). Tenure and retention of newly hired principals in Texas. Austin, TX: Texas High School Project Leadership Initiative. Glazerman, S., Goldhaber, D., Loeb, S., Raudenbush, S., Staiger, D. O., & Whitehurst, G. J. (2011). Passing muster: Evaluating evaluation systems. Washington, DC: Brown Center on Education Policy at Brookings.

51 References (continued) Goe, L. (2007). The link between teacher quality and student outcomes: A research synthesis. Washington, DC: National Comprehensive Center for Teacher Quality. Goe, L., Bell, C., & Little, O. (2008). Approaches to evaluating teacher effectiveness: A research synthesis. Washington, DC: National Comprehensive Center for Teacher Quality. Hassel, B. (Oct 30, 2009). How should states define teacher effectiveness? Presentation at the Center for American Progress, Washington, DC. how-should-states-define-teacher-effectiveness Herman, J. L., Heritage, M., & Goldschmidt, P. (2011). Developing and selecting measures of student growth for use in teacher evaluation. Los Angeles, CA: University of California, National Center for Research on Evaluation, Standards, and Student Testing (CRESST). Hill, H. C., Charalambous, C. Y., & Kraft, M. A. (2012). When rater reliability is not enough: Teacher observation systems and a case for the generalizability study. Educational Researcher, 41(2), Howes, C., Burchinal, M., Pianta, R., Bryant, D., Early, D., Clifford, R., et al. (2008). Ready to learn? Children's pre-academic achievement in pre-kindergarten programs. Early Childhood Research Quarterly, 23(1),

52 References (continued) Hunt, B. C. (2009). Teacher effectiveness: A review of the international literature and its relevance for improving education in Latin America. Washington, DC: Partnership for Educational Revitalization in the Americas (PREAL). Kane, T. J., Taylor, E. S., Tyler, J. H., & Wooten, A. L. (2010). Identifying effective classroom practices using student achievement data. Cambridge, MA: National Bureau of Economic Research. Koedel, C., & Betts, J. R. (2009). Does student sorting invalidate value-added models of teacher effectiveness? An extended analysis of the Rothstein critique. Cambridge, MA: National Bureau of Economic Research. McCaffrey, D., Sass, T. R., Lockwood, J. R., & Mihaly, K. (2009). The intertemporal stability of teacher effect estimates. Education Finance and Policy, 4(4), Peterson, K. D., Wahlquist, C., Brown, J. E., & Mukhopadhyay, S. (2003). Parent surveys for teacher evaluation. Journal of Personnel Evaluation in Education(17), Pianta, R. C., Belsky, J., Houts, R., & Morrison, F. (2007). Opportunities to learn in America’s elementary classrooms. [Education Forum]. Science, 315,

53 References (continued) Prince, C. D., Schuermann, P. J., Guthrie, J. W., Witham, P. J., Milanowski, A. T., & Thorn, C. A. (2006). The other 69 percent: Fairly rewarding the performance of teachers of non-tested subjects and grades. Washington, DC: U.S. Department of Education, Office of Elementary and Secondary Education. Race to the Top Application Rivkin, S. G., Hanushek, E. A., & Kain, J. F. (2005). Teachers, schools, and academic achievement. Econometrica, 73(2), Sartain, L., Stoelinga, S. R., & Krone, E. (2010). Rethinking teacher evaluation: Findings from the first year of the Excellence in Teacher Project in Chicago public schools. Chicago, IL: Consortium on Chicago Public Schools Research at the University of Chicago. Schochet, P. Z., & Chiang, H. S. (2010). Error rates in measuring teacher and school performance based on student test score gains. Washington, DC: National Center for Education Evaluation and Regional Assistance, Institute of Education Sciences, U.S. Department of Education.

54 References (continued) Springer, M., Lewis, J. L., Podgursky, M. J., Ehlert, M. W., Taylor, L. L., Lopez, O. S., et al. (2009). Governor’s Educator Excellence Grant (GEEG) Program: Year three evaluation report (Policy Evaluation Report). Nashville, TN: National Center on Performance Incentives. Redding, S., Langdon, J., Meyer, J., & Sheley, P. (2004). The effects of comprehensive parent engagement on student learning outcomes. Paper presented at the American Educational Research Association Weisberg, D., Sexton, S., Mulhern, J., & Keeling, D. (2009). The widget effect: Our national failure to acknowledge and act on differences in teacher effectiveness. Brooklyn, NY: The New Teacher Project. Yoon, K. S., Duncan, T., Lee, S. W.-Y., Scarloss, B., & Shapley, K. L. (2007). Reviewing the evidence on how teacher professional development affects student achievement (No. REL 2007-No. 033). Washington, D.C.: U.S. Department of Education, Institute of Education Sciences, National Center for Education Evaluation and Regional Assistance, Regional Educational Laboratory Southwest.

55 Questions?

56 Laura Goe, Ph.D National Comprehensive Center for Teacher Quality 1000 Thomas Jefferson Street, NW Washington, D.C