Presentation is loading. Please wait.

Presentation is loading. Please wait.

TEACHER EVALUATION IN NEW JERSEY: PREVIEW OF FINAL REPORT RUTGERS GRADUATE SCHOOL OF EDUCATION.

Similar presentations


Presentation on theme: "TEACHER EVALUATION IN NEW JERSEY: PREVIEW OF FINAL REPORT RUTGERS GRADUATE SCHOOL OF EDUCATION."— Presentation transcript:

1 TEACHER EVALUATION IN NEW JERSEY: PREVIEW OF FINAL REPORT RUTGERS GRADUATE SCHOOL OF EDUCATION

2 EDUCATIONAL ADMINISTRATION PROGRAM: RUTGERS GRADUATE SCHOOL OF EDUCATION Major Purpose: Introduce ourselves and potential to be of assistance. Specific Purpose: Talk about teacher evaluation: Bill Firestone: Preview of assessment of Teacher Evaluation Pilot (EE4NJ) Bruce Baker: How to understand value-added and growth data for teacher evaluation Melinda Mangin: Using teacher evaluation to improve teaching

3 Graduate School of Education OVERVIEW 2-Year Assessment of Teacher Evaluation Process Sample: –10 districts in 1 st year. –25 districts in 2 nd year. –Diverse with respect to size, wealth, and location. Mix of surveys (of teachers and administrators) and site visits. Focused more on local use of teacher practice measures than student growth measures. Asked about: –Perceptions of implementation –Orientations toward program. –Perceived facilitators & barriers 3

4 Graduate School of Education Implementation Ability to get complete set of observations of all teachers. –Year 1: Based on analysis of teacher practice data submitted by 10 districts to NJDOE Final summary evaluation provided for between 60% and 91% of teachers per district –Year 2: Based on administrator survey in 25 districts Number of project directors that expected to complete all required observations for each teacher: –2 nd year districts 8 of 11 –1 st year districts: 3 of 13 –In sum: Completing all observations is problematic. May be easier since fewer are required than in Year 2 of pilot. 4

5 Graduate School of Education Factors Affecting Perceived Accuracy & Fairness Consistency of raters in districts. –In 4 districts, focus group teachers noted that observers would disagree in rating same teacher. –In 3 districts, teachers noted that observers were consistent. Distance from classroom. –Measurement experts say observers with less contact = fewer incentives to raise ratings. –Focus group teachers noted that observers with less contact would not know unusual contextual factors that might affect their practice. –80% of surveyed teachers said important to “know my classroom well.” Observer’s content knowledge: –86% of surveyed teachers said it was important to have “content knowledge in your content area.” 5

6 Graduate School of Education ORIENTATIONS II: Purpose & Tenure Equal proportions of teachers think a major purpose of teacher evaluation is to: –Make tenure and promotion decisions: 57% –Provide information to help teachers improve practice: 58% Most tenured teachers think they are unlikely to lose tenure under the new system: 68% More new teachers think the new system will increase their chances of getting tenure--26% than decrease their chances-- 7%. Less than half teachers think evaluation data helped their teaching: –40-41% think observation data is helpful. –46-49% think student growth data is helpful 6

7 Graduate School of Education FACILITATORS/BARRIERS: Training on Observations Administrators more satisfied with their training than teachers. Percent getting 25 or more hours: –2 nd yr administrators: 37 –1 st yr administrators58 –2 nd year teachers03 –1 st year teachers07 Training formats: –Important to continue training after initial orientation. –Issue may not be face-to-face v. online or media-based training. –Challenge is to provide opportunities for active learning: Practice scoring (for administrators AND teachers) Peer discussions Model what teacher practice instruments say is good practice. 7

8 Graduate School of Education FACILITATORS/BARRIERS: Time From 86% to 96% of administrators said they were spending more time doing observations or entering data, depending on cohort & year. Things affected by time crunch (from interviews): –Getting all observations done –Providing good documentation –Providing teachers with helpful feedback –Other work, including handling discipline issues, meetings with parents, being visible in building. Things that helped get observations done: –An effective district-wide schedule for observations –Clear central administration focus on and support for doing observations –Sufficient staff or capacity to hire or reconfigure to create special observer cadre. 8

9 Graduate School of Education FACILITATORS/BARRIERS: Data management tool Can’t work with teacher practice data without data management tool: Teachscape, iObservation etc. Learning data management tools = a substantial 1 st year challenge. The tools continue to raise problems. Percent finding data management tool easy to use for: –Purposes related to providing feedback to teachers: 31-36% –Purposes related to analyzing data and sharing reports: 14-17% Reasons: –Lack of practice –Design flaws like excessive attention to individual confidentiality v. capacity to analyze. 9

10 Graduate School of Education FINAL NOTE: DISTRICT VARIATION Almost everything we looked at, substantial variation among districts. Appeared to be some districts that consistently got more observations done & were perceived to be more accurate, fair, and helpful to teachers. Have not yet been able to analyze what those districts did, but two hypotheses: –Pattern of integrating teacher evaluation into overall district improvement strategy. –Sufficient human capacity: numbers and knowledge of administrators in particular. 10


Download ppt "TEACHER EVALUATION IN NEW JERSEY: PREVIEW OF FINAL REPORT RUTGERS GRADUATE SCHOOL OF EDUCATION."

Similar presentations


Ads by Google