Selecting Growth Measures for School and Teacher Evaluations: Should Proportionality Matter? Mark Ehlert Cory Koedel Eric Parsons Michael Podgursky Department.

Slides:



Advertisements
Similar presentations
Place and Economic Activity: Key issues from the area effects debate Nick Buck ISER, University of Essex.
Advertisements

The Framework for Teaching Charlotte Danielson
Testing for Tomorrow Growth Model Testing Measuring student progress over time.
William Fallon May 08, 2012 Brown University Department of Economics A Theoretical and Empirical Analysis of Cross-Country Welfare.
Labor Market Equilibrium. We start with the assumption that each labor market is competitive. What does this mean? How is equilibrium price set? Why is.
Cory Koedel, Eric Parsons, Michael Podgursky and Mark Ehlert
Mark D. Reckase Michigan State University The Evaluation of Teachers and Schools Using the Educator Response Function (ERF)
The choice between fixed and random effects models: some considerations for educational research Claire Crawford with Paul Clarke, Fiona Steele & Anna.
1 Empirical Approaches to Assessing Discrimination.
A Guide to Education Research in the Era of NCLB Brian Jacob University of Michigan December 5, 2007.
Baseline for school surveys - Young Lives longitudinal survey of children, households & communities every 3 years since ,000 children Ethiopia,
Valuing and Evaluating the Liberal Arts in the Research University The Liberal Arts & Sciences in the Research University Today: Histories, Challenges,
Designing Influential Evaluations Session 5 Quality of Evidence Uganda Evaluation Week - Pre-Conference Workshop 19 th and 20 th May 2014.
BUSINESS AND FINANCIAL LITERACY FOR YOUNG ENTREPRENEURS: EVIDENCE FROM BOSNIA-HERZEGOVINA Miriam Bruhn and Bilal Zia (World Bank, DECFP)
1 Chapter 3 – Tools of Normative Analysis Public Finance McGraw-Hill/Irwin © 2005 The McGraw-Hill Companies, Inc., All Rights Reserved.
Maths Counts Insights into Lesson Study
Promoting a Positive Organizational Culture Prof. Brian Blume, Ph.D University of Michigan, Flint School of Management.
Using the WV Growth Model to Measure Student Achievement Nate Hixson Assistant Director, Office of Research.
APPRAISAL An evaluation into the progress and work of an employee over a period of time. Appraisal is part of the job for any person who is responsible.
ES INC: Economic and discounted cash flow techniques: a comparison with respect to the Requirements of the Management Control System.
Using Growth Models for Accountability Pete Goldschmidt, Ph.D. Assistant Professor California State University Northridge Senior Researcher National Center.
Efficient Estimation of Emission Probabilities in profile HMM By Virpi Ahola et al Reviewed By Alok Datar.
PART TWO EMPLOYMENT Chapters 5-7.
Gareth D. Myles University of Exeter.  The book advances the theory of cost benefit in directions beneficial for EU policy  It builds on the best of.
What Makes For a Good Teacher and Who Can Tell? Douglas N. Harris Tim R. Sass Dept. of Ed. Policy Studies Dept. of Economics Univ. of Wisconsin Florida.
Welcome to the Presentation on Marginal Costing
Experimental Design The Gold Standard?.
Classroom Assessment A Practical Guide for Educators by Craig A
John Cronin, Ph.D. Director The Kingsbury NWEA Measuring and Modeling Growth in a High Stakes Environment.
Assessment Group for Provincial Assessments, June Kadriye Ercikan University of British Columbia.
1 Comments on: “New Research on Training, Growing and Evaluating Teachers” 6 th Annual CALDER Conference February 21, 2013.
Human Capital Policies in Education: Further Research on Teachers and Principals 5 rd Annual CALDER Conference January 27 th, 2012.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Non-Experimental Methods Florence Kondylis.
Global Workshop on Development Impact Evaluation in Finance and Private Sector Rio de Janeiro, June 6-10, 2011 Mattea Stein Quasi Experimental Methods.
Sensitivity of Teacher Value-Added Estimates to Student and Peer Control Variables October 2013 Matthew Johnson Stephen Lipscomb Brian Gill.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Causal Inference and Adequate Yearly Progress Derek Briggs University of Colorado at Boulder National Center for Research on Evaluation, Standards, and.
Slide 1 Estimating Performance Below the National Level Applying Simulation Methods to TIMSS Fourth Annual IES Research Conference Dan Sherman, Ph.D. American.
Parents’ basic skills and children’s test scores Augustin De Coulon, Elena Meschi and Anna Vignoles.
6. Conceptualization & Measurement
Communicative Language Teaching
Measuring Student Growth and its Role in the Evaluation Process.
A Research-Based Response to Federal Non- Regulatory Guidance on Growth Models Mark Ehlert Cory Koedel Eric Parsons Michael Podgursky.
Issues in Assessment Design, Vertical Alignment, and Data Management : Working with Growth Models Pete Goldschmidt UCLA Graduate School of Education &
WORKSHOP ON TEACHING AND RESEARCH OF TRADE AND POVERTY: Conceptual and Methodological approaches and Policy Implications Peacock, Hotel, Dar-es-Salaam,
Research on teacher pay-for-performance Patrick McEwan Wellesley College (Also see Victor Lavy, “Using performance-based pay to improve.
Human Capital Policies in Education: Further Research on Teachers and Principals 5 rd Annual CALDER Conference January 27 th, 2012.
The Choice Between Fixed and Random Effects Models: Some Considerations For Educational Research Clarke, Crawford, Steele and Vignoles and funding from.
“Value added” measures of teacher quality: use and policy validity Sean P. Corcoran New York University NYU Abu Dhabi Conference January 22, 2009.
Impediments to the estimation of teacher value added Steven Rivkin Jun Ishii April 2008.
Using School Choice Lotteries to Test Measures of School Effectiveness David Deming Harvard University and NBER.
The measurement and comparison of health system responsiveness Nigel Rice, Silvana Robone, Peter C. Smith Centre for Health Economics, University of York.
LISA A. KELLER UNIVERSITY OF MASSACHUSETTS AMHERST Statistical Issues in Growth Modeling.
Public Finance and Public Policy Jonathan Gruber Third Edition Copyright © 2010 Worth Publishers 1 of 24 Copyright © 2010 Worth Publishers.
Copyright © 2014 American Institutes for Research and Cleveland Metropolitan School District. All rights reserved. March 2014 Interpreting Vendor Assessment.
Evaluation and Assessment Evaluation is a broad term which involves the systematic way of gathering reliable and relevant information for the purpose.
 Mark D. Reckase.  Student achievement is a result of the interaction of the student and the educational environment including each teacher.  Teachers.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
IEc INDUSTRIAL ECONOMICS, INCORPORATED Attributing Benefits to Voluntary Programs: Practical and Defensible Approaches Cynthia Manson, Principal June 23,
Having it both ways? Balancing personal and party representation
Access to Quality Teaching: Myths, Facts, and Potential Policy Solutions for the Future Dr. Michael Hansen Senior Fellow and Director, Brown Center on.
Measuring Results and Impact Evaluation: From Promises into Evidence
Quasi Experimental Methods I
How High Schools Explain Students’ Initial Colleges and Majors
Holli G. Bayonas, Ph.D & Eric S. Howard, M.A.
Quasi Experimental Methods I
International Livestock Research Institute
Maths Counts Insights into Lesson Study
Stat 112 Notes 4 Today: Review of p-values for one-sided tests
Ghent University, Belgium
Presentation transcript:

Selecting Growth Measures for School and Teacher Evaluations: Should Proportionality Matter? Mark Ehlert Cory Koedel Eric Parsons Michael Podgursky Department of Economics, University of Missouri -Columbia 1

Motivation Growth models are increasingly being incorporated into district, school and teacher evaluations across the United States. The question of how to model student test- score growth has resulted in lively policy debates What are objectives of the evaluation system? 2

Summary of Findings We argue that the three key objectives of an evaluation system in education are: Elicit optimal effort from agents Provide useful performance signals to educational actors Avoid exacerbating pre-existing inequities in the labor markets faced by advantaged and disadvantaged schools Given these objectives, the proper growth model for use in evaluation systems is neither the sparse model or a traditional VAM model. Instead, it is what we call the “proportional” VAM model. 3

A Model Menu The growth-model choice set essentially comes down to these three choices: 1)The sparse model (e.g., SGPs) 2)A single-equation VAM model (e.g., a standard value-added model from the research literature). 3)The proportional model (e.g., a two-step fixed effects model or random- effects model, less common in research) 4

5

Comparing the One-Step and Two-Step VAMs The key difference is that the two-step VAM partials out variation in test scores attributable to student and school characteristics before estimating the school effects. Specific example: Suppose that high-poverty schools really are of lower quality (causally). – In the one-step VAM, the model identifies poverty effects (F/R lunch) using within-school variation in student poverty status so it can separately identify differences in school quality between high- and low-poverty students – In the two-step VAM, the first step attributes any and all systematic performance differences between high- and low-poverty students to the first-step variables (e.g., it purges them from the residuals), including systematic differences in school quality. The implication is that high- and low-poverty schools are only compared to each other in the model output – not to dissimilar schools. 6

Missouri Schools, Median SGPs r =

Missouri Schools, one-step fixed effects VAM r =

Missouri Schools, two-step fixed effects VAM r =

Implications 10

Objective #1: Elicit optimal educator effort Barlevy and Neal (2012) cover this issue extensively. There is also a large literature in economics, outside of the education- evaluation context, that is very clear on how to design evaluation systems when some competitors are at an inherent disadvantage (e.g., see Schotter and Weigelt (1992), who study this issue in the context of affirmative action policy). A central lesson from these studies is that the right signal must be sent to agents in different circumstances to elicit optimal effort. This signal need not be a direct measure of absolute productivity; instead, it should be an indicator of performance relative to equally-circumstanced peers. This is precisely what the proportional model does (based on observable circumstances). 11

Objective #1: Elicit optimal educator effort Limitation: There is some evidence that the effort response margin in education in the United States is weak (Springer et al., 2010; on other hand … Fryer, et. al., 2012). 12

Objective #2: Provide useful performance signals It is a common conventional wisdom that growth-model output doesn’t help educational actors improve. Is this really true? – Growth model output can: Encourage effective schools (districts/teachers) to continue to refine and augment existing instructional strategies Serve as a point of departure for interventions/overhauls in ineffective schools (districts/teachers) Facilitate productive educator-to-educator learning by pairing low- and high-performing schools (districts/teachers). – The signaling value of an evaluation system is particularly important when it is difficult for individual schools (districts/teachers) to assess their performance, and the performance of others, accurately. 13

Objective #2: Provide useful performance signals We argue that the most useful performance signals come from the two-step “proportional” model. This is true even under the maintained assumption that the one-step VAM produces causal estimates. A key reason is that the causal estimates from the one-step VAM do not account for the counterfactual. – Example: Disadvantaged schools face weaker educator labor markets (Boyd et al., 2005; Jacob, 2007; Koedel et al., 2011; Reininger, 2012) Sparse models provide the least-useful performance signals (not controversial: acknowledged in SGP literature) 14

Example What do we tell Rough Diamond and Gold Leaf? What do we tell other schools about Rough Diamond and Gold Leaf? 15

Objective #3: Labor-market inequities The labor-market difficulties faced by disadvantaged schools have been well-documented (Boyd et al., 2005; Jacob, 2007; Koedel et al., 2011; Reininger, 2012). As stakes become attached to school rankings based on growth models, systems that disproportionately identify poor schools as “losers” will make positions at these schools even less desirable to prospective educators. 16

Summary thus far… We identify three key objectives of an evaluation system in education: 1.Elicit optimal effort from agents 2.Provide useful performance signals to educational actors 3.Avoid exacerbating pre-existing inequities in the labor markets faced by advantaged and disadvantaged schools When one considers these key objectives, the “proportionality” feature of the two-step model is preferred on all three. 17

But what about… The fact remains that schools serving disadvantaged students really do have lower test scores, and lower unconditional growth, than schools serving advantaged students. There seems to be general concern that this information will be hidden if we construct proportional growth models. Our view is that this concern is largely misguided. 18

Test-Score Levels 19

Concluding Remarks Growth models are quickly (very quickly) moving from the research space to the policy space. – The policy uses for growth models are not the same as the research uses for growth models. Starting with the right question is important: “What are the objectives of the evaluation system?” Beginning with this question, in our view, leads us to conclude that a “proportional” growth model is best-suited for use in educational evaluation programs for districts, schools and teachers. 20