The National Conference on Value-Added UW-Madison, April 22-24 Program Chairs Douglas Harris, Adam Gamoran, Steve Raudenbush Program Committee Members.

Slides:



Advertisements
Similar presentations
Objective: To learn and use other important principles of experimental design. HW: Read the rest of section 5.2 and complete pink ditto (double sided).
Advertisements

Agenda For Today! School Improvement Harris Poll Data PDSA PLC
Teacher Effectiveness and the Equitable Distribution of Effective Teachers 2009 National Forum on Education Policy Education Commission of the States July.
Experimental and Quasiexperimental Designs Chapter 10 Copyright © 2009 Elsevier Canada, a division of Reed Elsevier Canada, Ltd.
Postgraduate Course 7. Evidence-based management: Research designs.
Designing an impact evaluation: Randomization, statistical power, and some more fun…
Victorian Curriculum and Assessment Authority
Lecturette 2: Educational Mandates & Universally Designed Large Scale Assessments.
1 A Systematic Review of Cross- vs. Within-Company Cost Estimation Studies Barbara Kitchenham Emilia Mendes Guilherme Travassos.
Teacher Training, Teacher Quality and Student Achievement Douglas Harris Tim R. Sass Dept. of Educational Dept. of Economics Policy Studies Florida State.
T-tests continued.
Mywish K. Maredia Michigan State University
Webinar: How to handle PRP appeals Presented by Heather Mitchell, employment lawyer at Browne Jacobson.
Completing the Classroom Teacher and Non-Classroom Teacher Evaluations for Presented by: The Office of Talent Development Employee Evaluations.
Douglas N. Harris University of Wisconsin at Madison Evaluating and Improving Value-Added Modeling.
Student Learning Targets (SLT) You Can Do This! Getting Ready for the School Year.
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
Comparing Growth in Student Performance David Stern, UC Berkeley Career Academy Support Network Presentation to Educating for Careers/ California Partnership.
© Cambridge International Examinations 2013 Component/Paper 1.
Experimental Design Week 9 Lecture 1.
Writing with APA style (cont.) & Experiment Basics: Variables Psych 231: Research Methods in Psychology.
Research in Psychology. Questions What can we find out with research? Why should we believe scientists? Isn’t Psychology just common sense?
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Basic Business Statistics.
Statistics for Managers Using Microsoft® Excel 5th Edition
What Makes For a Good Teacher and Who Can Tell? Douglas N. Harris Tim R. Sass Dept. of Ed. Policy Studies Dept. of Economics Univ. of Wisconsin Florida.
Jingzi Huang, Ph.D School of Teacher Education, CEBS UNC 2013 Fall GTA Conference Presentation.
Postgraduate Course 6. Evidence based management: What is the best available evidence?
-- Preliminary, Do Not Quote Without Permission -- VALUE-ADDED MODELS AND THE MEASUREMENT OF TEACHER QUALITY Douglas HarrisTim R. Sass Dept. of Ed. LeadershipDept.
School-Wide Positive Behavioral Support School-wide Evaluation Tool January 2007.
Educational Psychology: Theory and Practice Chapter 1
1 Writing Effective Critical Elements Using the SMART or MARST Formats.
John Cronin, Ph.D. Director The Kingsbury NWEA Measuring and Modeling Growth in a High Stakes Environment.
The First Days of School: Harry Wong July 29, 2009 (#2) Research Based Practices.
Evaluating Teacher Performance Daniel Muijs, University of Southampton.
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
Evaluation of software engineering. Software engineering research : Research in SE aims to achieve two main goals: 1) To increase the knowledge about.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Evaluating a Research Report
CHAPTER 14: Nonparametric Methods
Methodology Part 1. Hindsight Bias “I knew it all along” The tendency to believe, after learning an outcome, that we knew the outcome.
Taking it to Another Level: Increasing Higher Order Thinking Session 3 BHCA PROFESSIONAL DEVELOPMENT SERIES.
Cambridge Pre-U Getting Started In-service Training Liberating learning Developing successful students.
Committee on the Assessment of K-12 Science Proficiency Board on Testing and Assessment and Board on Science Education National Academy of Sciences.
Selecting and Recruiting Subjects One Independent Variable: Two Group Designs Two Independent Groups Two Matched Groups Multiple Groups.
Line-up:. Tuesday, April 17 5:30 – 9:30 p.m.  Define characteristics of an Effective Teacher  Review and understand course syllabus, requirements,
Laying the Foundation for Scaling Up During Development.
An Introduction to Statistics and Research Design
EVAL 6000: Foundations of Evaluation Dr. Chris L. S. Coryn Nick Saxton Fall 2014.
“Value added” measures of teacher quality: use and policy validity Sean P. Corcoran New York University NYU Abu Dhabi Conference January 22, 2009.
Quantitative & Qualitative Approaches Harlina Nathania Lukman.
DWW: Doing What Works Recommendation 1. Make data part of an ongoing cycle of instructional improvement. Recommendation 2. Teach students to examine their.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Chapter 1: The Science of Biology. Science What is science? –An organized way of using evidence to learn about the natural world What is the goal of science?
Variables It is very important in research to see variables, define them, and control or measure them.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
S-005 Types of research in education. Types of research A wide variety of approaches: –Theoretical studies –Summaries of studies Reviews of the literature.
WHAT REALLY WORKS IN SPECIAL AND INCLUSIVE EDUCATION DoSE Meeting February 26, 2016 Lori Dehart, Behavior Consultant.
CHAPTER ONE: INTRODUCTION TO ACTION RESEARCH CONNECTING THEORY TO PRACTICE IMPROVING EDUCATIONAL PRACTICE EMPOWERING TEACHERS.
Oneway ANOVA comparing 3 or more means. Overall Purpose A Oneway ANOVA is used to compare three or more average scores. A Oneway ANOVA is used to compare.
Connecting Theory to Practice Improving Educational Practice Empowering Teachers Chapter One: Introduction to Action Research.
1 IT/Cybersecurity - ICRDCE Conference Day Using Blooms to Write Student Learning Outcomes (SLO’s)
Assessing Impact: Approaches and Designs
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Teacher Evaluation System
Doing your best for every student –
Instructional Personnel Performance Appraisal System
Teaching and Educational Psychology
Making Causal Inferences and Ruling out Rival Explanations
2018 OSEP Project Directors’ Conference
Student Assessment Methods David W. Dillard AVCTC.
Presentation transcript:

The National Conference on Value-Added UW-Madison, April Program Chairs Douglas Harris, Adam Gamoran, Steve Raudenbush Program Committee Members Tim Sass, Rob Meyer, Henry Braun, JR Lockwood Funders Carnegie Corp, Joyce and Spencer Foundations

Why Are We Here? (Part I) Student achievement is an important objective of schools and a tool in education policy Student achievement is an important objective of schools and a tool in education policy - Beginning in early 1990s, it became common to use snapshots of student achievement to measure and reward school improvement - NCLB has strongly reinforced shift toward “accountability” Because student achievement is an important objective, researchers use this as an important outcome variable in program evaluations Because student achievement is an important objective, researchers use this as an important outcome variable in program evaluations - The basis for education production function studies dating back at least to the Coleman Report

Why are We Here? (Part II) Common uses of student test scores for accountability and program evaluation are hindered by selection bias: Common uses of student test scores for accountability and program evaluation are hindered by selection bias: (1) Students are non-randomly assigned to teachers (2) Schools and classrooms are non-randomly assigned to programs In theory, value-added addresses these selection bias problems In theory, value-added addresses these selection bias problems

An Important Distinction As we will see, the “right” value-added model depends on how it is being used—that is, on the types of conclusions being drawn As we will see, the “right” value-added model depends on how it is being used—that is, on the types of conclusions being drawn Value-added for program evaluation (VAM-P) Value-added for program evaluation (VAM-P) Value-added for accountability (VAM-A) Value-added for accountability (VAM-A)

Encouraging Observations about VAM-A Differences between the lowest and highest value-added teachers seem large Differences between the lowest and highest value-added teachers seem large Some evidence that having a high-value-added teacher has lasting benefits for students Some evidence that having a high-value-added teacher has lasting benefits for students VAM-A measures of teacher effectiveness are positively correlated with principals’ subjective assessments of teachers VAM-A measures of teacher effectiveness are positively correlated with principals’ subjective assessments of teachers VAM-A measures have been validated in a random assignment experiment VAM-A measures have been validated in a random assignment experiment

Less Encouraging Observations Measured “teacher effects” are imprecise Measured “teacher effects” are imprecise - hard to say that one teacher is clearly better than another based on VAM-A Related point: teacher effects are unstable Related point: teacher effects are unstable Value-added of individual teachers varies-- sometimes considerably--depending on how the model is specified Value-added of individual teachers varies-- sometimes considerably--depending on how the model is specified

Possibly Problematic Assumptions VAM requires many strong assumptions, such as: VAM requires many strong assumptions, such as: - test scale - assignment based on fixed student and teacher characteristics - each teacher equally effective with all types of students - constant rate of decay in effects of past inputs Violation of these assumptions may explain some of the less encouraging results Violation of these assumptions may explain some of the less encouraging results

Critical Question How do we reconcile the encouraging findings with the less encouraging findings and problematic assumptions?

Broader Conference Goals To create greater clarity about the properties of VAM-A and VAM-P, from a statistical standpoint To create greater clarity about the properties of VAM-A and VAM-P, from a statistical standpoint To narrow the range of appropriate value-added models (may vary by VAM-A and VAM-P) To narrow the range of appropriate value-added models (may vary by VAM-A and VAM-P) To provide a sense of the policy implications of these technical findings for use of VAM-A To provide a sense of the policy implications of these technical findings for use of VAM-A Critical Point We must view VAM-A and VAM-P in comparison to the alternatives in comparison to the alternatives

Conference Ground Rules Each presenter has 20 minutes, followed by 10 minutes for discussants, followed by Q&A Each presenter has 20 minutes, followed by 10 minutes for discussants, followed by Q&A - please hold questions until the end of each session Remember, we have researchers from a wide variety of disciplines and fields—try to avoid jargon and explain the intuition Remember, we have researchers from a wide variety of disciplines and fields—try to avoid jargon and explain the intuition Remember our two purposes: technical issues and and policy issues, VAM-A and VAM-P Remember our two purposes: technical issues and and policy issues, VAM-A and VAM-P