How to Norm Rubrics Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.

Slides:



Advertisements
Similar presentations
Session Learning Target You will gain a better understanding of identifying quality evidence to justify a performance rating for each standard and each.
Advertisements

Item Analysis.
Carol Ann Gittens, Gail Gradowski & Christa Bailey Santa Clara University WASC Academic Resource Conference Session D1 April 25, 2014.
A Guide for College Assessment Leaders Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.
Vivian Mun, Ed.D. Accreditation What is a rubric? A rubric is a scoring tool that lists the criteria for a piece of work, or “what counts” (for.
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
Susan Malone Mercer University.  “The unit has taken effective steps to eliminate bias in assessments and is working to establish the fairness, accuracy,
Developing Rubrics Presented by Frank H. Osborne, Ph. D. © 2015 EMSE 3123 Math and Science in Education 1.
Crowd-Sourcing Innovative Practices: Assessing Integrative Learning at Large Research Institutions.
Measuring Student Learning March 10, 2015 Cathy Sanders Director of Assessment.
Applying Assessment to Learning
We accept learning as the fundamental purpose of our school and therefore are willing to examine all practices in light of the impact on learning.. - DuFour,
Graduate Program Assessment Report. University of Central Florida Mission Communication M.A. Program is dedicated to serving its students, faculty, the.
Item Analysis Ursula Waln, Director of Student Learning Assessment
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Standard Setting Inclusive Assessment Seminar Marianne.
Uses of Language Tests.
Dr. Robert Mayes University of Wyoming Science and Mathematics Teaching Center
Descriptive Rubrics Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.
performance INDICATORs performance APPRAISAL RUBRIC
Learning Outcomes from Report-Writing Unit
FLCC knows a lot about assessment – J will send examples
Assessment Literacy Series
Rhode Island Model Academy for Personnel Evaluating Teachers Day One Professional Practice.
Becoming a Teacher Ninth Edition
Ch 6 Validity of Instrument
Student Learning targets
October 31, Dialog about SLOs, assessment, and existing practices at TC Identify course level SLO to assess this semester Align SLO with TC’s institutional.
ERIKA HALL CENTER FOR ASSESSMENT PRESENTATION AT THE 2014 NATIONAL CONFERENCE ON STUDENT ASSESSMENT NEW ORLEANS JUNE 25, 2014 The Role of a Theory of Action.
January 29, 2010ART Beach Retreat ART Beach Retreat 2010 Assessment Rubric for Critical Thinking First Scoring Session Summary ART Beach Retreat.
Classroom Assessments Checklists, Rating Scales, and Rubrics
Assessing Standards Through Rubrics Milton Hershey School Houseparent Professional Development Session #1 Denise G. Meister, Ph.D. Penn State.
Classroom Assessment A Practical Guide for Educators by Craig A
Note: Because of slide animation, this ppt is intended to be viewed as a slide show.  While viewing the ppt, it may be helpful to obtain a sample Core.
Building Your Assessment Plan Esther Isabelle Wilder Lehman College.
Information for school leaders and teachers regarding the process of creating Student Learning Targets. Student Learning targets.
Evaluating a Research Report
Measuring Complex Achievement
Assessment in Education Patricia O’Sullivan Office of Educational Development UAMS.
March 26-28, 2013 SINGAPORE CDIO Asian Regional Meeting and Workshop on Engineering Education and Policies for Regional Leaders Programme Evaluation (CDIO.
Teaching Today: An Introduction to Education 8th edition
Student Growth in the Washington State Teacher Evaluation System Michelle Lewis Puget Sound ESD
Standard Setting Results for the Oklahoma Alternate Assessment Program Dr. Michael Clark Research Scientist Psychometric & Research Services Pearson State.
An overview of multi-criteria analysis techniques The main role of the techniques is to deal with the difficulties that human decision-makers have been.
VALUE/Multi-State Collaborative (MSC) to Advance Learning Outcomes Assessment Pilot Year Study Findings and Summary These slides summarize results from.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
STANDARD 4 & DIVERSITY in the NCATE Standards Boyce C. Williams, NCATE John M. Johnston, University of Memphis Institutional Orientation, Spring 2008.
Race to the Top (RTTT) and the New York State Regents Reform Agenda Dr. Timothy T. Eagen Assistant Superintendent for Instruction & Curriculum South Huntington.
What Are the Characteristics of an Effective Portfolio? By Jay Barrett.
Identifying Assessments
1 Scoring Provincial Large-Scale Assessments María Elena Oliveri, University of British Columbia Britta Gundersen-Bryden, British Columbia Ministry of.
Rubrics Staff development workshop 19/9/2014 Dr Ruth Fazakerley.
Foundations of American Education: Perspectives on Education in a Changing World, 15e © 2011 Pearson Education, Inc. All rights reserved. Chapter 11 Standards,
Refining Your Assessment Plan: Session 4 Use of the Results Ryan Smith and Derek Herrmann University Assessment Services.
Rubrics: Using Performance Criteria to Evaluate Student Learning PERFORMANCE RATING PERFORMANCE CRITERIABeginning 1 Developing 2 Accomplished 3 Content.
QCC General Education Assessment Task Force March 21 and 22, 2016 Faculty Forum on General Education Outcomes.
If I hear, I forget. If I see, I remember. If I do, I understand. Rubrics.
Copyright © Springer Publishing Company, LLC. All Rights Reserved. DEVELOPING AND USING TESTS – Chapter 11 –
Using Rubrics for Assessing Individual and/or Group Participation Marie Krbavac June 4, 2015.
Student Growth 2.0 NCESD Fellows November 17 th,
EVALUATING EPP-CREATED ASSESSMENTS
Designing Rubrics with the Three Categories of Knowledge
Classroom Assessment A Practical Guide for Educators by Craig A
Consider Your Audience
Phyllis Lynch, PhD Director, Instruction, Assessment and Curriculum
Test Design & Construction
General Education Assessment
Director, Institutional Research
Effective Use of Rubrics to Assess Student Learning
Links for Academic Learning: Planning An Alignment Study
Developing a Rubric for Assessment
Presentation transcript:

How to Norm Rubrics Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College

What is a Rubric? A rubric is a scoring guide such as: A checklist A rating scale A matrix or list containing descriptions of student work characteristics at different levels of sophistication, a.k.a. a descriptive rubric Holistic (containing performance levels but not separating out criteria) Analytic (providing descriptions for each criterion at each performance level)

Consideration of Diverse Points of View Checklist Sanitized hands Verified the patient’s fasting status Asked about latex sensitivity Selected appropriate gloves and tourniquet Assembled necessary supplies Positioned the patient properly Rating Scale   Beginning (1) Developing (2) Proficient (3) Developed key ideas Addressed important details Organized information logically Used proper writing mechanics Descriptive Rubric   Beginning (1) Developing (2) Proficient (3) Consideration of Diverse Points of View Wholly dismisses or disparages points of view that diverge from own worldview Identifies valid components of differing perspectives but responds in accordance with own worldview without reflection Analyzes the complexity and validity of differing perspectives and re-evaluates own perspectives in light of alternative worldviews

Objective vs. Subjective scoring With the exception of checklists, rubrics are used to lend a level of objectivity to evaluation that is inherently subjective Checklists are for use when the demonstration of learning either is or is not present, with no in-between degrees of manifestation Checklists do not require norming Rating scales are the most subjective because they rely on the scorer’s interpretation of the performance-level headings Descriptive rubrics can essentially eliminate subjectivity by clearly identifying indicators of distinct levels of performance Objective vs. Subjective scoring And Selecting the Right Tool

Why Norm a Rubric? Rubric norming is called for when more than one person will be scoring students’ work and the results will be aggregated for assessment of student learning To develop shared understanding of the outcome(s) assessed To discover any need for editing of the rubric To develop scoring consistency among raters Minimize the variable of individual expectations regarding rigor Minimize potential for differences in interpretation of criteria tied to identification of performance levels

Norming RubricS or Norming Raters? Norming rating scales = developing consensus among raters What the different performance levels are intended to capture What level of rigor should be applied in distinguishing the levels All who will be raters should be involved in the norming session(s) Norming descriptive rubrics = perfecting the rubric The better written the rubric, the less possibility it allows for differences in interpretation Identify and fix gray areas and ambiguities in the rubric Reduce need for scorers to conform to a group standard for interpretation

What You’ll Need A rubric A facilitator Some work samples Student learning outcome statements to which the rubric is tied An outline of the steps of the norming process Raters (the faculty who will be doing the scoring)

Selecting Work Samples Real student work or mock-up samples If using real student work, redact any identifying information Select samples that demonstrate different performance levels Plan to have 1 to 3 for each of 2 to 4 scoring sessions Determine number based on the time and complexity of scoring If the rubric contains multiple criteria, select samples that display differing levels of performance on differing criteria (i.e., samples that are neither all good nor all bad)

Outline of the Steps Orientation Discuss levels and rating criteria/thought processes Score the samples Compare scores Discuss (and possibly modify the rubric) Repeat above two steps as needed until consensus is reached

Orienting the Raters to the Process Components The student learning outcome(s) being assessed The purpose of the norming session The rubric itself How it came to be Its intended alignment to the SLO(s) Its intended use An outline of the process

Discussing levels & criteria Model the thought processes involved in using the rubric Entertain perspectives regarding: The number of performance levels and their headings The construct validity of the criteria that have been identified Perceived distinctions between performance levels Perceptions regarding how the faculty think most of the student work will be scored and the discriminative value of the rubric Discussing levels & criteria Part of the Orientation

Scoring the Samples All raters score the same 1 - 3 samples concurrently Raters refrain from discussing the works and/or their scores during the scoring sessions Each time, samples provide a range of skill demonstration Start with the most straight-forward samples and work up to those that require more refined decision-making

Comparing Scores Look for consistencies and inconsistencies Confirm and summarize the rationale behind consistencies Ask raters to articulate the rationale behind inconsistencies Review the scoring criteria Encourage discussion

Reconciling Inconsistent Scores Descriptive rubrics: Can the criteria or descriptions be revised in a way that produces agreement? Strive for natural delineations in performance that reflect discernable steps or benchmarks in the development of proficiency Rating scales: Can the criteria be revised in a way that reduces the opportunity for rater bias Strive for consensus or at least democratic agreement

If the process gets stuck If disagreement persists, change course Consider breaking to calculate percentages of agreement May help raters calibrate their judgments by seeing how their scores compare to those of other raters Calculate the percentage of agreement between pairs of raters and then calculate the mean of the percentages overall (shown on next slide) Consider asking for all to commit to a democratically established convention for scoring in this particular context Those who are the outliers agree to disagree but nonetheless concede to score in accordance with the majority’s bias for the sake of consistency If the process gets stuck Part of Reconciliation

Percentage of Agreement Raters Compared Number of Agreements Total Items Percentage of Agreement 1 & 2 3 7 14% 1 & 3 6 86% 1 & 4 2 & 3 4 57% 2 & 4 3 & 4 100% Overall 30 42 71% Percentage of Agreement How to Calculate It

Final Notes Rubric norming is a means for creating solidarity among the faculty regarding what the shared goals are (as reflected in the student learning outcome statements) and what student performance looks like when the outcomes are partially versus wholly achieved. Rubric norming is most effective when assessment is steeped in an ethos of inquiry, scholarly analysis, and civil academic discourse that encourages faculty participation in decision-making.