Download presentation
Presentation is loading. Please wait.
Published byOphelia Byrd Modified over 8 years ago
1
Developing Rubrics within the Context of Assessment Methods Peggy Maki Senior Scholar Assessing for Learning AAHE pmaki@aahe.org
2
Rubrics Rubrics establish a basis upon which you ascertain how well a student is achieving or performing or using or integrating “x” Raters use these to judge work Professors use them to give students feedback Students use them to develop work and to understand how their work meets those standards
3
What are the key elements in a rubric? Levels of achievement Criteria that distinguish good work from poor work Descriptions of criteria at each level of achievement
4
Examples of Levels of Achievement Verbal or numerical rankings (such as conceptual understanding is apparent; conceptual understanding only adequate; conceptual understanding not adequate; does not attempt conceptual understanding Mastery levels (levels of attainment) Holistic (one score overall); analytic National standards for scoring Departmental standards for scoring (developed through consensus)
5
Examples of Criteria for Conceptual Attainment in Fundamental Mathematics Conceptual understanding apparent Consistent notation, with only an occasional error Logical formulation Complete or near complete solution/response
6
Ways to Derive Rubrics: Develop through consensus with colleagues Derive from examples of student work (A-F) Derive from knowledge about transition from novice to expert Draw on students’ experiences
7
Format for Scoring Sheet Level Indicators Trait 1 Trait 2 Trait 3 Trait 4 Trait 5
8
Develop a rubric for one outcome:
9
Where You Seek Evidence… Course-embedded In-class Out-of-class Off-campus On-line
10
Some Methods That Provide Direct Evidence Student work samples Collections of student work (e.g., Portfolios) Capstone projects Course-embedded assessment Observations of student behavior Internal juried review of student projects
11
Direct Evidence (continued) External juried review of student projects Externally reviewed internship Performance on a case study/problem Performance on problem and analysis (Student explains how he or she solved a problem)
12
Direct Evidence (continued) Performance on national licensure examinations Response to critical incident Locally developed tests Standardized tests Pre- and post-tests Essay tests blind scored across units Mapping
13
Methods That Do Not Provide Direct Evidence but May Be Combined with Other Methods – Indirect Methods Faculty publications (unless students are involved) Courses selected or elected by students Faculty/student ratios Percentage of students who study abroad Enrollment trends Percentage of students who graduate within five-six years Diversity of student body
14
Focus groups (representative of the population) Interviews (representative of the population) Surveys
15
Other sources of information that contribute to your inference making: NSSE results, grades, participation rates or persistence in support services, course-taking patterns, majors
16
Identify Methods that Will Assess Your Outcome Statement Using Your Rubric:
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.