Michigan Assessment Consortium Common Assessment Development Series Putting Together The Test Blueprint.

Slides:



Advertisements
Similar presentations
Matching the Assessment Method to the Learning Target
Advertisements

Performance Assessment
Instructional Decision Making
Evaluating the Alignment and Quality of the Assessed Curriculum Virginia Department of Education Office of School Improvement
Introduction to Creating a Balanced Assessment System Presented by: Illinois State Board of Education.
TEST DEVELOPMENT: CONNECTING CURRICULUM MAPS/PACING GUIDES, ITEM SPECIFICATIONS, AND TEST BLUEPRINTS DEVELOPED BY: HEATHER P. WRIGHTCHERLOTTLA ARGRETT.
Summative Assessment Kansas State Department of Education ASSESSMENT LITERACY PROJECT1.
Learning Outcomes Participants will be able to analyze assessments
The ABCs of Assessment Improving Student Learning Through New Approaches to Classroom Assessment.
Understanding Depth 0f knowledge
The Network of Dynamic Learning Communities C 107 F N Increasing Rigor February 5, 2011.
Assessment Literacy Series
Benchmark Assessment Item Bank Test Chairpersons Orientation Meeting October 8, 2007 Miami-Dade County Public Schools Best Practices When Constructing.
Writing High Quality Assessment Items Using a Variety of Formats Scott Strother & Duane Benson 11/14/14.
DEEPENING ASSESSMENT LITERACY Fall Objective  Identify best practices for local assessment development  Provide a working knowledge of the WPSD.
Michigan Assessment Consortium Common Assessment Development Series Module 6 – The Test Blueprint.
The mere imparting of information is not education. Above all things, the effort must result in helping a person think and do for himself/herself. Carter.
Building an Accurate SBPR RECORD KEEPING ASSESSMENT AND INSTRUCTION GRADE LEVEL STANDARDS SBPR 1.
Webb’s Depth of Knowledge USD 457. Objectives CO – Analyze and apply the four levels of Webb’s Depth of Knowledge. LO – Read for the purpose of understanding.
MCAS-Alt: Alternate Assessment in Massachusetts Technical Challenges and Approaches to Validity Daniel J. Wiener, Administrator of Inclusive Assessment.
Principles of High Quality Assessment
Atlantic Canada Mathematics Assessment Resource Entry - 3 Introduction to the Resource CAMET (Council of Atlantic Ministers of Education & Training )
Assessment Literacy Series
SEPT 20 8:00-11:00 WHAT ARE WE MEASURING? HOW DO WE MEASURE? DHS English Department Professional Development.
Introduction to Depth of Knowledge
Evaluation: A Challenging Component of Teaching Darshana Shah, PhD. PIES
2012 Secondary Curriculum Teacher In-Service
NEXT GENERATION BALANCED ASSESSMENT SYSTEMS ALIGNED TO THE CCSS Stanley Rabinowitz, Ph.D. WestEd CORE Summer Design Institute June 19,
Building Effective Assessments. Agenda  Brief overview of Assess2Know content development  Assessment building pre-planning  Cognitive factors  Building.
Assessment Literacy Module
DLM Early Childhood Express Assessment in Early Childhood Dr. Rafael Lara-Alecio Dr. Beverly J. Irby
Clear Purpose: Assessment for and of Learning: A Balanced Assessment System “If we can do something with assessment information beyond using it to figure.
Classroom Assessment A Practical Guide for Educators by Craig A
Quality Instruction The ♥ of the matter is the depth of knowledge.
Teaching Today: An Introduction to Education 8th edition
Michigan Assessment Consortium Common Assessment Development Series Common Assessment Development Series Rubrics and Scoring Guides.
MAC Common Assessment Training Modules Session F3 Michigan School Testing Conference February 23, 2012.
Developing Assessments for and of Deeper Learning [Day 2b-afternoon session] Santa Clara County Office of Education June 25, 2014 Karin K. Hess, Ed.D.
Data for Student Success August, 2009 Mission Pointe “It is about focusing on building a culture of quality data through professional development and web.
Stages 1 and 2 Wednesday, August 4th, Stage 1: Step 5 National and State Standards.
Building Assessment Literacy in Michigan through Quality Common Assessment Development.
After lunch - Mix it up! Arrange your tables so that everyone else seated at your table represents another district. 1.
Fourth session of the NEPBE II in cycle Dirección de Educación Secundaria February 25th, 2013 Assessment Instruments.
Michigan Assessment Consortium Building and Using Common Assessments: A Professional Development Series Module 1 Overview of the Series.
Goals for Webinar: Applications of SEC Alignment Analysis
Learning Targets January 21, 2008 Londa Richter & Jo Hartmann TIE.
An Analysis of Three States Alignment Between Language Arts and Math Standards and Alternate Assessments Claudia Flowers Diane Browder* Lynn Ahlgrim-Delzell.
Summer 2012 DataCamp June/August, Everyone is using standards…
A Balanced Assessment Plan For Petoskey Public Schools.
Assessment Power! Pamela Cantrell, Ph.D. Director, Raggio Research Center for STEM Education College of Education University of Nevada, Reno.
Determining Student Mastery: Achieving learning potential using assessment Drew Maerz Asheboro City Schools July 8, 2014.
Standards-Based Instruction Implementation: Badge 4-Analyzing Standards.
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
Michigan Assessment Consortium Common Assessment Training Modules Session B5 Michigan School Testing Conference February 23, 2011.
SBAC-Mathematics November 26, Outcomes Further understand DOK in the area of Mathematics Understand how the new SBAC assessments will measure student.
Statewide System of Support For High Priority Schools Office of School Improvement.
GOING DEEPER INTO STEP 1: UNWRAPPING STANDARDS Welcome!
Instructional Leadership Supporting Common Assessments.
Designing Quality Assessment and Rubrics
Dr. Marciano B. Melchor University of Ha’il, KINGDOM OF SAUDI ARABIA May 2013.
Educational Assessments – Sound Design
Assessment and Evaluation
How Assessment Supports RTI2 and Closing the Achievement Gap
Curriculum Structures
Assessing the Common Core Standards
Critically Evaluating an Assessment Task
Michigan Assessment Consortium Common Assessment Development Series Module 21 Assessment Administration and Scoring MAC CAD-PD Mod-6 BRF
Assessment Practices in a Balanced Assessment System
Developing Quality Assessments
Presentation transcript:

Michigan Assessment Consortium Common Assessment Development Series Putting Together The Test Blueprint

Developed and Narrated by Bruce R. Fay, PhD Assessment Consultant Wayne RESA

Support The Michigan Assessment Consortium professional development series in common assessment development is funded in part by the Michigan Association of Intermediate School Administrators in cooperation with MDE, MSU, Ingham & Ionia ISDs, Oakland Schools, and Wayne RESA. The Michigan Assessment Consortium professional development series in common assessment development is funded in part by the Michigan Association of Intermediate School Administrators in cooperation with MDE, MSU, Ingham & Ionia ISDs, Oakland Schools, and Wayne RESA.

What You Will Learn Test blueprints…what they are and why you need them Test blueprints…what they are and why you need them The components of a test blueprint The components of a test blueprint Criteria for a good test blueprint Criteria for a good test blueprint Test blueprint examples Test blueprint examples

If you don't know where you're going, any road will take you there. George Harrison ( ) "Any Road", Brainwashed, 2002

Assessment with a Purpose Educational assessment is not something incidental to teaching and learning. It is an equal partner with curriculum and instruction. It is the critical 3 rd leg through which both students and teachers receive feedback about the effectiveness of the teaching and learning process in achieving desired learning outcomes. Assessment closes the loop. Educational assessment is not something incidental to teaching and learning. It is an equal partner with curriculum and instruction. It is the critical 3 rd leg through which both students and teachers receive feedback about the effectiveness of the teaching and learning process in achieving desired learning outcomes. Assessment closes the loop.

Closed–Loop (Feedback) Systems Home Heating System (Teaching & Learning) Desired Temperature (Learning Target) Actual Temperature (Test Results)

C – I – A Alignment Requires thoughtful alignment – ensuring that the items on a test fairly represent the… Requires thoughtful alignment – ensuring that the items on a test fairly represent the… Intended learning targets (intended curriculum) Intended learning targets (intended curriculum) Actual learning targets (taught curriculum) Actual learning targets (taught curriculum) Test what you teach, teach what you test

Target–Level Alignment Relative importance of those targets Relative importance of those targets Level of cognitive complexity associated with those targets Level of cognitive complexity associated with those targets

Useful feedback requires tests that are… Reliable (consistent; actually measure something) Reliable (consistent; actually measure something) Fair (Free from bias or distortions) Fair (Free from bias or distortions) Valid (contextually meaningful or interpretable; can reasonably support the decisions we make based on them) Valid (contextually meaningful or interpretable; can reasonably support the decisions we make based on them)

Test Blueprints The Big Idea A simple but essential tool, used to: A simple but essential tool, used to: Design tests that can meet the preceding requirements Design tests that can meet the preceding requirements Define the acceptable evidence to infer mastery of the targets Define the acceptable evidence to infer mastery of the targets Build in evidence for validity Build in evidence for validity

The Test Blueprint or Table of Test Specifications Explicitly map test items to: Explicitly map test items to: Learning Targets Learning Targets Levels of Complexity Levels of Complexity Relative Importance Relative Importance Provides common definition of the test Provides common definition of the test

Learning Targets & Standards Frameworks Standards as structured hierarchical frameworks. Michigans is: Standards as structured hierarchical frameworks. Michigans is: Strands Strands Standards Standards Domains Domains Content Expectations Content Expectations Detailed curriculum is usually left to local districts or classroom teachers. Detailed curriculum is usually left to local districts or classroom teachers.

A Simple Taxonomy of Cognitive Complexity Norm Webbs Depth Of Knowledge (1997) (highest to lowest) Extended Thinking Extended Thinking Strategic Thinking Strategic Thinking Skill / concept use / application Skill / concept use / application Recall Recall

Putting it all together… A Basic Test Blueprint Table (matrix) format (spreadsheet) Table (matrix) format (spreadsheet) Rows = learning targets (one for each) Rows = learning targets (one for each) Columns = Depth of Knowledge Columns = Depth of Knowledge Cells = number of items and points possible Cells = number of items and points possible

Summary Information Number of items and points possible: Number of items and points possible: Row Margins = for that target Row Margins = for that target Column Margins = for that level of complexity Column Margins = for that level of complexity Lower Right Corner = for the test Lower Right Corner = for the test

Example 1 – Basic Blueprint for a test with 5 learning targets Learning Targets Recall # (pts) Use # (pts) Strategic # (pts) Extended #(pts) Target Totals # (pts) Target 13 (3)2 (2)5 (5) Target 21 (1)2 (2)2 (4)5 (7) Target 32 (2)1 (3)3 (5) Target 43 (3)1 (2)4 (5) Target 52 (4)1 (4)3 (8) Level Totals # (pts) 6 (6)10 (14)4 (10)20 (30)

Is this reasonable? Rule of Thumb Criteria… At least 3 items per target for reliability At least 3 items per target for reliability Appropriate: Appropriate: Distribution of items over targets Distribution of items over targets Levels of complexity for targets/instruction Levels of complexity for targets/instruction Distribution of items over levels of complexity Distribution of items over levels of complexity (all items are NOT at the lowest or highest level)

Professional Judgment Like all things in education, the development of assessments and the use of the results are dependent on professional judgment, which can be improved through… Like all things in education, the development of assessments and the use of the results are dependent on professional judgment, which can be improved through… Experience Experience Collaboration Collaboration Reflection on methods Results Reflection on methods Results

Limitations… Shows total points for each target/level combination, but not how those points apply to each item Shows total points for each target/level combination, but not how those points apply to each item Doesnt show item types Doesnt show item types Doesnt indicate if partial credit scoring can/will be used (but may be implied) Doesnt indicate if partial credit scoring can/will be used (but may be implied) But…it was easy to construct, is still a useful blueprint, and is much better than not making one! But…it was easy to construct, is still a useful blueprint, and is much better than not making one!

Add details on item type and format to ensure… Appropriate match to learning targets and associated levels of complexity Appropriate match to learning targets and associated levels of complexity Balanced use within tests and across tests over time Balanced use within tests and across tests over time Specification of test resources, i.e. – Specification of test resources, i.e. – Calculators, dictionaries, measuring tools… Calculators, dictionaries, measuring tools… Track on same or separate spreadsheet Track on same or separate spreadsheet

Common item types include… Selected-response Selected-response Multiple-choice Multiple-choice Matching Matching Constructed-response Constructed-response Brief (fill-in-the-blank, short answer, sort a list) Brief (fill-in-the-blank, short answer, sort a list) Extended (outline, essay, etc.) Extended (outline, essay, etc.) Performance Performance Project Project Portfolio Portfolio

Complexity vs. Utility Your test blueprint could get complicated if you try to account for too much in one spreadsheet. Your test blueprint could get complicated if you try to account for too much in one spreadsheet. Make sure your test blueprint covers the basics, is not a burden to create, and is useful to you Make sure your test blueprint covers the basics, is not a burden to create, and is useful to you The following example is slightly more complicateded, but still workable The following example is slightly more complicateded, but still workable

Target Code Item # Item type SR Item type CR-B Item type CR-E DOK Recall Pts DOK Use Pts DOK Strategic Pts Trgt Tots (pts) x1 2x1 3x x3 5x58 etc Col Tots Example 2 – Blueprint with Explicit Items and Item Types

Beyond the Test Blueprint Answer key (selected-response items) Answer key (selected-response items) Links to scoring guides & rubrics Links to scoring guides & rubrics Specs for external test resources Specs for external test resources Item numbering for alternate test forms Item numbering for alternate test forms

Conclusions Destination & Road Map Destination & Road Map Alignment/balance of items/types for… Alignment/balance of items/types for… learning targets (curriculum /content) learning targets (curriculum /content) size (complexity) of targets size (complexity) of targets cognitive level of targets cognitive level of targets relative importance of targets relative importance of targets Spec or document other aspects of the test Spec or document other aspects of the test