What is Open response?.  A Situation Reading Open Response will have a story, a poem, or an article to read.  A Task Set of questions/prompt to answer.

Slides:



Advertisements
Similar presentations
Writing constructed response items
Advertisements

Strategy and Presentation Developed by
Assessment Adapted from text Effective Teaching Methods Research-Based Practices by Gary D. Borich and How to Differentiate Instruction in Mixed Ability.
NJ ASK 3 New Jersey Assessment of Skills and Knowledge Grade 3
When the bartender asked, "How's it going, Norm
Kentucky Department of Education
Gradual Release of Responsibility & Feedback
Designing Instruction Objectives, Indirect Instruction, and Differentiation Adapted from required text: Effective Teaching Methods: Research-Based Practice.
Aligning Depth of Knowledge with the TEKS and the STAAR
The Network of Dynamic Learning Communities C 107 F N Increasing Rigor February 5, 2011.
Domain 3: Instruction. ElementUnsatisfactoryBasicProficientDistinguished Expectations for learningTeacher’s purpose in a lesson or unit is unclear to.
MCAS COACHING “Preserving Our Past… Focusing on Our Future” March 14, 2008 Deb Donahue Amy Polito Kathy Troy.
Beginning With the End in Mind Planning Higher-Level Multiple Choice and Open Response Assessment Items Tom Stewart.
Assessment: Reliability, Validity, and Absence of bias
Building an Accurate SBPR RECORD KEEPING ASSESSMENT AND INSTRUCTION GRADE LEVEL STANDARDS SBPR 1.
Intellectual Challenge of Teaching
Classroom Assessment FOUN 3100 Fall Assessment is an integral part of teaching.
Assessing Student Learning
Methods of Instruction. Learning Objectives Upon completion of this lesson, participants will be able to: – Compare and contrast a range of instructional.
Understanding Validity for Teachers
Classroom Assessment A Practical Guide for Educators by Craig A. Mertler Chapter 9 Subjective Test Items.
Stages of Second Language Acquisition
Teaching with Depth An Understanding of Webb’s Depth of Knowledge
Goals and Objectives.
Access to the GPS – Making the link between student skill, academic content and activity Access to the GPS/GAA Elluminate Session 1 September 11, 2008.
Hallie Hundemer-Booth Halliehbooth.com
Assessing Content Proficiency MAINE CURRICULUM LEADERS’ ASSOCIATION, SEPTEMBER 29, 2014 HEIDI MCGINLEY, EXECUTIVE DIRECTOR,
Bank of Performance Assessment Tasks in English
Evaluation and Testing course: Exam information 6 th semester.
Component 1 – Content Knowledge You may sit in common Certificate Areas.
The Framework for Teaching Charlotte Danielson
Building Effective Assessments. Agenda  Brief overview of Assess2Know content development  Assessment building pre-planning  Cognitive factors  Building.
Task 4 Mathematics Boot Camp Fall, 2015.
ASSESSMENT OF STUDENT LEARNING Manal bait Gharim.
Day 3 (1:00 – 2:00) Room 5 Presented by: Heidi Fulcher.
Aces…A strategy for mastering constructed responses
Classroom Assessments Checklists, Rating Scales, and Rubrics
CONSTRUCTING OBJECTIVE TEST ITEMS: MULTIPLE-CHOICE FORMS CONSTRUCTING OBJECTIVE TEST ITEMS: MULTIPLE-CHOICE FORMS CHAPTER 8 AMY L. BLACKWELL JUNE 19, 2007.
Become familiar with constructed response in the law Understand the meaning of constructed response Review open-response questions Know when to assess.
Setting High Academic Expectations that Ensure Academic Achievement TEAM PLANNING STANDARDS & OBJECTIVES TEACHER CONTENT KNOWLEDGE.
What is the TPA? Teacher candidates must show through a work sample that they have the knowledge, skills, and abilities required of a beginning teacher.
Technology in the classroom. UM weather Great way to begin a day or class –Links to 300 weather sites –Links to 700 web cams to view weather –Radar and.
LAS LINKS DATA ANALYSIS. Objectives 1.Analyze the 4 sub-tests in order to understand which academic skills are being tested. 2.Use sample tests to practice.
Teaching Today: An Introduction to Education 8th edition
Open Response Questions
Power verbs.
ASSESSING STUDENT ACHIEVEMENT Using Multiple Measures Prepared by Dean Gilbert, Science Consultant Los Angeles County Office of Education.
Construct-Centered Design (CCD) What is CCD? Adaptation of aspects of learning-goals-driven design (Krajcik, McNeill, & Reiser, 2007) and evidence- centered.
Jill Kerper Mora San Diego State University. 1.Why do we need to write instructional objectives? What are IOs good for? 2.How do instructional objectives.
Richard Woods, Georgia’s School Superintendent “Educating Georgia’s Future” gadoe.org Assessment for Learning Series Module 4: Working through Complex.
SHOW US YOUR RUBRICS A FACULTY DEVELOPMENT WORKSHOP SERIES Material for this workshop comes from the Schreyer Institute for Innovation in Learning.
Assessment Specifications Gronlund, Chapter 4 Gronlund, Chapter 5.
Depth of Knowledge (DOK)
Assessment and Testing
Bridge Year (Interim Adoption) Instructional Materials Criteria Facilitator:
Alternate Proficiency Assessment Erin Lichtenwalner.
Tuesday 08/12 Grab DOK handouts and put them in your “Units” tab. Warm-up: Look over the Academic Integrity Policy that you researched for homework. Respond.
RESEARCH METHODS IN INDUSTRIAL PSYCHOLOGY & ORGANIZATION Pertemuan Matakuliah: D Sosiologi dan Psikologi Industri Tahun: Sep-2009.
Kentucky Extended-Response Questions General Scoring Guide Score Point 4 Score You complete all important components of the question and communicate.
Essay Questions. Two Main Purposes for essay questions 1. to assess students' understanding of and ability to think with subject matter content. 2. to.
Performance Task and the common core. Analysis sheet Phases of problem sheet Performance task sheet.
Test Question Writing Instructor Development ANSF Nurse Training Program.
The New Face of Assessment in the Common Core Wake County Public Schools Common Core Summer Institute August 6, 2013.
Building Effective Content Literacy Tasks. The Cycle of Assessment Teach: Does the instruction and the tasks align to the identified learning target(s)?
Understanding Depth of Knowledge. Depth of Knowledge (DOK) Adapted from the model used by Norm Webb, University of Wisconsin, to align standards with.
EDU 385 Session 8 Writing Selection items
Preparation for the American Literature Eoc
Do Now – day 1 What is RACES, and what is its purpose?
Strategy and Presentation Developed by
Georgia Department of Education Assessment and Accountability Division
Presentation transcript:

What is Open response?

 A Situation Reading Open Response will have a story, a poem, or an article to read.  A Task Set of questions/prompt to answer using what you have read.  A Scoring Guide Explanation of how task is scored. 4 Characteristics of quality Open Response: Directly tied to one or more content standard Consists of a Situation --Task--Scoring Guide

What is an Open Response task? An open response item is a task that requires students to both demonstrate content knowledge and to apply that knowledge in some way. It is the application component of an open response question that distinguishes it from the more familiar essay. For all tasks, content is the foundation of student answers. Students must demonstrate content knowledge in order to successfully support their answers to communicate clearly what they know and are able to do.

Comparing KCCT Items for State Assessment and Classroom Use Multiple Choice and Open Response Items Addresses Lesson Objectives Results allow for instructional planning and practice Embed into curriculum & instruction Developed within the classroom content STATE Addresses state expectations (Academic Expectations and Core Content for Assessment) Annual measurement Results measure school performance Evaluates long-term learning CLASSROOM

Advantages of Open Response Items They allow for more depth of knowledge to be demonstrated than do multiple choice items. They allow students to demonstrate more complex cognitive behaviors, such as comparing, relating, analyzing, inferring, concluding, predicting, generalizing, solving and/or applying.

Disadvantages of Open Response Items They are much more difficult and time consuming to score. Because of the time required to answer them, there must usually be fewer open response items on an assessment than multiple choice items. Effectiveness of Open Response items is dependant on the scoring guide and answer information provided.

Open Response items may: have a correct answer which students can determine and explain through a variety of methods or in varying degrees of correctness; have multiple successful answers for which students must apply their analytical skills to a response; OR combine requirements: one part requires a student to provide a single correct answer and a subsequent part asks the student to extend his/her knowledge in another way, such as applying the knowledge to another situation or by predicting an outcome.

Tips for designing directions Specify exactly what you want students to answer. Specify numbers of responses if appropriate. Specify three examples if you require three, as opposed to ambiguous terms like “some” or “several.” Use simple and direct language. You are evaluating what students know rather than attempting to determine if they can decipher the question. Use language that is both age and grade-level appropriate. Use simple, basic vocabulary when appropriate, and technical vocabulary when you are assessing the students’ knowledge of the meaning of the technical word/phrase.

1. Scaffolded questions Scaffolded questions have multiple parts, with each question or direction the student is to address presented and labeled separately(A. B. C.). The order is arranged so that successive questions depend upon the response to the previous question.

1. Scaffolded Question Example The framers of the U.S. Constitution wanted to prevent the new federal government from becoming a dictatorship. To keep the government from becoming too powerful, they divided its powers among three branches. A. For each of the three branches of government identify one power given to it by the Constitution. B. Explain why each power you identified in part a is important to our system of government. Support your answer with real-life examples. (Note: answering part b of this question requires that the student be able to list branches of government in part a.)

2. Single Dimension/Component questions Single Dimension/Component forms ask a straight-forward question which requires explanation, examples, description or evidence as support.

2. Single Dimension/Component Example Rivers provide several advantages to cities. Many Kentucky cities are located near large rivers. Describe three important advantages that the rivers provide these cities. Explain why each advantage is important.

3. Two or More Relatively Independent Components questions Two or More Relatively Independent Components are signaled by A, B, C parts. The parts may be about the same prompt but have little relation to each other in that a correct response to one question is not dependent upon the response to the other questions.

3. Two or More Relatively Independent Components Example Fossils provide important clues about things that have lived in the past. A. Describe two ways that fossils can form. B. Explain one way that fossils can help us understand how livings things have changed over time. (Note: answering part b of this question does not require the student be able to successfully answer part a. the reverse is also true.)

4. Student Choice: Topics/Options Provided questions Student Choice with provided topics or options that force students to choose from the selections. They offer students more opportunities to demonstrate their individual learning, but may provide more scoring difficulty because there are many more “correct” answers.

4. Student Choice Example Some of Earth’s materials are listed below: soil water gases of the atmosphere rocks A. Choose TWO materials from the list. Explain how a PLANT uses each of these materials to live. B. Choose TWO materials from the list. Explain how an ANIMAL uses each of these materials to live.

5. Response to Provided Information questions Students must be able to manipulate raw materials such as data, readings or graphics in order to respond to specific questions. This question type is combined with another type of question. In the example below, the student is responding to a text passage, but the question is scaffolded as well.

5. Response to Provided Information Example Note: the student was required to read a text passage before completing this question In the story “First Light,” Matthew woke up in another time period, the 1850s. A. Describe FOUR things Matthew discovered that were different from what he was used to in his present life. B. Explain how EACH of those differences affected him. Use information from the story to support your answer.

Scoring guides are sets of criteria which describe the characteristics of responses at each identified level. They provide the tool necessary to accurately evaluate student success for each individual question. Scoring guides are developed following the actual writing of the question. This will allow the teacher to discover potential problems with the question. Constructing the scoring guide will also help ensure that the question is rich enough to support various levels of student responses. SCORING GUIDES

A scoring guide should: Include a clear explanation of what is expected in a quality student response. Define the various levels of possible student responses and place a value on each level. If there is more than one way a student may achieve a given score level, the scoring guide should include those different possibilities. Enable scoring to be consistent, accurate and as objective as possible. The scoring guide should provide a scorer with the details necessary to score a response. Use simple language and repeat significant descriptive words used in the question. Ensure that what is required for a top-level response is clearly indicated in the description.

Level 4 of the scoring guide may typically characterize responses as effective, thorough, complete, successful, insightful, in depth, efficient and/or sophisticated. Level 3 response will usually use terms such as adequate, satisfactory, understanding of major concepts, completes most and/or clear. Level 2 will often be described as having gaps or leaps, incomplete, some important points, demonstrates basic understanding and/or some errors. Level 1are typically labeled as minimal, completes only small part, little understanding, not logical, unclear and/or major errors.

SCORE 4 You complete all important components of the question and communicate ideas clearly. You demonstrate in-depth understanding of the relevant concepts and/or processes. Where appropriate, you choose more efficient and/or sophisticated processes. Where appropriate, you offer insightful interpretations of extensions (generalizations, applications, analogies). SCORE 3 You complete most important components of the question and communicate clearly. You demonstrate an understanding of major concepts even though you overlook or misunderstand some less important ideas or details.

SCORE 2 You complete some important components of the question and communicate those components clearly. You demonstrate that there are gaps in your conceptual understanding. SCORE 1 You show minimal understanding of the question. You address only a small portion of the question. SCORE 0 Your answer is totally incorrect or irrelevant. BLANK--You did not give any answer at all.

Open Response Hatchet Following Brain's discovering of the survival rifle and matches he feels changed and isn't sure he likes the change. a. Describe how the discovery of the rifle and matches changes him. b. Explain what you believe he means by the idea of not liking the change. Use evidence from the story to support your answer.

4 Specific and detailed description of how the discovery of the rifle and the matches changes Brain. Clear and concise information that supports the changes are given. Specific and detailed explanation of what not liking the changes means. Clear and concise information is given to support the answer. 3 Sufficient description of how the discovery changes Brian. The student supports how the discovery changes Brian with clear information. The student gives an adequate explanation of what Brian means when he says that he does not like the change. The student provides adequate information to support his/her answer.

2 The student vaguely describes how the discovery changed Brian. The student uses limited details to support how the discovery changes Brian. The student provides an unclear explanation of what Brian means when he says that he does not like the change. The student provides limited information to support his/her answer. 1 The student does not provide a clear explanation of how the items changed Brian. The student does not support the answer with details. The student leaves gaps when explaining what Brian means when he says that he does not like the change. The student does not provide information to support the answer.

Much of the information for this presentation was adapted from Developing Quality Open Response and Multiple Choice Items for the Classroom Office of Teaching & Learning (available on-line)