Writing Modified Achievement Level Descriptors Presented at OSEP Conference January 16, 2008 by Marianne Perie Center for Assessment.

Slides:



Advertisements
Similar presentations
Depths of Knowledge and Reading
Advertisements

Bringing it all together!
CLOSE READING WITH MNEMONICS, MEDIA AND MEANING WHY CLOSE READING? Thoughtful, Critical Analysis of Text Focus on Patterns Develops Deep, Precise Understanding.
NCLB Basics From “What Parents of Students with Disabilities Need to Know & Do” National Center on Educational Outcomes University of Minnesota
Standards-Based IEPs Aligning the IEP and Academic Content Standards to Improve Academic Achievement.
The Network of Dynamic Learning Communities C 107 F N Increasing Rigor February 5, 2011.
1 © 2006 Curriculum K-12 Directorate, NSW Department of Education and Training Implementing English K-6 Using the syllabus for consistency of teacher judgement.
American History Foundations
What Parents Need to Know  TABS (Texas Assessment of Basic Skills)  TEAMS (Texas Educational Assessment of Minimum Skills)  TAAS (Texas Assessment.
1 New England Common Assessment Program (NECAP) Setting Performance Standards.
WORKING TOGETHER ACROSS THE CURRICULUM CCSS ELA and Literacy In Content Areas.
ACT Close and Critical Reading Using ACT Content Passages Macomb Intermediate School District September 15 th, 2010.
Consistency of Assessment
Intellectual Challenge of Teaching
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Standard Setting Inclusive Assessment Seminar Marianne.
Principles of High Quality Assessment
Overview of the Dynamic Learning Maps Alternate Assessment System
 Here’s What... › The State Board of Education has adopted the Common Core State Standards (July 2010)  So what... › Implications and Impact in NH ›
Washington Educational Research Association Conference Washington Educational Research Association Conference Beth Dorr Reading Assessment Specialist
Science Inquiry Minds-on Hands-on.
Assessment Cadre #3: “Assess How? Designing Assessments to Do What You Want”
Moving to the Common Core Janet Rummel Assessment Specialist Indiana Department of Education.
Dr. MaLinda Hill Advanced English C1-A Designing Essays, Research Papers, Business Reports and Reflective Statements.
DeAnn Huinker & Kevin McLeod University of Wisconsin-Milwaukee Designing High Quality Professional Development Knowledge, Management, & Dissemination Conference.
PSSA Reading Test.
NCSC Project Description
Understanding Depth of Knowledge
1 Summer 2012 Educator Effectiveness Academies English Language Arts Transitioning to the CCSS by Making Strategic and Informed Choices in the Classroom.
Wisconsin Extended Grade Band Standards
Building Effective Assessments. Agenda  Brief overview of Assess2Know content development  Assessment building pre-planning  Cognitive factors  Building.
Framework for Diagnostic Teaching. Framework The framework for diagnostic teaching places a premium on tailoring programs that specifically fit all readers.
The Framework for Teaching Domain 1 Planning and Preparation.
Out with the Old, In with the New: NYS Assessments “Primer” Basics to Keep in Mind & Strategies to Enhance Student Achievement Maria Fallacaro, MORIC
Webb’s Depth of Knowledge (DOK) Aligning Assessment Questions to DOK Levels Assessing Higher-Order Thinking.
© 2012 Board of Regents of the University of Wisconsin System, on behalf of the WIDA Consortium Introduction to the WIDA English Language Development.
T 7.0 Chapter 7: Questioning for Inquiry Chapter 7: Questioning for Inquiry Central concepts:  Questioning stimulates and guides inquiry  Teachers use.
The New English Curriculum September The new programme of study for English is knowledge-based; this means its focus is on knowing facts. It is.
Why We Got Involved… Bellwood-Antis Middle School PSSA Data 2003.
ENGLISH LANGUAGE ARTS MODULE 4 FEBRUARY 2013 Reading Common Core Focus: Text Complexity.
Developing Assessments for and of Deeper Learning [Day 2b-afternoon session] Santa Clara County Office of Education June 25, 2014 Karin K. Hess, Ed.D.
QAR: Question Answer Relationships
What is it? Why do it? How do you do it?
COUNCIL OF CHIEF STATE SCHOOL OFFICERS (CCSSO) & NATIONAL GOVERNORS ASSOCIATION CENTER FOR BEST PRACTICES (NGA CENTER) JUNE 2010.
Constructed Response Developing this writing practice as part of ongoing classroom assessment The value of constructed response is that it is teaching.
MRS. LIMA AP Literature & Composition. What are AP Courses? Provide the opportunity for academically prepared and motivated students to complete.
Graphic Organizers 1. The human continuum Scale -questions about Graphic Organizers! 2. What are Graphic Organizers? 3. Different Types of Graphic Organizers.
Lesson Plans the UWG way Block One. Components of a Lesson Lesson Title : Main Idea or focus of the lesson Content Standard : What standard(s) are most.
Implementing the Common Core State Standards Monday, January 23rd - 4pm EST Deconstructing the Common Core Standards: Analyzing for Content, Level of Cognition.
Steps to Creating Standards-Based Individualized Education Programs The following highlights the major steps Committees on Special Education (CSEs) can.
Assessment Literacy and the Common Core Shifts. Why Be Assessment Literate? ED Test and Measurement 101 was a long time ago Assure students are being.
The Standards-based IEP Process: What You Need to Know Standards-Based IEP State-Directed Project - January 2011.
INTRODUCTION TO THE WIDA FRAMEWORK Presenter Affiliation Date.
Tia Juana Malone, English Professor Ruth Ronan, Course Developer Assessment Strategies That Promote Student Engagement.
 Participants will leave knowing how to determine if Kurzweil is appropriate for your students.  Participants will begin the discussion of how to.
In the Age of Common Core. Close Reading of text involves an investigation of a short piece of text, with multiple readings done over multiple instructional.
Learning Goals & Targets
Understanding Your Child’s Report Card
Reciprocal Teaching At Work
M-Step Overview, Practice and Instructional Implications
Advanced English 6 November 1-2, 2017
Parent Involvement Committee EQAO Presentation
Assessment Information
Instructional Learning Cycle:
Writing Business Reports and Proposals
Assessment Literacy: Test Purpose and Use
Goals for Tonight Present the new Georgia Milestones Assessment System
Goals for Tonight Present the new Georgia Milestones Assessment System
Advanced English 6 November 10, 14
Standards-based Individualized Education Program (IEP) Module One: Introduction SBIEP Module one: Introduction - The standards-based reform movement has.
Norman L Webb.
Presentation transcript:

Writing Modified Achievement Level Descriptors Presented at OSEP Conference January 16, 2008 by Marianne Perie Center for Assessment

2 Purpose of this Presentation This won’t give you the answers of what a “good” modified achievement level descriptor (MALD) looks like However, it should provide some thoughts and a process for your consideration States may choose different paths, resulting in different MALDs—what are the implications? Choices should be deliberate as they will have a strong influence not only on interpreting test results, but on instructional choices as well

Center for Assessment 3 Interrelation of All Activities Who Are the Students? What Are the Characteristics of Items They Do Well/Poorly On? Appropriate Modifications to Items Development of Modified ALDs Instruction Figuring out the answer to this question… And examining extant data to determine this… Informs all of these activities

Center for Assessment 4 Importance of Writing Thoughtful MALDs MALDs guide the interpretation of the test scores of the AA-MAS MALDs influence teacher expectations (e.g., what students CAN do versus what they CAN’T do) MALDs guide instructional activities

Center for Assessment 5 Considerations in Writing MALDs Determine how the MALDs should reflect the content standards Decide on the relationship between the MALDs and the test blueprint Weigh the importance of skill development with content Think holistically about how a student moves from one performance level to the next, one grade to the next, and one assessment group (1% → 2% → general assessment) to the next

Center for Assessment 6 Reflecting the Content Standards and Test Blueprints Consider how the emphasis in the content standards is reflected in the test blueprint.  Strong emphasis on numbers and operations may result in a blueprint with 50% of the items assessing this content strand Consider how the emphasis in the content standards and test blueprints should be reflected in the MALDs  Do the learning descriptions in your MALDs mirror your assessment emphasis?  How should the weights in the blueprint be translated into descriptors?

Center for Assessment 7 Interaction between Content and Process In reading, “content” could be the level and type of reading passage Process would be how the student works with the information, such as recalling specific text or locating direct references to interpreting, analyzing, comparing, or making inferences and drawing conclusions. How do we expect greater knowledge and skill to manifest itself?  Does a student demonstrate greater understanding by staying at the same processing level, but increase the difficulty/complexity of the text?  Does a student demonstrate greater understanding by continuing to work with the same text, but increase the level of skill applied to that text?  Both? How do they interact?

Center for Assessment 8 Movement Across Grades and Levels Think about movement along content-specific learning progressions within and across grades How does a student who is Proficient at grade 3 continue to show progress?  Move to Advanced at grade 3, to Basic at grade 4, to Proficient at grade 4?  Over how long a period? Ask that same question for students who are below Basic at Grade 3  If a student moved from below Basic at grade 3 to Advanced at grade 3 in one academic year, we would call that progress, but he would then be operating off-grade level  So, how does Advanced at grade 3 relate to performance at grade 4?

Center for Assessment 9 Movement across Assessments Consider how the AA-MAS is supposed to fit between the AA-AAS and the general assessment  Do we expect to see smooth transitions from one to the next?  How do the content expectations relate?  Is Advanced on the AA-MAS similar in nature to Basic on the general assessment? Our expectation for the AA-MAS is that it may provide a stepping stone for students to reach proficient on the general assessment That expectation needs to be reflected in the AA- MAS  Consider the grade-level PLDs when writing MALDs

Center for Assessment 10 Scaffolding / Content Supports Think about the use of scaffolding. How true is it that a proficient student on the modified assessment may have a similar set of knowledge and skills as the proficient student on the general assessment, but may require more supports to demonstrate that knowledge.

Center for Assessment 11 Example of Scaffolding One student may be able to answer a broader/more generalized question, such as about the author’s purpose or theme, immediately after reading a longer text. Another student may need to first consider questions more directly connected to aspects of the same text (e.g., about the main events, the conflict, the resolution) before being able to make an interpretive statement about the author’s purpose or theme. Ultimately, the answers may be equally correct, but one student is able to produce a purpose statement without any supporting direction, while the other student needs to be directed toward the answer through a stepwise progression. Note that in neither situation is the instructor (or prior test items) providing the correct answer, only a way to think about determining the answer.

Center for Assessment 12 Implications for MALDs If scaffolding is important to a student’s success, then it needs to be incorporated into the descriptors The knowledge and skills required to reach proficiency might not be very different between the general assessment and the modified assessment, but the supports may vary Consider scaffolding in addition to the learning progressions always keeping in mind transitions across levels, grades, and assessments

Center for Assessment 13 Now Let’s Make this Practical Supplement theory with data Convene committees to draft MALDs

Center for Assessment 14 Data Analysis Supplement the conceptual understanding of learning progressions with data on what these students currently know and can do Once the population has been defined we can examine data on student performance that should already exist Data on how the students are currently performing may help those drafting the MALDs balance what the students currently know and can do with what they should know and can do Of course the data analysis will also help inform the modifications to the assessment, which is why these three pieces must be considered together

Center for Assessment 15 Gather Data to Inform Development of MALDs Identify the population to be assessed under the 2% assessment (e.g., students with IEPs who are consistent low performers on the general assessment) Then, examine their general assessment results  What are the characteristics of items that this population does well on?  What are the characteristics of items that this population struggles with?

Center for Assessment 16 Item Characteristics Have content experts summarize findings considering both the content standards measured and specific item features such as  Conceptual understanding, fact-based content, or processes/skills  Level of reasoning required (locate vs. infer)  “Closeness” of distracters to each other  Vocabulary load within items (not the vocabulary term being tested)  Difficulty/abstractness of vocabulary (e.g., use of figurative language) or ideas presented  Concepts tested in items (e.g., fact versus opinion, author’s purpose)  Reading: genres, text structures, and length of the passages  Mathematics: multi-step problems, and/or supports with graphics

Center for Assessment 17 Committee Meeting 1. Bring together a committee of content experts (e.g., classroom teachers and curriculum leaders) and special education teachers.  Content experts should make up about 2/3rd of the committee.  Need around 5–8 participants per subject area, but if you’re developing MALDs for multiple grade levels, consider inviting more participants and splitting them into teams 2. Start with background information on this population and discuss what you have learned about this population giving specific examples  Disaggregated assessment data  Teacher perceptions

Center for Assessment 18 Committee Meeting (continued) 3. Discuss interactions of process and content and what it takes to move across both performance levels and grade levels  Are the knowledge and skills required of Proficient on the MAS the same as on the GLAS but more supports are scaffolding is needed, or are the knowledge and skills different?  If they are different, is the content different or the processes? e.g., both can make inferences at the Proficient level but the GLAS requires that the inferences are made in a more complex context than the MAS, or GLAS can make inferences, while MAS can only draw basic conclusions from concepts presented directly How do they map to the grade-level content standards?

Center for Assessment 19 Committee Meeting (continued) 4. Discuss the transition from this assessment to the general assessment – how are they linked? 5. Focus first on the proficient level and brainstorm what a student should know regarding each content strand (or substrand/benchmark/indicator) in order to be proficient  Keep a list of the ideas in bullet format 6. Move to Basic and write statements for that level  Should be parallel to a degree, although all skills and content might not be included at all performance levels  Compared to Proficient, does Basic imply a different breadth/depth of content, a different level of processes, or the ability to apply knowledge to different contexts 7. Move to Advanced (and any other levels) and repeat

Center for Assessment 20 Committee Meeting (continued) 8. Now consider adjacent grade(s)  How should Advanced in the prior grade relate to Below Basic/ Basic/ Proficient in the subsequent grade?  How do you envision students moving across grades?  How does Proficient in one grade compare to Proficient in the next? 9. Write “final” MALDs  Format could be the bulleted list, or you could rewrite that into a descriptive paragraph. 10. End the meeting with a summary of all MALDs across all levels and applicable grades.  Can you see a clear progression?  Will this be translatable to instruction? (Just as in the general assessment, states will then need to have MALDs adopted formally)

Center for Assessment 21 Concluding Thoughts A lot of work needs to go into developing a theoretical model of learning progressions  Should be based on research in this field  Model can then drive development of both the assessment and the descriptors Data on current achievement can be examined once the population has been identified to help inform the MALDs Possibility that states could go different ways – what are the implications?  One state chooses to require higher inferences on lower-level texts while another chooses to require lower-level inferences on more complex text This work has the potential to influence the instruction of all students  Consider low achievers who are not special education students