Assessing Learning for Students with Disabilities Tom Haladyna Arizona State University.

Slides:



Advertisements
Similar presentations
Ed-D 420 Inclusion of Exceptional Learners. CAT time Learner-Centered - Learner-centered techniques focus on strategies and approaches to improve learning.
Advertisements

Action Research Not traditional educational research often research tests theory not practical Teacher research in classrooms and/or schools/districts.
You can use this presentation to: Gain an overall understanding of the purpose of the revised tool Learn about the changes that have been made Find advice.
The Journey Toward Accessible Assessments Karen Barton CTB/McGraw-Hill Validity & Accommodations:
MCR Michael C. Rodriguez Research Methodology Department of Educational Psychology.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
VALIDITY AND RELIABILITY
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 6 Validity.
Validity in Action: State Assessment Validity Evidence for Compliance with NCLB William D. Schafer, Joyce Wang, and Vivian Wang University of Maryland.
Chapter 4A Validity and Test Development. Basic Concepts of Validity Validity must be built into the test from the outset rather than being limited to.
Chapter Fifteen Understanding and Using Standardized Tests.
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Consequential Validity Inclusive Assessment Seminar Elizabeth.
Lesson Eight Standardized Test. Contents Components of a Standardized test Reasons for the Name “Standardized” Reasons for Using a Standardized Test Scaling.
National Center on Educational Outcomes N C E O What the heck does proficiency mean for students with significant cognitive disabilities? Nancy Arnold,
1 Some Key Points for Test Evaluators and Developers Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland October.
Large Scale Assessment Conference June 22, 2004 Sue Rigney U.S. Department of Education Assessments Shall Provide for… Participation of all students Reasonable.
Consistency/Reliability
C R E S S T / U C L A Improving the Validity of Measures by Focusing on Learning Eva L. Baker CRESST National Conference: Research Goes to School Los Angeles,
Research Methods in MIS
Chapter 6 Ethical and Legal Issues in Psychoeducational Assessment
Understanding Validity for Teachers
Chapter 4. Validity: Does the test cover what we are told (or believe)
Presented to: Dr. Ava Clare Marie O. Robles Class Schedule: TFr /1:00-2:30 pm Presented by: Ierine Joy L. Caserial Republic of the Philippines MINDANAO.
Measurement in Exercise and Sport Psychology Research EPHE 348.
High Stakes Testing EDU 330: Educational Psychology Daniel Moos.
Reliability and Validity what is measured and how well.
Psychometric Issues in the Use of Testing Accommodations Chapter 4 David Goh.
Assessing Learning-Centered Leadership Andrew C. Porter University of Pennsylvania Joseph Murphy, Ellen Goldring, & Stephen N. Elliott Vanderbilt University.
Including Quality Assurance Within The Theory of Action Presented to: CCSSO 2012 National Conference on Student Assessment June 27, 2012.
WELNS 670: Wellness Research Design Chapter 5: Planning Your Research Design.
Stephen C. Court Educational Research and Evaluation, LLC A Presentation at the First International Conference on Instructional Sensitivity Achievement.
CCSSO Criteria for High-Quality Assessments Technical Issues and Practical Application of Assessment Quality Criteria.
Reliability & Validity
Validity Is the Test Appropriate, Useful, and Meaningful?
Introduction to Validity
Modifying Achievement Test Items: A Theory-Guided & Data-Based Approach Stephen N. Elliott Learning Sciences Institute and Department of Special Education.
Enhancing the Technical Quality of the North Carolina Testing Program: An Overview of Current Research Studies Nadine McBride, NCDPI Melinda Taylor, NCDPI.
The Do’s and Don’ts of High-Stakes Student Achievement Testing Andrew Porter Vanderbilt University August 2006.
CAROLE GALLAGHER, PHD. CCSSO NATIONAL CONFERENCE ON STUDENT ASSESSMENT JUNE 26, 2015 Reporting Assessment Results in Times of Change:
Evaluating Consequential Validity of AA-AAS Presented at OSEP Conference January 15, 2008 by Marianne Perie Center for Assessment.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Do Your Assessment Instruments Measure Up? Gerald Tindal UO.
Catholic College at Mandeville Assessment and Evaluation in Inclusive Settings Sessions 3 & /14/2015 Launcelot I. Brown Lisa Philip.
ARG symposium discussion Dylan Wiliam Annual conference of the British Educational Research Association; London, UK:
Slide 1, Moving From Challenge To Action: Accountability Supporting Student Learning Joan L. Herman UCLA Graduate School of Education & Information.
Introduction to Item Analysis Objectives: To begin to understand how to identify items that should be improved or eliminated.
C R E S S T / U C L A Validity Issues for Accountability Systems Eva L. Baker AERA April 2002 UCLA Graduate School of Education & Information Studies.
What does exam validity really mean? Andrew Martin Purdue Pesticide Programs.
Chapter 6 - Standardized Measurement and Assessment
C R E S S T / CU University of Colorado at Boulder National Center for Research on Evaluation, Standards, and Student Testing Design Principles for Assessment.
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
Critical Issues Related to ELL Accommodations Designed for Content Area Assessments The University of Central Florida Cocoa Campus Jamal Abedi University.
Copyright © 2009 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 47 Critiquing Assessments.
The evidence is in the specifications
Take-Home Message: Principles Unique to Alternate Assessments
Introduction to the Validation Phase
Test Blueprints for Adaptive Assessments
Concept of Test Validity
Introduction to the Validation Phase
Validity and Reliability
Week 3 Class Discussion.
پرسشنامه کارگاه.
Federal Policy & Statewide Assessments for Students with Disabilities
Validating Interim Assessments
William D. Schafer, Joyce Wang, and Vivian Wang University of Maryland
Investigations into Comparability for the PARCC Assessments
Presentation transcript:

Assessing Learning for Students with Disabilities Tom Haladyna Arizona State University

Useful Sources Standards for Educational & Psychological Testing, AERA, APA, NCME (1999) Standards for Educational & Psychological Testing, AERA, APA, NCME (1999) Tindal & Haladyna (2002) Tindal & Haladyna (2002) Large-scale assessment programs for all students: Validity, technical adequacy, and implementation. Downing & Haladyna (2006) Handbook of test development Downing & Haladyna (2006) Handbook of test development Haladyna & Downing (2004). Haladyna & Downing (2004). Construct-irrelevant variance in high-stakes testing. Educational Measurement: Issues and Practice. Kane (2006) Content-related validity evidence. Handbook of test development. Kane (2006) Content-related validity evidence. Handbook of test development. Kane (In press). Validation. Educational Measurement (4 th ed.) Kane (In press). Validation. Educational Measurement (4 th ed.)

Assessment vs. Testing Assessment is the act of judging the indicators of student achievement for the benefit of planning future instruction. Assessment is the act of judging the indicators of student achievement for the benefit of planning future instruction. Testing is a way of providing one valid source of information for assessment Testing is a way of providing one valid source of information for assessment A test is NEVER a valid source of information for assessment unless corroborated by other evidence.—Multiple indicators A test is NEVER a valid source of information for assessment unless corroborated by other evidence.—Multiple indicators

Validity of a Test Score Interpretation or Use Way of reasoning about test scores. Way of reasoning about test scores. Concerned about the accuracy of any interpretation or use. Concerned about the accuracy of any interpretation or use. Involves an argument about how an assessment or a test score can be validly interpreted or used. Involves an argument about how an assessment or a test score can be validly interpreted or used. Involves a claim by the developer/user Involves a claim by the developer/user Involves evidence that might support this claim Involves evidence that might support this claim

Validation’s Steps Developmental Phase Developmental Phase State a purpose for the test. State a purpose for the test. Define the trait (construct). Define the trait (construct). Content Content Cognitive demand Cognitive demand Develop the test. Develop the test. Validate—conduct the study. Validate—conduct the study. Investigative phase Investigative phase

Two Types of Evidence That supports our claim That supports our claim That weakens or threatens validity That weakens or threatens validity Construct under representation Construct under representation Construct-irrelevant variance Construct-irrelevant variance

Two Types of Evidence Includes procedures known to strengthen our argument and support our claim. Includes procedures known to strengthen our argument and support our claim. Includes statistical/empirical information that also strengthens our argument and supports our claim Includes statistical/empirical information that also strengthens our argument and supports our claim

More Types of Evidence Content-related Content-related Reliability Reliability Item quality Item quality Test design Test design Test administration Test administration Test scoring Test scoring Test reporting Test reporting Consequences Consequences

Content Structure—sub scores??? Structure—sub scores??? Concurrent—how it correlates with other information Concurrent—how it correlates with other information Does it represent the construct (content)? Does it represent the construct (content)?

Reliability Very important type of validity evidence. Very important type of validity evidence. Can be applied to individual or group scores. Can be applied to individual or group scores. Group scores tend to be very reliable. Group scores tend to be very reliable. Can focus at reliability at a decision point. Can focus at reliability at a decision point. Subjective judgment a factor in reliability. Subjective judgment a factor in reliability.

Random Error Basis for reliability Basis for reliability Can be large or small Can be large or small Can be positive or negative Can be positive or negative We never know. We never know. We just guess. We just guess. Guessing allows us to speculate about where a student’s true score lies and what action we take. Guessing allows us to speculate about where a student’s true score lies and what action we take.

Item Quality Universal item design Universal item design Format issues Format issues Item reviews Item reviews Field tests Field tests

Test Design Breadth Breadth Scope Scope Depth Depth Length Length Formats Formats

Test Administration Standardized Standardized Accommodations Accommodations Standards Standards

Test Scoring Avoid errors. Avoid errors. Quality control is important. Quality control is important. Invalidate scores when evidence suggests that. Invalidate scores when evidence suggests that.

Score Reporting Helpful to teachers for assessment Helpful to teachers for assessment Meets requirements for accountability Meets requirements for accountability Meet Standards (Ryan, 2006) Meet Standards (Ryan, 2006)

Advice Document what you do. Technical Report Document what you do. Technical Report Build the case for validity. Build the case for validity. Do validity studies when possible. Do validity studies when possible. Stay focused on the real reason for assessment and testing: helping students learn not satisfying someone in DC. Stay focused on the real reason for assessment and testing: helping students learn not satisfying someone in DC.