Evaluating Consequential Validity of AA-AAS Presented at OSEP Conference January 15, 2008 by Marianne Perie Center for Assessment.

Slides:



Advertisements
Similar presentations
Leon County Schools Performance Feedback Process August 2006 For more information
Advertisements

School Based Assessment and Reporting Unit Curriculum Directorate
Intelligence Step 5 - Capacity Analysis Capacity Analysis Without capacity, the most innovative and brilliant interventions will not be implemented, wont.
Action Research Not traditional educational research often research tests theory not practical Teacher research in classrooms and/or schools/districts.
Purpose of Instruction
Considering Assistive Technology and the AT Plan Sara Menzel, ATP UCP Michigan Assistive Technology Center
Continuing QIAT Conversations Joan Breslin Larson Follow up webinar post Feb for AT Conference for AT Teams Hosted by Oklahoma.
Domain 1: Planning and Preparation
Screen 1 of 24 Reporting Food Security Information Understanding the User’s Information Needs At the end of this lesson you will be able to: define the.
Standards for Qualitative Research in Education
Consistency of Assessment
1 An Introduction to Validity Arguments for Alternate Assessments Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland.
NCTM’s Focus in High School Mathematics: Reasoning and Sense Making.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Consequential Validity Inclusive Assessment Seminar Elizabeth.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Standard Setting Inclusive Assessment Seminar Marianne.
Introduction & Background Laurene Christensen National Center on Educational Outcomes National Center on Educational Outcomes (NCEO)
National Center on Educational Outcomes N C E O What the heck does proficiency mean for students with significant cognitive disabilities? Nancy Arnold,
1 Some Key Points for Test Evaluators and Developers Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland October.
Consistency/Reliability
Business research methods: data sources
Chapter 3 Preparing and Evaluating a Research Plan Gay and Airasian
GOAL SETTING CONFERENCES BRIDGEPORT, CT SEPTEMBER 2-3,
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Alignment Inclusive Assessment Seminar Brian Gong Claudia.
Understanding Validity for Teachers
What is program success? Wendy Tackett, Ph.D., Evaluator Valerie L. Mills, Project Director Adele Sobania, STEM Oakland Schools MSP, Michigan.
OCTOBER ED DIRECTOR PROFESSIONAL DEVELOPMENT 10/1/14 POWERFUL & PURPOSEFUL FEEDBACK.
1 Early Childhood Special Education Connecticut State Department of Education Early Childhood Special Education Maria Synodi.
Meeting SB 290 District Evaluation Requirements
Assessing the Curriculum Gary L. Cates, Ph.D., N.C.S.P.
Becoming a Teacher Ninth Edition
Argumentation in Middle & High School Science Victor Sampson Assistant Professor of Science Education School of Teacher Education and FSU-Teach Florida.
Connecting State Performance Plans and Technical Assistance Through Evidence-Based Education Melissa Price Higher Education Support Center for SystemsChange.
Interstate New Teacher Assessment and Support Consortium (INTASC)
Sociology 3322a. “…the systematic assessment of the operation and/or outcomes of a program or policy, compared to a set of explicit or implicit standards.
The Evaluation Plan.
Martha Thurlow and Laurene Christensen National Center on Educational Outcomes CEC Preconvention Workshop #4 April 21, 2010.
DLM Early Childhood Express Assessment in Early Childhood Dr. Rafael Lara-Alecio Dr. Beverly J. Irby
40 Performance Indicators. I: Teaching for Learning ST 1: Curriculum BE A: Aligned, Reviewed and Monitored.
Classroom Assessments Checklists, Rating Scales, and Rubrics
Semester 2: Lecture 9 Analyzing Qualitative Data: Evaluation Research Prepared by: Dr. Lloyd Waller ©
INTERNATIONAL SOCIETY FOR TECHNOLOGY IN EDUCATION working together to improve education with technology Using Evidence for Educational Technology Success.
Evaluating a Research Report
Including Quality Assurance Within The Theory of Action Presented to: CCSSO 2012 National Conference on Student Assessment June 27, 2012.
Planning and Integrating Curriculum: Unit 4, Key Topic 1http://facultyinitiative.wested.org/1.
Teaching Today: An Introduction to Education 8th edition
A Principled Approach to Accountability Assessments for Students with Disabilities CCSSO National Conference on Student Assessment Detroit, Michigan June.
Fourth session of the NEPBE II in cycle Dirección de Educación Secundaria February 25th, 2013 Assessment Instruments.
Exploring Evidence.
Research Methods in Psychology Chapter 2. The Research ProcessPsychological MeasurementEthical Issues in Human and Animal ResearchBecoming a Critical.
P.R.I.D.E. School Professional Day :45 am- 3:30 pm.
Data Report July Collect and analyze RtI data Determine effectiveness of RtI in South Dakota in Guide.
LEARNING RESEARCH AND DEVELOPMENT CENTER © 2004 University of Pittsburgh 1 Principles of Learning: Accountable Talk SM Accountability to the Learning Community.
The Development and Validation of the Evaluation Involvement Scale for Use in Multi-site Evaluations Stacie A. ToalUniversity of Minnesota Why Validate.
Catholic College at Mandeville Assessment and Evaluation in Inclusive Settings Sessions 3 & /14/2015 Launcelot I. Brown Lisa Philip.
Student Learning Objectives (SLO) Resources for Science 1.
A presentation of methods and selected results 1.
Chapter 6 Curriculum-Based Classroom Assessment Techniques.
Chapter 14: Affective Assessment
Assessment Small Learning Communities. The goal of all Small Learning Communities is to improve teaching, learning, and student outcomes A rigorous, coherent.
Weighting components of teacher evaluation models Laura Goe, Ph.D. Research Scientist, ETS Principal Investigator for Research and Dissemination, The National.
Candidate Support. Working Agreements Attend cohort meetings you have agreed upon. Start and end on time; come on time and stay for the whole time. Contribute.
CONNECT WITH CAEP | | Measures of Teacher Impact on P-12 Students Stevie Chepko, Sr. VP for Accreditation.
Conducting a research project. Clarify Aims and Research Questions Conduct Literature Review Describe methodology Design Research Collect DataAnalyse.
Welcome to the (ENTER YOUR SYSTEM/SCHOOL NAME) Data Dig 1.
Research Philosophies, Approaches and Strategies Levent Altinay.
Implementing the Professional Growth Process Session 3 Observing Teaching and Professional Conversations American International School-Riyadh Saturday,
The IEP: Progress Monitoring Process. Session Agenda Definition Rationale Prerequisites The Steps of Progress Monitoring 1.Data Collection –Unpack Existing.
DATA COLLECTION METHODS IN NURSING RESEARCH
Consequential Validity
An Introduction to NETS*T
Presentation transcript:

Evaluating Consequential Validity of AA-AAS Presented at OSEP Conference January 15, 2008 by Marianne Perie Center for Assessment

2 Purpose of this Presentation Define and summarize the literature on consequential validity Discuss the rationale for examining consequential validity of AA-AAS Propose hypotheses for testing Develop research designs for testing hypotheses

Center for Assessment 3 Definition of Validity “Validity refers to the degree to which evidence and theory support the interpretations of test scores entailed by proposed uses of tests.” (AERA, APA, NCME, 1999, p. 9)

Center for Assessment 4 Developing a Validity Argument Validity argument provides “an overall evaluation of the plausibility of the proposed interpretations and uses of test scores.” (Cronbach, 1988; Kane, 2002) In order to develop a validity argument “it is necessary to be clear about what the interpretation claims.”… The interpretive argument “provides an explicit statement of a proposed interpretation/use and a framework for developing a validity argument.” (Kane, 2002, p. 31)

Center for Assessment 5 Consequential Validity Consequential validity focuses on the implementation and use of the test as well as decisions made and actions based on test interpretation. Concept first appeared in third edition of Educational Measurement when Messick, 1989 talked about the consequential basis or aspect of validity. Controversy over whether consequences should be considered a part of validity or a separate aspect of test use. Regardless, there is much agreement that the consequences/ results/impact of using a test should be evaluated.

Center for Assessment 6 Application to AA-AAS NCLB requires that state assessment systems (including alternate assessments) “be valid for the purposes for which the assessment system is used; be consistent with relevant nationally recognized professional and technical standards, and be supported by evidence of adequate technical quality for each purpose” (NCLB, 2002 §200.2(b)(4)(i,ii)) Developing a “unified view of validity” is an Important part of evaluating and documenting the AA-AAS (Marion & Pellegrino, 2006) EAG(s) examine consequential validity of AA-AAS

Center for Assessment 7 Application to AA-AAS (continued) Some states have still not had their alternate assessment systems approved by peer review While consequential validity is not a requirement of peer review, developing a coherent validity argument will go a long way toward satisfying peer review requirements Those who are in the early stages should focus on construct validity but incorporate consequential validity into the full validity argument Those who are further along should be paying close attention to consequential validity

Center for Assessment 8 Considering Consequential Validity from the Beginning You should not wait until the assessment has been in place for several years before beginning an evaluation of consequential validity Early on, document purposes and intended outcomes of your alternate assessment program This will later provide the basis for determining what evidence you should gather on consequential validity

Center for Assessment 9 Think of the Big Picture Clearly state the theory behind test development and the ideology that supports test use. Develop statements of expected consequences Collect evidence to support planned uses and consequences Reckase, 1998 Let’s take a moment to do this – jot down some ideas about the effects you expect the AA-AAS to have in your state.

Center for Assessment 10 What Parts of Education Might Statewide Assessment Systems Impact? Implemented curriculum Instructional content and strategies Content and format of classroom assessment Student, teacher, and administrator motivation and effort Improvement of learning for all students Nature of professional development support Teacher participation in the administration, development, and scoring of the assessment Student, teacher, administrator, and public awareness and beliefs about assessment From Lane, Parke, & Stone, 1998

Center for Assessment 11 Additional Impacts of AA-AAS IEP development Identification rates of special education students Attitudes towards special education students Self-efficacy of special education students Classroom inclusion

Center for Assessment 12 Also Consider Negative Impacts Narrowing of curriculum Cheating Bad test preparation activities Further stigma on or bias toward special education students Changing student designations based solely on test assignment Increased turnover of special education teachers

Center for Assessment 13 Develop Hypotheses on How Testing Might Impact Education Systems From Lane & Stone, 2002  School administrators and teachers are motivated to adapt the instruction and curriculum to the standards  Professional development support is being provided  Instruction and curriculum will be adapted  Students are motivated to learn as well as to put forth their best effort on the assessment  Improved performance is related to changes in instruction For AA-AAS, add  Students are appropriately identified for participation in the AA-AAS  Student IEPs are written to reflect appropriately the content assessed

Center for Assessment 14 To Test Hypotheses Use multiple stakeholders to get multiple perspectives Look at outcomes at multiple levels Triangulate results using multiple measures Cautions:  When making causal attributions be sure to rule out alternative hypotheses  Be open to finding unexpected (and sometimes negative) consequences

Center for Assessment 15 Stakeholders Teachers Students Parents School administrators Community members Policymakers

Center for Assessment 16 Outcomes at Multiple Levels Consider different levels of outcomes when planning studies How does the use of AA-AAS affect the  State  District  School  Classroom  Family  Community

Center for Assessment 17 Multiple Measures Surveys Focus Groups  Gather richer more qualitative data to explain or probe more deeply into survey findings Interviews  e.g., with district curriculum or special ed coordinator Direct measures  e.g., instructional materials Direct (classroom) observation

Center for Assessment 18 Surveys Much work has been done at UKY (e.g., Towles- Reeves & Kleinert, 2006) The primary focus of these surveys has been on changes in instruction and IEP development They also consider changes in inclusion and perception Cautions:  May appear biased against finding negative impacts  May provide more informative results if baseline data are collected first

Center for Assessment 19 Sample Questions from UKY Surveys What benefits to students have accrued from participation in AA- AAS? What is the impact of the AA-AAS on students’ IEP development? What student, teacher, and instructional variables influence parents’/teachers’ perceptions regarding the AA-AAS? What benefits to teachers have accrued from the participation of students in the AA-AAS? To what extent are AA-AAS [activities]* a part of daily classroom routine? What is the impact of the AA-AAS on teachers’ daily instruction? What is the relationship between student performance in AA-AAS and post-school life outcomes? * substitute as appropriate: portfolios, structured tasks

Center for Assessment 20 Other Questions to Consider Questions about the performance levels:  What impact does a student’s performance level have on instruction?  What impact do the PLDs have on determining the sequence for curriculum and instruction? Questions about affect  Student and teacher Questions about test preparation  Look for negative consequences Specific questions about score usage  Look for both positive and negative consequences Questions about costs of using AA-AAS

Center for Assessment 21 Focus Groups and Interviews Use information from surveys to find areas to explore in depth Conduct focus groups or interviews with various stakeholders Provides qualitative information to support quantitative information gathered through the surveys

Center for Assessment 22 Direct Evidence and Observation Evidence of changes in curriculum or classroom instruction could be gathered and analyzed in a document review Evidence of changes in classroom assessment also might be available in document form Other changes (e.g., attitude, motivation, time allocation) may need to be observed through systematic classroom observation  Researcher  School administrator filling out specific form

Center for Assessment 23 Triangulate Results Draw conclusions when you have evidence from multiple sources Example  A teacher says on a survey response that she has changed her instruction to include more math curriculum after seeing the AA-AAS emphasis on academic content. This result is confirmed through examining copies of lesson plans from before and after the teacher implemented the change. In addition, the principal observed the classroom and provided information on the amount of time spent on academic math. A parent survey also indicated that students in this classroom were learning more academic math. Caution  Remember to disprove alternate hypotheses—are there other reasons for the change in emphasis?

Center for Assessment 24 Prioritizing Studies What is critical to know?  Give more weight to studies that are central to your premise of how the AA-AAS operates What do you really not know?  Give more weight to studies that provide new information rather than those that confirm what you already know What are the costs  May choose the least expensive option, at least to start with Timing  Some studies will be more appropriate at certain points in the assessment cycle  Constantly re-examine the priorities over the life of the assessment program

Center for Assessment 25 Conclusion Start by looking at the big picture  Ask what impacts you expect the alternate assessment to have and on whom Develop hypotheses regarding impacts of AA-AAS Determine appropriate stakeholders Decide on the best methods for collecting data Determine the best timing for collecting each type of data Collect and examine data and make adjustments to the AA-AAS program as needed Data collection should be continuous over the life of the assessment

Center for Assessment 26 References AERA, APA, NCME. (1999). Standards for educational and psychological testing. Washington, DC: American Educational Research Association. Cronbach, L. (1988). Five perspectives on validity argument. In H. Wainer (Ed.) Test Validity (p. 3-17). Hillsdale, NJ: Erlbaum Kane, M. (2002). Validating high-stakes testing programs. Educational Measurement: Issues and Practices, 21(1), Lane, S., Parke, C., & Stone, C. (1998). A framework for evaluating the consequences of assessment programs. Educational Measurement: Issues and Practices, 17(2), Lane, S. & Stone, C. (2002). Strategies for evaluating the consequences of assessment and accountability programs. Educational Measurement: Issues and Practices, 21(1), Marion, S. & Pellegrino, J., (2006). A validity framework for evaluating the technical quality of alternate assessments. Educational Measurement: Issues and Practices, 25(4), Messick, S. (1989). Validity. In R. L. Linn (Ed.) Educational Measurement (3 rd ed.). New York: American Council on Education. Reckase, M. (1998). Consequential validity from the test developer’s perspective. Educational Measurement: Issues and Practices, 17(2), Towles-Reeves, E & Kleinert, H. (2006). Impact of one state’s alternate assessment upon instruction and IEP development. Rural Special Education Quarterly, 25(3),