Comprehensive Assessment System Webinar #6 December 14, 2011.

Slides:



Advertisements
Similar presentations
1 New York State English as a Second Language Achievement Test (NYSESLAT) Presented by: Vanessa Lee Mercado Assistant in Educational Testing Office of.
Advertisements

Fairness, Accuracy, & Consistency in Assessment
TESTING SPEAKING AND LISTENING
Daniel Peck January 28, SLOs versus Course Objectives Student Learning Outcomes for the classroom describe the knowledge, skills, abilities.
Chapter 1 What is listening?
Testing What You Teach: Eliminating the “Will this be on the final
National Center on Response to Intervention RTI Implementer Webinar Series: What is Screening?
Student Learning Targets (SLT) You Can Do This! Getting Ready for the School Year.
ASSESSMENT LITERACY PROJECT4 Student Growth Measures - SLOs.
California English Language Development Test Review of the Test Composition.
A Terse Self-Test about Testing
Effective Intervention Using Data from the Qualitative Reading Inventory (QRI-5) Developed by the authors of the Qualitative Reading Inventory (QRI) -5,
FORMATIVE/INTERIM ASSESSMENTS OCM BOCES NETWORK TEAM.
Student Learning Objectives Session 3 Denver Public Schools Assessment, Research and Evaluation, 2014.
DEEPENING ASSESSMENT LITERACY Fall Objective  Identify best practices for local assessment development  Provide a working knowledge of the WPSD.
Teaching and Testing Pertemuan 13
BASIC PRINCIPLES OF ASSSESSMENT RELIABILITY & VALIDITY
Understanding Validity for Teachers
Questions to check whether or not the test is well designed: 1. How do you know if a test is effective? 2. Can it be given within appropriate administrative.
Chapter 4. Validity: Does the test cover what we are told (or believe)
Stages of testing + Common test techniques
is a receptive skill can be described as the process of extracting meaning from printed or written material.
Foreign language and English as a Second Language: Getting to the Common Core of Communication. Are we there yet? Marisol Marcin
6 th semester Course Instructor: Kia Karavas.  What is educational evaluation? Why, what and how can we evaluate? How do we evaluate student learning?
Student Growth Measures in Teacher Evaluation Module 2: Selecting Appropriate Assessments 1.
PARCC Information Meeting FEB. 27, I Choose C – Why We Need Common Core and PARCC.
Comprehensive Assessment System Webinar #7 January 11, 2012.
Information on New Regents Examinations for SCDN Presentation September 19, 2007 Steven Katz, Director Candace Shyer, Bureau Chief Office of Standards,
Science This school year, all of my 6 th grade science students will demonstrate measurable growth in their ability to apply the scientific practices.
Launching the Common Core State Standards Embrace Initiative Presented by Brittany Austin Literacy Interventionist.
Technical Adequacy Session One Part Three.
Ensuring State Assessments Match the Rigor, Depth and Breadth of College- and Career- Ready Standards Student Achievement Partners Spring 2014.
Human Capital Management Office of Human Resources Office of Professional Development & Evaluation.
Annual Professional Performance Review (APPR). What are the components of APPR? Teacher Evaluation –60 points (observation*/goal setting) –20 points (State.
What is Open response?.  A Situation Reading Open Response will have a story, a poem, or an article to read.  A Task Set of questions/prompt to answer.
Joanne Chen Irvine Valley College.  SLOs are statements that specify what students will know, be able to perform and to demonstrate.  SLOs specify an.
1 Michigan Educational Assessment Program (MEAP) Fall 2005 Test Design English Language Arts Grades 3-8, High School Updated March 2005.
Understanding the TerraNova Test Testing Dates: May Kindergarten to Grade 2.
Chap. 2 Principles of Language Assessment
Arizona English Language Learner Assessment AZELLA
Alternate Assessments: A Case Study of Students and Systems: Gerald Tindal UO.
Everything You Need to Know for Spring 2010 S. Kashima.
Reliability vs. Validity.  Reliability  the consistency of your measurement, or the degree to which an instrument measures the same way each time it.
Illustration of a Validity Argument for Two Alternate Assessment Approaches Presentation at the OSEP Project Directors’ Conference Steve Ferrara American.
Introduction to Advanced Placement Language and Composition.
EQAO Assessments and Rangefinding
Student Learning Objectives. Introductions Training Norms Be present Actively participate in activities Respect time boundaries Use electronics respectfully.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Assessment. Workshop Outline Testing and assessment Why assess? Types of tests Types of assessment Some assessment task types Backwash Qualities of a.
Alternate Proficiency Assessment Erin Lichtenwalner.
ANALYSIS AND ATTRIBUTES OF APPROPRIATE ASSESSMENTS Coastal Carolina University.
Colorado Student Assessment Program Colorado Department of Education Unit of Student Assessment CSAP Administration Training 2008.
EDUC 5535 Spring  An artifact of the eugenics movement (in the 1920’s) - an attempt to sort people by their perceived intelligence or ability.
Nurhayati, M.Pd Indraprasta University Jakarta.  Validity : Does it measure what it is supposed to measure?  Reliability: How the representative is.
1 Scoring Provincial Large-Scale Assessments María Elena Oliveri, University of British Columbia Britta Gundersen-Bryden, British Columbia Ministry of.
SLO Check-in December 2015 SLOs in LEAP Midyear Conversations.
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
Assessment Literacy and the Common Core Shifts. Why Be Assessment Literate? ED Test and Measurement 101 was a long time ago Assure students are being.
Handout Regarding Development of Quality Assessments to Determine Student Growth Objectives.
English IV MSL/ Common Exams. Purpose of the Assessments  They replace locally developed assessments.  Common Exam scores will count as final exam grades,
 WIDA MODEL: Grades 1-12 Measure of Developing English Language.
March, 2016 SLO End of Course Command Levels. OUTCOMES Teachers will… be prepared to determine end of course command levels for each student. be prepared.
Principles of Language Assessment
Reliability & Validity
Confidential - For internal NYSED Use Only - Not for Distribution
Your introduction to this year’s English exam.
A study guide for students.
Designing Your Performance Task Assessment
Your SLO.
Qualities of a good data gathering procedures
Presentation transcript:

Comprehensive Assessment System Webinar #6 December 14, 2011

Session Topic: Validity & Reliability

Session Objectives The purpose of this session is to: 1)Define validity and reliability 2)Distinguish between valid and invalid inferences 3)Understand how to apply knowledge of validity and reliability to the selection and development of assessments

Defining Validity Validity refers to the accuracy of inferences drawn from an assessment. It is the degree to which the assessment measures what it is intended to measure.

Types of Validity Construct validity- the assessment actually measures what it is designed to measure. A actually is A

Types of Validity Concurrent validity- the assessment correlates with other assessments that measure the same construct. A correlates with B

Types of Validity Predictive validity- the assessment predicts performance on a future assessment. A predicts B

Valid Inferences Validity is closely tied to the purpose or use of an assessment. DON’T ASK: “Is this assessment valid?” ASK: “Are the inferences I’m making based on this assessment valid for my purpose?”

Evidence-Centered Design Validity is about providing strong evidence Evidence-centered design boosts validity – What do you want to know? – How would you know? – What should the assessment look like?

Defining Reliability Reliability refers to consistency and repeatability. A reliable assessment provides a consistent picture of what students know, understand, and are able to do.

Remember! An assessment that is highly reliable is not necessarily valid. However, for an assessment to be valid, it must also be reliable.

Purchasing & Developing Assessments If necessary, consider commercial assessments or create a new assessment What assessments do you already have that purport to measure this? What are you trying to measure? PurposeReviewPurchaseDevelop

Considerations Using what you have – Is it carefully aligned to your purpose? Purchasing a new assessment – Is it carefully matched to your purpose? – Do you have the funds (for assessment, equipment, training)? Developing a new assessment – Do you have the in-house content knowledge? – Do you have the in-house assessment knowledge? – Does your team have time for development? – Does your team have the knowledge and time needed for proper scoring?

Improving Validity & Reliability Ensure questions are based on taught curricula Ensure questions are based on standards Allow students to demonstrate knowledge/skills in multiple ways Ensure a variety of item types (multiple-choice, constructed response) Ask questions at varying Depth of Knowledge levels Ensure accurate test administration Include items that address the full range of standards Include multiple items that assess the same standard Review scorer reliability, when necessary

V&R : Student Learning Objectives What makes high-quality evidence for SLOs: Aligned to the content standards (construct validity) Being used for the purpose for which it was designed Administered properly

Ex. SLO Objective: Students will demonstrate grade- level proficiency in reading, writing, and speaking French, including the accurate use of past and present tenses. How would you know if students were proficient in reading, writing, and speaking French?

Ex. SLO Evidence: 1. Written final exam measuring reading comprehension, vocabulary, conjugation/agreement in past and present tenses word written composition in French, using past and present tense in a familiar content theme minute conversation on one of 3 pre-selected topics, using past and present tense.

Ex. SLO Administration & Scoring The exam and composition will be part of the written final, administered during the final exam period. – I will score the compositions using the Foreign Language Department level 2 writing rubric, which includes vocabulary, tense, subject-verb agreement, spelling, level of detail, etc. – Approximately 20% of the compositions will also be double-scored by the other French teacher. The oral assessment will be administered one-on-one in the last week of school, prior to the exam period. I will develop the rubric with the other French teacher and have it approved by the Department Chair. I will administer and score most oral exams myself, though I will schedule my Department Chair to sit in on and double-score the first 20%.

Questions?

Upcoming Webinars January 11 th 9:30-10:30 Cultural & Linguistic Demands of Assessment