Validity Is the Test Appropriate, Useful, and Meaningful?

Slides:



Advertisements
Similar presentations
Chapter 8 Flashcards.
Advertisements

Psychometrics William P. Wattles, Ph.D. Francis Marion University.
VALIDITY AND RELIABILITY
Research Methodology Lecture No : 11 (Goodness Of Measures)
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 6 Validity.
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Chapter 4A Validity and Test Development. Basic Concepts of Validity Validity must be built into the test from the outset rather than being limited to.
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
Part II Knowing How to Assess Chapter 5 Minimizing Error p115 Review of Appl 644 – Measurement Theory – Reliability – Validity Assessment is broader term.
Assessment: Reliability, Validity, and Absence of bias
RESEARCH METHODS Lecture 18
Chapter 4 Validity.
VALIDITY.
Concept of Measurement
Developing a Hiring System Reliability of Measurement.
Validity Does test measure what it says it does? Is the test useful? Can a test be reliable, but not valid? Can a test be valid, but not reliable?
Lecture 7 Psyc 300A. Measurement Operational definitions should accurately reflect underlying variables and constructs When scores are influenced by other.
Chapter 7 Evaluating What a Test Really Measures
Classroom Assessment A Practical Guide for Educators by Craig A
Reliability and Validity. Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? How do we judge.
Understanding Validity for Teachers
Chapter 4. Validity: Does the test cover what we are told (or believe)
Validity and Reliability
Ch 6 Validity of Instrument
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Technical Adequacy Session One Part Three.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
Principles of Test Construction
The World of Assessment Consider the options! Scores based on developmental levels of academic achievement Age-Equivalent scores.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Week # 4 Copyright © 2004 by R. Halstead. All rights reserved.
EDU 8603 Day 6. What do the following numbers mean?
Measurement Validity.
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Session 4 Reliability and Validity. Validity What does the instrument measure and How well does it measure what it is supposed to measure? Is there enough.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Evaluating Survey Items and Scales Bonnie L. Halpern-Felsher, Ph.D. Professor University of California, San Francisco.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Chapter 4 Validity Robert J. Drummond and Karyn Dayle Jones Assessment Procedures for Counselors and Helping Professionals, 6 th edition Copyright ©2006.
The Theory of Sampling and Measurement. Sampling First step in implementing any research design is to create a sample. First step in implementing any.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Measurement Experiment - effect of IV on DV. Independent Variable (2 or more levels) MANIPULATED a) situational - features in the environment b) task.
Chapter 6 - Standardized Measurement and Assessment
Chapter 3 Selection of Assessment Tools. Council of Exceptional Children’s Professional Standards All special educators should possess a common core of.
Standards-Based Tests A measure of student achievement in which a student’s score is compared to a standard of performance.
Measuring Research Variables
Lesson 3 Measurement and Scaling. Case: “What is performance?” brandesign.co.za.
Reliability and Validity
VALIDITY by Barli Tambunan/
Questions What are the sources of error in measurement?
Concept of Test Validity
Reliability & Validity
Test Validity.
Validity and Reliability
Journalism 614: Reliability and Validity
Reliability & Validity
Week 3 Class Discussion.
پرسشنامه کارگاه.
5. Reliability and Validity
Reliability and Validity of Measurement
RESEARCH METHODS Lecture 18
Methodology Week 5.
TESTING AND EVALUATION IN EDUCATION GA 3113 lecture 1
Presentation transcript:

Validity Is the Test Appropriate, Useful, and Meaningful?

Properties of Validity Does the test measure what it purports to measure? Validity is a property of inferences that can be made. Validity should be established over multiple inferences.

Evidence of Validity Meaning of test scores Reliability Adequate standardization and norming Content validity Criterion-related validity Construct validity

Content Validity How well does the test represent the domain? Appropriateness of items. Completeness of items. How the items assess the content.

Do the Items Match the Content? Face validity Is the item considered part of the domain Curriculum match Do the items reflect what has been taught Homogeneity with the test Is the item positively correlated with the test score? (Minimum.25) Point bi-serial r xx

Are the Total Items Representative? Are items included representing various parts of the domain? Are different parts differentially represented? How well are the parts represented?

How Is the Content Measured? Different types of measures Multiple choice, short answer Performance-based vs. Factual recall Measures should match the type of content taught Measures should match HOW the content was taught

Criterion-related (Cr)validity How well does performance on the test reflect performance on what the test purports to measure? Expressed as r xx Based on concurrent CR validity And predictive CR validity

Concurrent Criterion-related Validity How well does performance on the test estimate knowledge and/or skill on the criterion measure? Does the reading test estimate the students current reading performance? Compares performance on one test with performance with other similar tests. KTEA w/ Woodcock-Johnson

Predictive Criterion-related Validity Will performance on a test now be predictive of performance at a later time? Will a score derived now from one test be as accurate as a score derived by another and later test ? Will a student’s current reading scores accurately reflect reading measured at a later time by another test?

Criterion-related Validity Should Describe criterion measures accurately Provide rationales for choices. Provide sufficient information to judge adequacy of the criterion Describe adequately the sample of students used. Include analytical data used to determine predictiveness

Criterion-related Validity Should Basic statistics should include Number of cases Reasons for eliminating cases Central tendency estimates Provide analysis of the limits toward generalizability of the test What kind of inferences about the content can be made

Construct Validity How validly does the test measure the underlying constructs it purports to measure? Do IQ tests measure intelligence? Do self-concept scales measure self concept?

Definition of Construct A psychological or personality trait E.G. Intelligence, learning style Or A psychological concept, attribute, or theoretical characteristic E.G. Problem solving, locus of control, or learning disability

Ways to Measure Construct Validity Developmental change: determining expected differences among identified groups Assuming content validity Assuming reliability Convergent /divergent validity: high correlation with similar tests and low correlation with tests measuring different constructs

Ways to measure Construct Validity Predictive Validity: High scores on one test should predict high scores on similar tests. Accumulation of evidence FAILING TO DISPROVE: If the concept or trait tested can be influenced by intervention, intervention effects should be reflected by pre and posttest scores. If the test score should not be influenced by intervention, changes should not be reflected in pre and posttest scores.

Factors Affecting Validity Unsystematic error Lack of reliability Systematic error Bias

Reliability Effects on Validity The validity of a measure can NEVER exceed the measure’s reliability Reliability measures error Validity measures expected traits (content, constructs, criteria) r xy = r x(t)y(t) r xx r yy

Systematic Bias Method of Measurement Behaviors in testing Item selection Administration errors Norms

Who is Responsible for Valid Assessment? The Test Author Authors are reponsible for ensuring and publishing validation. The Test Giver Test administrators are reponsible for following procedures outlined in administration guidelines.

Guidelines for Giving Tests Exact Administration Read the administration instructions. Note procedures for establishing baselines. Note procedures for individual items. Practice giving the test. Appropriate Pacing Develop fluency with the test.