Concept of Test Validity

Slides:



Advertisements
Similar presentations
Chapter 8 Flashcards.
Advertisements

Survey Methodology Reliability and Validity EPID 626 Lecture 12.
The Research Consumer Evaluates Measurement Reliability and Validity
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
VALIDITY AND RELIABILITY
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Chapter 4A Validity and Test Development. Basic Concepts of Validity Validity must be built into the test from the outset rather than being limited to.
Assessment: Reliability, Validity, and Absence of bias
RESEARCH METHODS Lecture 18
Chapter 4 Validity.
Test Validity: What it is, and why we care.
VALIDITY.
Concept of Measurement
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Research Methods in MIS
Chapter 7 Correlational Research Gay, Mills, and Airasian
Chapter 7 Evaluating What a Test Really Measures
Classroom Assessment A Practical Guide for Educators by Craig A
Standardized Test Scores Common Representations for Parents and Students.
Understanding Validity for Teachers
Chapter 4. Validity: Does the test cover what we are told (or believe)
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Measurement Concepts & Interpretation. Scores on tests can be interpreted: By comparing a client to a peer in the norm group to determine how different.
Technical Issues Two concerns Validity Reliability
Measurement and Data Quality
Reliability, Validity, & Scaling
Instrumentation.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Technical Adequacy Session One Part Three.
Validity & Practicality
Chapter 4: Test administration. z scores Standard score expressed in terms of standard deviation units which indicates distance raw score is from mean.
Validity. Face Validity  The extent to which items on a test appear to be meaningful and relevant to the construct being measured.
Validity Is the Test Appropriate, Useful, and Meaningful?
6. Evaluation of measuring tools: validity Psychometrics. 2012/13. Group A (English)
Measurement Validity.
Selecting a Sample. Sampling Select participants for study Select participants for study Must represent a larger group Must represent a larger group Picked.
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
Criteria for selection of a data collection instrument. 1.Practicality of the instrument: -Concerns its cost and appropriateness for the study population.
Chapter 6 - Standardized Measurement and Assessment
VALIDITY, RELIABILITY & PRACTICALITY Prof. Rosynella Cardozo Prof. Jonathan Magdalena.
Classroom Assessment Chapters 4 and 5 ELED 4050 Summer 2007.
Reliability EDUC 307. Reliability  How consistent is our measurement?  the reliability of assessments tells the consistency of observations.  Two or.
TEST SCORES INTERPRETATION - is a process of assigning meaning and usefulness to the scores obtained from classroom test. - This is necessary because.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
Measurement and Scaling Concepts
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Survey Methodology Reliability and Validity
VALIDITY by Barli Tambunan/
Lecture 5 Validity and Reliability
DUMMIES RELIABILTY AND VALIDITY FOR By: Jeremy Starkey Lijia Zhang
Reliability and Validity
Questions What are the sources of error in measurement?
Reliability and Validity in Research
Test Validity.
Reliability & Validity
Human Resource Management By Dr. Debashish Sengupta
پرسشنامه کارگاه.
Reliability and Validity of Measurement
RESEARCH METHODS Lecture 18
Chapter 4 Characteristics of a Good Test
Qualities of a good data gathering procedures
Presentation transcript:

Concept of Test Validity SOWUNMI E. T.

Meaning and Nature of Test Validity Validity deals with the degree to which a measures what it claims to measure. Test validity is the degree to which the test measures what it is supposed to measure. Reliability has more to do with consistency of a measure (repeatability) while validity has to do with accuracy (how accurately it measure what it is purported to measure from the start).

Validity is a matter of degree in nature, which means that it ranges from low through medium to high. The statistics for indicating the degree of validity is known as coefficient of validity which ranges from zero to one.

Methods of Measuring of an Instruments Validity 1. Content Validity: it involves the systematic examination of the test content to determine whether it covers a representative sample of the behaviour domain it is purported to measure. It could be checked by subject specialist or panel experts or It could also be done by preparing a table of specification. The closer the test contents to the subject matter, the higher the content validity.

2. Criterion-Related Validity: It correlate or compare the test with other existing standard measures (criteria). Criterion related validity looks at the relationship between a test score and an outcome. Example: intelligent quotient tests are often validated against measures of academic performance (the criterion) NOW, If the test data and criterion data are collected at the same time, this is referred to as Concurrent validity evidence. . If the new test is validated by a comparison with a currently existing criterion, we have concurrent validity.

Criterion-Related Validity (contd.) If the test data are collected first in order to predict criterion data collected at a later point in time, then this is referred to as Predictive validity evidence. Predictive validity measures the degree to which a test accurately predicts or foretell a criterion that will occur in the future. It refers to the power or usefulness of test scores to predict future performance.

3. Construct validity: It is referred to as the degree to which a test/instrument measures the underlying theoretical construct it is supposed to measure. It is the extent to which a test captures a specific trait known as construct. In order to demonstrate construct validity, evidence that the test measures what it purports to measure as well as evidence that the test does not measure irrelevant attributes are both required. These are referred to as convergent and discriminant validity. i. convergent validity: if the two different tests end up ranking students similarly i.e when two different tests measures related skills. Ii. discriminant validity: providing evidence that two tests that do not measure closely related skills or types of knowledge do not correlate strongly.

Methods of establishing construct validity: Correlations with other tests that has been proven valid for the same trait or construct. Criterion group studies: when a test is administered to groups of students’ judged to be good or poor in a construct and it was proven so at the end of an administered test. Analysis of the mental process required by the test Comparing pre-test and post-test scores of persons exposed to some relevant treatment: it indicate how a treatment or condition may or may not affect a particular set of learners.

4. Face Validity: It is simply an estimate as to whether a test appears to measure what it claims to measure. It can be related to the question of whether a non- professional testers such as parents think the test is appropriate or not appropriate. A direct measurement of face validity is obtained by asking people to rate the validity of a test as it appears to them. The question that the face validity provides its answer is how much does a test look like an appropriate measure of the skills or body of knowledge concerned?

Principles for Test Validity Appropriate interpretation: tests are valid to the degree that evidence which supports their appropriateness and correctness can be identified. Appropriate uses: a test is valid to the degree by which evidence that supports their correctness and appropriateness use of the test result. Appropriate values: when the values implied by them are appropriate. Appropriate consequences: when the consequences of these interpretations and uses are consistent with appropriate values.

Factors affecting Test Validity These factors has been categorized into four main headings (a) Factors in the test: Difficulty level of test items should be able to discriminate effectively between the good and poor students Lack of clear instructions: test that lack in this tend to confuse the testee. Poor quality of test items: test items may be inappropriate for the cognitive level of the respondents thereby invalidating the results. Test items should not be vague (it should be clear not having different meaning to different testees) Test length: a test must not be too short nor too long otherwise it shows that the test do not adequately represent the content or has been over-stretched or over emphasis.

(a) Factors in the test (contd.) Difficult vocabulary: when a test items contain difficult words, it is more or less measuring another trait which is not the achievement in that subject. Identifiable patterns of answers: when students identify the pattern of answer in multiple choice tests, they easily guess the answer to some items. This tends to lower the validity of test result. (b) Factors in test administration and scoring: factors such as cheating, undue assistance given to students, insufficient time and unreliable scoring of essay answers will tend to reduce the validity of test results.

Factors affecting Test Validity (contd.) (C) Factors affecting validity coefficients: The larger the spread of scores, the higher the validity coefficient. Therefore, factors such as speed of test and item difficulty which reduces discrimination and spread of scores may reduce the validity coefficient. (d) Factors affecting students responses: variables such as sex (gender), age, and personality traits. Likewise, unhealthy physical and psychological conditions of the testing hall, emotional disturbance of the students may prevent students from achieving maximally.

End of the first topic.