Validity and Validation: An introduction Note: I have included explanatory notes for each slide. To access these, you will probably have to save the file.

Slides:



Advertisements
Similar presentations
Questionnaire Development
Advertisements

Standardized Scales.
Chapter 8 Flashcards.
Reliability and Validity
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
The Research Consumer Evaluates Measurement Reliability and Validity
Reliability and Validity
1Reliability Introduction to Communication Research School of Communication Studies James Madison University Dr. Michael Smilowitz.
Research Methodology Lecture No : 11 (Goodness Of Measures)
Measurement Reliability and Validity
Reliability and Validity of Research Instruments
Concept of Measurement
Class 3: Thursday, Sept. 16 Reliability and Validity of Measurements Introduction to Regression Analysis Simple Linear Regression (2.3)
Measurement: Reliability and Validity For a measure to be useful, it must be both reliable and valid Reliable = consistent in producing the same results.
Personality, 9e Jerry M. Burger
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
FINAL REPORT: OUTLINE & OVERVIEW OF SURVEY ERRORS
Reliability and Validity. Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? How do we judge.
Rosnow, Beginning Behavioral Research, 5/e. Copyright 2005 by Prentice Hall Ch. 6: Reliability and Validity in Measurement and Research.
Understanding Validity for Teachers
Measurement Concepts & Interpretation. Scores on tests can be interpreted: By comparing a client to a peer in the norm group to determine how different.
The Practice of Social Research
Measurement and Data Quality
Reliability, Validity, & Scaling
Measurement in Exercise and Sport Psychology Research EPHE 348.
Instrumentation.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
I Can Distinguish the types of validity Distinguish the types of reliability Identify if an example is objective or subjective Copyright © Allyn & Bacon.
Final Study Guide Research Design. Experimental Research.
CS130 – Software Tools Fall 2010 Statistics and PASW Wrap-up 1.
Lecture 6: Reliability and validity of scales (cont) 1. In relation to scales, define the following terms: - Content validity - Criterion validity (concurrent.
Validity. Face Validity  The extent to which items on a test appear to be meaningful and relevant to the construct being measured.
Reliability & Validity
6. Evaluation of measuring tools: validity Psychometrics. 2012/13. Group A (English)
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Measurement Validity.
VALIDITY AND VALIDATION: AN INTRODUCTION Note: I have included explanatory notes for each slide. To access these, you will probably have to save the file.
Confidence Interval Estimation For statistical inference in decision making:
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
The Theory of Sampling and Measurement. Sampling First step in implementing any research design is to create a sample. First step in implementing any.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
MOI UNIVERSITY SCHOOL OF BUSINESS AND ECONOMICS CONCEPT MEASUREMENT, SCALING, VALIDITY AND RELIABILITY BY MUGAMBI G.K. M’NCHEBERE EMBA NAIROBI RESEARCH.
Validity: Introduction. Reliability and Validity Reliability Low High Validity Low High.
Measurement Issues General steps –Determine concept –Decide best way to measure –What indicators are available –Select intermediate, alternate or indirect.
 Measuring Anything That Exists  Concepts as File Folders  Three Classes of Things That can be Measured (Kaplan, 1964) ▪ Direct Observables--Color of.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Confidence Interval Estimation For statistical inference in decision making: Chapter 9.
1 Announcement Movie topics up a couple of days –Discuss Chapter 4 on Feb. 4 th –[ch.3 is on central tendency: mean, median, mode]
TEST SCORES INTERPRETATION - is a process of assigning meaning and usefulness to the scores obtained from classroom test. - This is necessary because.
Assistant Instructor Nian K. Ghafoor Feb Definition of Proposal Proposal is a plan for master’s thesis or doctoral dissertation which provides the.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
Questionnaire-Part 2. Translating a questionnaire Quality of the obtained data increases if the questionnaire is presented in the respondents’ own mother.
RELIABILITY AND VALIDITY Dr. Rehab F. Gwada. Control of Measurement Reliabilityvalidity.
CRITICALLY APPRAISING EVIDENCE Lisa Broughton, PhD, RN, CCRN.
Measurement and Scaling Concepts
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Reliability and Validity
Chapter 4 Research Methods in Clinical Psychology
Reliability and Validity in Research
Concept of Test Validity
Reliability & Validity
Introduction to Measurement
Human Resource Management By Dr. Debashish Sengupta
Week 3 Class Discussion.
Analyzing Reliability and Validity in Outcomes Assessment Part 1
Measurement Concepts and scale evaluation
Analyzing Reliability and Validity in Outcomes Assessment
Presentation transcript:

Validity and Validation: An introduction Note: I have included explanatory notes for each slide. To access these, you will probably have to save the file to your computer, then look at it in the “normal view”

Conceptions of Validity 1.Does this test measure what it is supposed to measure? 2.Does the measure correlate with other (perhaps more expensive or invasive) measures of the concept? 3.Validity is present when the attribute to be measured causes variations in the measurement outcomes 4.What interpretations can fairly be placed on scores produced by this measurement? 5.How well does this measure fit in with a hypothesized network of related, and contrasting, constructs?

The Poetry of Donald Rumsfeld may offer insights into validity...? The Unknown As we know, There are known knowns: there are things we know we know. We also know There are known unknowns. That is to say We know there are some things We do not know. But there are also unknown unknowns, The ones we don't know We don't know. —Feb. 12, 2002, Department of Defense news briefing Clarity I think what you'll find, I think what you'll find is, Whatever it is we do substantively, There will be near-perfect clarity As to what it is. And it will be known, And it will be known to the Congress, And it will be known to you, Probably before we decide it, But it will be known. —Feb. 28, 2003, Department of Defense briefing Source:

Reliability and Validity Reliability Low High Validity Low High Biased result! ☺ Average of these inaccurate results is not bad. This is probably how screening questionnaires (e.g., for depression) work

Validity viewed as error in measurement. Types of Error in Repeated Measurements True value ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ individual measures Bias! Random error

The Sequence of Validation Checks (1) “Does it measure what it’s supposed to measure?” 1.First, decide and define what you want to measure (if it’s an abstract concept, explain the conceptual basis) 2.Select indicators or items that represent that topic (this involves content validity: a sampling of potential questions) 3.Check that items are clear, comprehensible and relevant (face validity, “sensibility”) 4.This produces a pool of items ready for the item analysis stage, which involves administering the test and analyzing responses (next slide).

Validation Sequence (2): Checking the internal structure “Item analysis” refers to a series of checks on the performance of each “item” (or question). Some analyses fall under the heading of reliability, some validity. Faulty items are discarded or replaced. Analyses include: 2.1 Item distributions & missing values: an item that does not vary or that people don’t answer cannot measure anything 2.2 Correlations among items, maybe using factor analysis 2.3 Item response theory (IRT) analyses.

Validation Sequence (3): External associations 3.1 Compare the measure against a criterion, if a “gold standard” exists. Sensitivity & specificity are the normal statistics. 3.2 Where there is no single, clear gold standard you can use correlations with other indicators. This leads to construct validation. You begin from a set of hypotheses covering the expected relationships among as wide a range of indicators as possible. 3.3 Correlations often divided into convergent and discriminant coefficients, according to hypothesized associations. These analyses tend to use the entire test administered to selected samples; inadequate performance leads back to basic design.

Validation Sequence (4): Group Discrimination Once you show that test scores correlate with other measures as intended, its actual performance is evaluated in rating groups of respondents. Analyses generally use representative samples 4.1 “Known groups” (can it distinguish well from sick? Similar to criterion validity) 4.2 Responsiveness: sensitivity to change over time (which is important in an evaluative measure) 4.3 Do scores show ceiling or floor effects?

Conclusion Validation is rarely complete. Many instruments continue to be checked for validity years after their invention. Times change, phrasing makes old items obsolete, and you can also test its validity for different purposes. Validation is long and expensive. Basic test development and validation may take years: it’s not a thesis project. Remember: validity is about the interpretation of scores. It is a relative concept: a test is not valid or invalid, but only valid or not for a particular application. But recall the Viagra principle: a test intended for one purpose may prove good for an unanticipated application.