VALIDITY AND VALIDATION: AN INTRODUCTION Note: I have included explanatory notes for each slide. To access these, you will probably have to save the file.

Slides:



Advertisements
Similar presentations
Assessing Student Performance
Advertisements

Standardized Scales.
Chapter 8 Flashcards.
Conceptualization, Operationalization, and Measurement
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
ASSESSING RESPONSIVENESS OF HEALTH MEASUREMENTS. Link validity & reliability testing to purpose of the measure Some examples: In a diagnostic instrument,
The Research Consumer Evaluates Measurement Reliability and Validity
Taking Stock Of Measurement. Basics Of Measurement Measurement: Assignment of number to objects or events according to specific rules. Conceptual variables:
1Reliability Introduction to Communication Research School of Communication Studies James Madison University Dr. Michael Smilowitz.
Research Methodology Lecture No : 11 (Goodness Of Measures)
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Measurement Reliability and Validity
Chapter 13 Survey Designs
Concept of Measurement
Measurement Validity and Reliability. Reliability: The degree to which measures are free from random error and therefore yield consistent results.
Beginning the Research Design
Personality, 9e Jerry M. Burger
Chapter 13 Survey Designs
FINAL REPORT: OUTLINE & OVERVIEW OF SURVEY ERRORS
Validity and Validation: An introduction Note: I have included explanatory notes for each slide. To access these, you will probably have to save the file.
Chapter 9 Descriptive Research. Overview of Descriptive Research Focused towards the present –Gathering information and describing the current situation.
Survey Designs EDUC 640- Dr. William M. Bauer
Measurement Concepts & Interpretation. Scores on tests can be interpreted: By comparing a client to a peer in the norm group to determine how different.
Proposal Writing.
Measurement and Data Quality
Reliability, Validity, & Scaling
‘Hints for Designing Effective Questionnaires ’
Chapter 10 Questionnaire Design Chapter Objectives explain why it is important for managers or business researchers to know how to design good questionnaires.
Power Point Slides by Ronald J. Shope in collaboration with John W. Creswell Chapter 13 Survey Designs.
Measurement in Exercise and Sport Psychology Research EPHE 348.
MICS4 Survey Design Workshop Multiple Indicator Cluster Surveys Survey Design Workshop Data Analysis and Reporting.
Instrumentation.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
Introduction Theoretical Perspectives Research.  Sampling : Identifying the appropriate population of people to be studied.  Random Sample : Each member.
The Genetics Concept Assessment: a new concept inventory for genetics Michelle K. Smith, William B. Wood, and Jennifer K. Knight Science Education Initiative.
Evaluating a Research Report
Lecture 6: Reliability and validity of scales (cont) 1. In relation to scales, define the following terms: - Content validity - Criterion validity (concurrent.
Quantitative and Qualitative Approaches
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Measurement Validity.
Learning Objective Chapter 9 The Concept of Measurement and Attitude Scales Copyright © 2000 South-Western College Publishing Co. CHAPTER nine The Concept.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Evaluating Survey Items and Scales Bonnie L. Halpern-Felsher, Ph.D. Professor University of California, San Francisco.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
MOI UNIVERSITY SCHOOL OF BUSINESS AND ECONOMICS CONCEPT MEASUREMENT, SCALING, VALIDITY AND RELIABILITY BY MUGAMBI G.K. M’NCHEBERE EMBA NAIROBI RESEARCH.
Validity: Introduction. Reliability and Validity Reliability Low High Validity Low High.
JS Mrunalini Lecturer RAKMHSU Data Collection Considerations: Validity, Reliability, Generalizability, and Ethics.
Measurement Issues General steps –Determine concept –Decide best way to measure –What indicators are available –Select intermediate, alternate or indirect.
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
Assessing Responsiveness of Health Measurements Ian McDowell, INTA, Santiago, March 20, 2001.
Nurhayati, M.Pd Indraprasta University Jakarta.  Validity : Does it measure what it is supposed to measure?  Reliability: How the representative is.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Chapter Eight: Quantitative Methods
Chapter 6 - Standardized Measurement and Assessment
Stages of Test Development By Lily Novita
Copyright © 2008 by Pearson Education, Inc. Upper Saddle River, New Jersey All rights reserved. John W. Creswell Educational Research: Planning,
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
Questionnaire-Part 2. Translating a questionnaire Quality of the obtained data increases if the questionnaire is presented in the respondents’ own mother.
CRITICALLY APPRAISING EVIDENCE Lisa Broughton, PhD, RN, CCRN.
1 Measurement Error All systematic effects acting to bias recorded results: -- Unclear Questions -- Ambiguous Questions -- Unclear Instructions -- Socially-acceptable.
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 25 Critiquing Assessments Sherrilene Classen, Craig A. Velozo.
QUESTIONNAIRE DESIGN AND VALIDATION
Concept of Test Validity
Test Validity.
Reliability & Validity
پرسشنامه کارگاه.
Chapter Eight: Quantitative Methods
Presentation transcript:

VALIDITY AND VALIDATION: AN INTRODUCTION Note: I have included explanatory notes for each slide. To access these, you will probably have to save the file to your computer, then look at it in the “normal view”

Conceptions of Validity 1. Does this test measure what it is supposed to measure? 2. Does the measure correlate with other (perhaps more expensive or invasive) measures of the concept? 3. Validity is present when the attribute to be measured causes variations in the measurement outcomes 4. What interpretations can fairly be placed on scores produced by this measurement? 5. How well does this measure fit in with a hypothesized network of related, and contrasting, constructs?

Reliability and Validity Reliability Low High Validity Low High Biased result! ☺ Average of these inaccurate results is not bad. This is probably how screening questionnaires (e.g., for depression) work

Validity viewed as error in measurement. Two types of error in repeated measurements True value (which we are trying to record) ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ ¦ individual measures Bias! Random error

The idea of Bias: Measured versus Self-Reported Height and Weight Canadian Community Health Survey (N=4,080 adults 18+) MeasuredReported Height (cm)M F Weight (kg)M F Bias: 1 cm taller 2 kg lighter BMean BMI (kg/m 2 ) MeasuredReported M F Bias: BMI underestimated by 1 unit Prevalence of Obesity MeasuredReported % Normal MalesOverweight Obese Normal FemalesOverweight Obese A B C Bias: 10% underestimate in obesity 6% underestimate

First steps in Validation: the design stage “Does it measure what it’s supposed to measure?” 1. First, decide and define what you want to measure (if it’s an abstract concept, explain the conceptual basis) 2. Select indicators or “items” that represent the topic (this involves content validity: a sampling of potential questions) 3. Check that items are clear, comprehensible and relevant (face validity, “sensibility”) 4. This produces a pool of items ready for the item analysis stage, which involves administering the test and analyzing responses (next slide).

Clarity of question wording Most questionnaires are written by educated people, and they tend to use long and complicated words that may not be understood by respondents with less language ability. ‘Cognitive testing’ should be a component of developing any questionnaire. E.g., discuss with the respondent in a pilot testing session how he/she understood the question. Checking the language using a ‘readability formula’ may be helpful: Dale-Chall formula; Flesch Reading Ease; Fog Index; Fry Graph These often show that questionnaires (and consent forms!) are too difficult for many people to understand.

Second steps in validation: field testing Administer your measurement to a sample of the types of person for whom it will be used: Typically a pre-test on selected respondents, to ensure the questions are clear Often using interviews, followed by qualitative de-briefing interviews with respondents May lead to re-wording questions, changing procedures, etc. Field testing on larger sample Data collected and analyzed, both for the scale as a whole to test overall performance, and item by item (item analysis) to identify items that are not performing. Analyses of items and overall performance may be done in any sequence, often with re-analyses discarding items that do not perform.

Item analyses: Checking the internal structure Item analysis refers to a series of checks on the performance of each “item” (e.g. question). Some analyses fall under the heading of reliability, some validity. Faulty items are discarded or replaced. Item analyses include: Item distributions & missing values: an item that does not vary or that people don’t answer cannot measure anything Internal structure: correlations among items, maybe leading to factor analysis. Factorial validity Item response theory (IRT) analyses checks contributions of each item to the overall scale.

Performance of the overall test: External associations Compare the scores against a criterion, if a “gold standard” exists. Criterion validity: sensitivity & specificity are the normal statistics. Compare against other measures: concurrent validation. Correlations often divided into convergent and discriminant coefficients, according to hypothesized associations. Comparisons against a set of other indicators leads to construct validation. You begin from hypotheses covering the expected correlations with as wide a range of indicators as possible.

Performance of the overall test: Group Discrimination Once you show that scores correlate with other measures as intended, its performance is evaluated in contrasting between different groups of respondents. “Known groups validation” (can it distinguish well from sick people? Similar to criterion validity) Responsiveness to change: sensitivity to change over time (this is important in an evaluative measure) Do scores show ceiling or floor effects? If it is not performing as expected, return to item analyses to diagnose where the weakness lies. This may lead to redesign and further field testing.

Comments Validation is rarely complete. Many instruments continue to be checked for validity years after their development. Times change, phrasing makes old items obsolete, and you can also test validity for different purposes. Validation is long and expensive. Basic test development and validation may take years: it’s not a thesis project. Remember: validity is about the interpretation of scores. It is a relative concept: a test is not valid or invalid, but only valid or not for a particular application. But recall the Viagra principle: a test intended for one purpose may prove good for an unanticipated application.