Reliability and Validity

Slides:



Advertisements
Similar presentations
Agenda Levels of measurement Measurement reliability Measurement validity Some examples Need for Cognition Horn-honking.
Advertisements

Chapter 8 Flashcards.
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
The Research Consumer Evaluates Measurement Reliability and Validity
Reliability And Validity
VALIDITY AND RELIABILITY
Research Methodology Lecture No : 11 (Goodness Of Measures)
RESEARCH METHODS Lecture 18
Reliability and Validity Dr. Roy Cole Department of Geography and Planning GVSU.
Concept of Measurement
Measurement: Reliability and Validity For a measure to be useful, it must be both reliable and valid Reliable = consistent in producing the same results.
Psych 231: Research Methods in Psychology
Validity, Reliability, & Sampling
Rosnow, Beginning Behavioral Research, 5/e. Copyright 2005 by Prentice Hall Ch. 6: Reliability and Validity in Measurement and Research.
Test Validity S-005. Validity of measurement Reliability refers to consistency –Are we getting something stable over time? –Internally consistent? Validity.
Measurement and Data Quality
Reliability, Validity, & Scaling
Reliability and Validity what is measured and how well.
Instrumentation.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Technical Adequacy Session One Part Three.
Principles of Test Construction
Lecture 6: Reliability and validity of scales (cont) 1. In relation to scales, define the following terms: - Content validity - Criterion validity (concurrent.
Measurement Validity.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Evaluating Survey Items and Scales Bonnie L. Halpern-Felsher, Ph.D. Professor University of California, San Francisco.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
The Theory of Sampling and Measurement. Sampling First step in implementing any research design is to create a sample. First step in implementing any.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Measurement Issues General steps –Determine concept –Decide best way to measure –What indicators are available –Select intermediate, alternate or indirect.
Reliability and Validity Themes in Psychology. Reliability Reliability of measurement instrument: the extent to which it gives consistent measurements.
Chapter 6 - Standardized Measurement and Assessment
VALIDITY, RELIABILITY & PRACTICALITY Prof. Rosynella Cardozo Prof. Jonathan Magdalena.
Reliability a measure is reliable if it gives the same information every time it is used. reliability is assessed by a number – typically a correlation.
Lesson 3 Measurement and Scaling. Case: “What is performance?” brandesign.co.za.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
WHS AP Psychology Unit 7: Intelligence (Cognition) Essential Task 7-3:Explain how psychologists design tests, including standardization strategies and.
RELIABILITY AND VALIDITY Dr. Rehab F. Gwada. Control of Measurement Reliabilityvalidity.
Measurement and Scaling Concepts
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
MGMT 588 Research Methods for Business Studies
Reliability and Validity
Chapter 2 Theoretical statement:
Unit 8: Intelligence (Cognition)
Ch. 5 Measurement Concepts.
VALIDITY by Barli Tambunan/
Lecture 5 Validity and Reliability
Questions What are the sources of error in measurement?
Concept of Test Validity
Measurement: Part 1.
Test Validity.
CHAPTER 5 MEASUREMENT CONCEPTS © 2007 The McGraw-Hill Companies, Inc.
Measurement: Part 2.
Understanding Results
Tests and Measurements: Reliability
Journalism 614: Reliability and Validity
Reliability & Validity
Introduction to Measurement
Human Resource Management By Dr. Debashish Sengupta
Week 3 Class Discussion.
پرسشنامه کارگاه.
5. Reliability and Validity
Reliability and Validity of Measurement
Unit IX: Validity and Reliability in nursing research
RESEARCH METHODS Lecture 18
Experiment Basics: Variables
Measurement Concepts and scale evaluation
Reliability and Validity
AS Psychology Research Methods
Presentation transcript:

Reliability and Validity Sa’ed H. Zyoud

The error of research Random error (Chance) (increase sample number) (Precision: free from RE) Precision is the degree to which repeated measurements under unchanged conditions show the same results Systematic error (Bias) (matching+ blinding) (Accuracy: free from SE) Accuracy is the degree of closeness of measurements of a quantity to that quantity's actual (true) value Sampling error (type 1 error, type 2 error) Measurement error (wrong stat, wrong in scoring)

Reliability Means "repeatability" or "consistency". A measure is considered reliable if it would give us the same result over and over again (assuming that what we are measuring isn't changing!). There are four general classes of reliability estimates, each of which estimates reliability in a different way.

Reliability (continued) Inter-Rater or Inter-Observer Reliability Intra-Rater or Intra-Observer Reliability (Test-Retest Reliability) Internal Consistency Reliability Inter-method reliability.

Inter-Rater or Inter-Observer Reliability Used to assess the degree to which different raters/observers give consistent estimates of the same phenomenon (i.e. is the variation in measurements when taken by different persons but with the same method or instruments.

Establish reliability on pilot data or a subsample of data and retest often throughout. There are a number of statistics which can be used to determine inter-rater reliability. Cohen's kappa, intra-class correlation.

Test-Retest Reliability Used to assess the consistency of a measure from one time to another. This approach assumes that there is no substantial change in the construct being measured between the two occasions. The amount of time allowed between measures is critical. intra-rater reliability is the degree of agreement among multiple repetitions of a diagnostic test performed by a single rater

Internal Consistency Reliability Used to assess the consistency of results across items within a test. We are  looking at how consistent the results are for different items for the same construct within the measure. Example   "I like to eat bran bread“ “I've enjoyed eating bran bread in the past” “I hate bran bread” -

Kinds of Internal Reliability There are a number of statistics which can be used to determine Internal Reliability. Inter-item Correlation Cronbach's Alpha (a)

cronbach's alpha ----------------------Internal consistency α ≥ .9-----------------------------------------Excellent. 9 > α ≥ .8------------------------------------Good. 8 > α ≥ .7------------------------------------Acceptable. 7 > α ≥ .6---------------------Questionable and Acceptable. 6 > α ≥ .5-----------------------------------Poor. 5 > α----------------------------------------Unacceptable

Inter-method reliability- is the variation in measurements of the same target when taken by a different methods or instruments, but with the same person

Validity Validity is the degree of closeness of measurements of a quantity to that quantity's actual (true) value (i.e. Does it measure what you think it measures. This is more familiarly called Construct Validity.

Types of Construct Validity Translation validity Face validity Content validity Criterion-related validity (Known groups validity) Predictive validity Concurrent validity Convergent validity Discriminant validity

Translation validity

Face Validity “On its face" does it seems like a good translation of the construct. i.e. Does it appear to measure what it is supposed to measure? Weak Version: If you read it does it appear to ask questions directed at the concept. Strong Version: If experts in that domain assess it, they conclude it measures that domain.

Content Validity How well elements of the test relate to the content domain? For example, a depression measure should cover the checklist of depression symptoms

Criteria-Related Validity (Known groups validity) involves the correlation between the test and a criterion variable (or variables)

Predictive Validity A high correlation would provide evidence for predictive validity – i.e. it would show that our measure can correctly predict something that we theoretically thing it should be able to predict.

Concurrent Validity Assess the operationalization's ability to distinguish between groups that it should theoretically be able to distinguish between. As in any discriminating test, the results are more powerful if you are able to show that you can discriminate between two groups that are very similar.

Convergent Validity Examine the degree to which the operationalization is similar to (converges on) other operationalizations that it theoretically should be similar to. To show the convergent validity of a test of arithmetic skills, one might correlate the scores on a test with scores on other tests that purport to measure basic math ability, where high correlations would be evidence of convergent validity.

Discriminant Validity Examine the degree to which the operationalization is not similar to (diverges from) other operationalizations that it theoretically should be not be similar to. To show the discriminant validity of a test of arithmetic skills, we might correlate the scores on a test with scores on tests that of verbal ability, where low correlations would be evidence of discriminant validity.

Conclusion validity

Differentiate between prevalence and incidence Prevalence: The total number of cases of a disease in a given population at a specific time Incidence:  the number of new cases of a specific disease occurring during a certain period in a population at risk.

odds ratio odds ratio: the ratio, used particularly in case-control studies or cohort study (exposed versus non-exposed), estimates the chances of a particular event occurring in one population in relation to its rate of occurrence in another population. Odds ratio= a/b / c/d Odds ratio= a*d/c*b a b a/b c d c/d

a b a/b c d c/d Relative risk is a ratio of the probability of the event occurring in the exposed group versus a non-exposed group Here, a = 20, b = 80, c = 1, and d = 99. Then the relative risk of cancer associated with smoking would be Smokers would be twenty times as likely as non-smokers to develop lung cancer.

Smokers are ???? times more likely to have lung cancer than non-smokers