Presentation is loading. Please wait.

Presentation is loading. Please wait.

Final Session of Summer: Psychometric Evaluation

Similar presentations


Presentation on theme: "Final Session of Summer: Psychometric Evaluation"— Presentation transcript:

1 Final Session of Summer: Psychometric Evaluation
Ron D. Hays, Ph.D. August 23, 2002 (9:30-11:30 am)

2 Four Types of Data Collection Errors
Coverage Error Does each person in population have an equal chance of selection? Sampling Error Are only some members of the population sampled? Nonresponse Error Do people in the sample who respond differ from those who do not? Measurement Error Are inaccurate answers given to survey questions?

3 What’s a Good Measure? Same person gets same score (reliability)
Different people get different scores (validity) People get scores you expect (validity) It is practical to use (feasibility)

4 How Are Good Measures Developed?
Review literature Expert input (patients and clinicians) Define constructs you are interested in Draft items (item generation) Pretest Cognitive interviews Field and pilot testing Revise and test again Translate/harmonize across languages

5 Scales of Measurement and Their Properties
Property of Numbers Type of Scale Equal Interval Rank Order Absolute 0 Nominal No No No Ordinal Yes No No Interval Yes Yes No Ratio Yes Yes Yes

6 Measurement Range for Health Outcome Measures
Nominal Ordinal Interval Ratio

7 Indicators of Acceptability
Response rate Administration time Missing data (item, scale)

8 Variability • All scale levels are represented
• Distribution approximates bell-shaped "normal"

9 Measurement Error observed = true score systematic error random error
+ + (bias)

10 Flavors of Reliability
• Inter-rater (rater) Equivalent forms (forms) • Internal consistency (items) • Test-retest (administrations)

11 Test-retest Reliability of MMPI 317-362 r = 0.75
True False 169 15 184 True MMPI 362 False 21 95 116 190 110 I am more sensitive than most other people.

12 Kappa Coefficient of Agreement (Corrects for Chance)
(observed - chance) kappa = (1 - chance) • Kappa can only reach 1.0 if the marginal distributions are equal • Adjusted kappa divides kappa by the maximum possible given the marginals kappa =

13 Example of Computing KAPPA
Rater A Row Sum 1 2 3 4 5 1 2 3 4 5 1 1 2 2 Rater B 2 2 2 3 2 10 Column Sum

14 Example of Computing KAPPA (Continued)
(1 x 2) + (3 x 2) + (2 x 2) + (2 x 2) + (2 x 2) P c = = 0.20 (10 x 10) 9 P obs. = = 0.90 10 Kappa = 0.87 =

15 Guidelines for Interpreting Kappa
Conclusion Kappa Conclusion Kappa Poor < Poor < 0.0 Fair Slight Good Fair Excellent > Moderate Substantial Almost perfect Fleiss (1981) Landis and Koch (1977)

16 Ratings of Height of Houseplants
Baseline Height Follow-up Height Experimental Condition Plant A1 R R A2 R R B1 R R B2 R R C1 R R

17 Ratings of Height of Houseplants (Cont.)
Baseline Height Follow-up Height Experimental Condition Plant C2 R R D1 R R D2 R R E1 R R E2 R R

18 Reliability of Baseline Houseplant Ratings
Ratings of Height of Plants: 10 plants, 2 raters Baseline Results Source DF SS MS F Plants Within Raters Raters x Plants Total

19 Sources of Variance in Baseline Houseplant Height
Source dfs MS Plants (N) (BMS) Within (WMS) Raters (K) (JMS) Raters x Plants (EMS) Total 19

20 Intraclass Correlation and Reliability
Model Reliability Intraclass Correlation One-Way MS MS MS MS MS MS (K-1)MS Two-Way MS MS MS MS Fixed MS MS (K-1)MS Two-Way N (MS MS ) MS MS Random NMS MS MS MS (K-1)MS K(MS MS )/N BMS WMS BMS WMS BMS BMS WMS BMS EMS BMS EMS BMS EMS EMS BMS EMS BMS EMS BMS JMS EMS BMS EMS JMS EMS

21 Summary of Reliability of Plant Ratings
Baseline Follow-up RTT RII RTT RII One-Way Anova Two-Way Random Effects Two-Way Fixed Effects Source Label Baseline MS Plants BMS Within WMS Raters JMS Raters X Plants EMS BMS - WMS BMS + (K - 1) * WMS BMS - EMS BMS + (K - 1) * EMS + K(JMS - EMS)/n BMS - EMS BMS + (K - 1) * EMS ICC (1,1) = ICC (2,1) = ICC (3,1) =

22 Cronbach’s Alpha Respondents (BMS) 4 11.6 2.9 Items (JMS) 1 0.1 0.1
Resp. x Items (EMS) Total Source SS MS df Alpha = = = 0.62 2.9 2.9

23 Alpha by Number of Items and Inter-item Correlations
_ K r alphast = _ 1 + r (K - 1 ) K = number of items in scale

24 Alpha for Different Numbers of Items and Homogeneity
Average Inter-item Correlation ( r ) Number of Items .0 .2 .4 .6 .8 1.0

25 Number of Items and Reliability for Three Versions of the Mental Health Inventory (MHI)

26 Spearman-Brown Prophecy Formula
( ) N • alpha x alpha = y 1 + (N - 1) * alpha x N = how much longer scale y is than scale x

27 Reliability Minimum Standards
For Group Comparisons – 0.70 or above For Individual Assessment – 0.90 or higher

28 Reliability of a Composite Score

29 Hypothetical Multitrait/Multi-Item Correlation Matrix

30 Multitrait/Multi-Item Correlation Matrix for Patient Satisfaction Ratings
Technical Interpersonal Communication Financial Technical * 0.63† 0.67† * 0.54† 0.50† * † * † * 0.60† 0.56† * 0.58† 0.57† 0.23 Interpersonal * 0.63† † 0.58* 0.61† † 0.65* 0.67† † 0.57* 0.60† * 0.58† † 0.48* 0.46† 0.24 Note – Standard error of correlation is Technical = satisfaction with technical quality. Interpersonal = satisfaction with the interpersonal aspects. Communication = satisfaction with communication. Financial = satisfaction with financial arrangements. *Item-scale correlations for hypothesized scales (corrected for item overlap). †Correlation within two standard errors of the correlation of the item with its hypothesized scale.

31 IRT

32 What are IRT Models? Mathematical equations that relate observed
survey responses to a persons location on an unobservable latent trait (i.e., intelligence, patient satisfaction).

33 Latent Trait and Item Responses
P(X1=1) P(X1=0) 1 Item 2 Response P(X2=1) P(X2=0) Item 3 Response P(X3=0) P(X3=2) 2 P(X3=1) Latent Trait In IRT, the observed variables are responses to specific items (e.g., 0, 1). The latent variable (trait) influences the probabilities of responses to the items (e.g, P(X1=1)). Notice that the number of response options varies across items. In IRT, the response formats do no need to be the same across items.

34 Types of IRT Models Dichotomous and polytomous Parameterization
Unidimensional and multidimensional Dichotomous and polytomous Parameterization One parameter: difficulty (location) Two Parameter: difficulty and slope (discrimination) Three Parameters: difficulty, slope, and guessing

35 1-Parameter Logistic Model for (Dichotomous Outcomes)

36 Item Characteristic Curves (1-Parameter Model)

37 2-Parameter Logistic Model (Dichotomous Outcomes)

38 Item Characteristic Curves (2-Parameter Model)

39 IRT Model Assumptions Unidimensionality
One construct measured by items in scale. Local Independence Items uncorrelated when latent trait(s) have been controlled for.

40 Item Responses and Trait Levels
Person 1 Person 2 Person 3 Trait Continuum

41 Information Conditional on Trait Level
Item information proportional to inverse of standard error: Scale/Test information is the sum over item information:

42 Item Information (2-parameter model)

43 Linking Item Content to Trait Estimates

44 Dichotomous Items Showing DIF (2-Parameter Model)
Hispanics Whites DIF – Location (Item 1) DIF – Slope (Item 2) Hispanics Whites

45 Forms of Validity Content Criterion Construct Validity

46 Construct Validity • Does measure relate to other measures in ways
consistent with hypotheses? Responsiveness to change

47 Relative Validity Analyses
Form of "known groups" validity Relative sensitivity of measure to important clinical difference One-way between group ANOVA

48 Relative Validity Example
Severity of Heart Disease Relative None Mild Severe F-ratio Validity Scale #1 Scale #2 Scale #3

49 Responsiveness to Change and Minimally Important Difference
• HRQOL measures should be responsive to interventions that changes HRQOL • Evaluating responsiveness requires assessment of HRQOL – pre-post intervention of known efficacy – at two times in tandem with gold standard

50 Two Essential Elements
External indicator of change (Anchors) mean change in HRQOL scores among people who have a “minimal” change in HRQOL. Amount of HRQOL change

51 External Indicator of Change (A)
Overall has there been any change in your asthma since the beginning of the study? Much improved; Moderately improved; Minimally improved No change Much worse; Moderately worse; Minimally worse

52 External Indicator of Change (B)
Rate your overall condition. This rating should encompass factors such as social activities, performance at work or school, seizures, alertness, and functional capacity; that is, your overall quality of life. 7 response categories; ranging from no impairment to extremely severe impairment

53 External Indicator of Change (C)
“changed” group = seizure free (100% reduction in seizure frequency) “unchanged” group = < 50% change in seizure frequency

54 Responsiveness Indices
(1) Effect size (ES) = D/SD (2) Standardized Response Mean (SRM) = D/SD† (3) Guyatt responsiveness statistic (RS) = D/SD‡ D = raw score change in “changed” group; SD = baseline SD; SD† = SD of D; SD‡ = SD of D among “unchanged”

55 Effect Size Benchmarks
Small: 0.20->0.49 Moderate: >0.79 Large: 0.80 or above

56 Treatment Impact on PCS

57 Treatment Impact on MCS

58 Treatment Impact on SF-12 MCS
Prozac for depressed (3 points) Erythropoetin for dialysis patients (5 points)

59


Download ppt "Final Session of Summer: Psychometric Evaluation"

Similar presentations


Ads by Google