Presentation is loading. Please wait.

Presentation is loading. Please wait.

Foundations of Evidence-Based Outcome Measurement.

Similar presentations


Presentation on theme: "Foundations of Evidence-Based Outcome Measurement."— Presentation transcript:

1 Foundations of Evidence-Based Outcome Measurement

2 Basic Question “What is the best way to measure this client’s problems so his or her progress can be monitored over time in a way that will result in the most favorable outcomes for this client?”

3 Measurement Systematic process that involves assigning labels (usually numbers), to characteristics of people, objects, or events, using explicit and consistent rules so, ideally, the labels accurately represent the characteristic measured

4 Measurement Plan Overall strategy used to measure a client’s outcomes, including the methods and instruments used, how to obtain the information, who can best provide the information, and when, where, and how often the information should be collected

5 Measurement Method Class of measurement procedures (e.g., standardized self-report scales)

6 Measurement Instrument Specific measurement tool (e.g., a specific standardized self-report scale measuring depression)

7 Measurement Errors Discrepancies between measured and actual (“true”) values of a variable Caused by flaws in the measurement process e.g., characteristics of clients or other respondents, measurement conditions, properties of measures)

8 Random Measurement Errors Discrepancies between measured and actual (“true”) values of a variable which are equally likely to be higher or lower than the actual values because they are caused by chance fluctuations in measurement Caused by flaws in the measurement process Tend to cancel each other out and average to zero but they increase the variability of measured values

9 Systematic Measurement Errors Discrepancies between measured and actual (“true”) values of a variable which are more likely to be higher or lower than the actual values of the variable Caused by flaws in the measurement process Lead to over- or under estimates of the actual values of a variable Also known as “bias” in measurement

10 Systematic and Random Measurement Errors

11 Correlation Originally “co-relation” Sir Francis Galton’s idea Born: 16 Feb 1822 in Sparkbrook, England Died: 17 Jan 1911 in Grayshott House, Haslemere, Surrey, England Charles Darwin’s cousin Revelation occurred about 1888 while he took cover during a rainstorm

12 Correlation Karl Pearson, Galton’s colleague, worked out the details about 1896 Born: 27 March 1857 in London, England Died: 27 April 1936 in London, England Coined the term “standard deviation”

13 Visualizing Correlations http://www.duxbury.com/authors/mcclella ndg/tiein/johnson/correlation.htm http://www.duxbury.com/authors/mcclella ndg/tiein/johnson/correlation.htm http://www.stat.uiuc.edu/courses/stat100/ java/GCApplet/GCAppletFrame.html http://www.stat.uiuc.edu/courses/stat100/ java/GCApplet/GCAppletFrame.html

14 Pearson Product-Moment Correlation (r) Indicates direction and magnitude of linear relationship Range from -1 to +1 0 indicates no linear relationship r 2 indicates the amount of variance in the DV accounted for by the IV

15 Fostering Challenges Assess foster parent applicants’ skills and abilities to manage some of the unique challenges of fostering Vignettes ask applicants what they would do if faced with common dilemmas that foster parents often experience

16 Reliability General term for the consistency of measurements, and unreliability means inconsistency caused by random measurement errors

17 Methods for Determining Reliability Test-retest Alternate form (not discussed) Internal consistency Interrater/interobserver

18 Test-retest Degree to which scores on a measure are consistent over time Independently measure the same people, with the same measure, under the same circumstances

19 Construct Complex concept (e.g., intelligence, well- being, depression) Inferred or derived from a set of interrelated attributes (e.g., behaviors, experiences, subjective states, attitudes) of people, objects, or events Typically embedded in a theory Oftentimes not directly observable but measured using multiple indicators

20 Internal Consistency Degree to which responses to a set of items on a standardized scale measure the same construct consistently Independently measure the same people, with a single multiple-item measure, under the same circumstances

21 Internal Consistency (cont’d) Coefficient alpha Statistic typically used to quantify the internal consistency reliability of a standardized scale Also known as “Cronbach’s alpha” and, when items are dichotomous, “KR-20”

22 Interrater/Interobserver Degree of consistency in ratings or observations across raters, observers, or judges Multiple observers independently observe the same people, using the same measure, under the same circumstances

23 Father (●) and mother ( ○ ) Report of Time with Family

24 Adequate Reliability?.90+, excellent.80-.89, good.70-.79, acceptable <.70 suspect

25 Measurement Validity General term for the degree to which accumulated evidence and theory support interpretations and uses of scores derived from a measure More important, but more difficult to determine than reliability

26 Methods for Determining Validity Face Content Criterion Concurrent and Predictive Construct Convergent, Discriminant, and Sensitivity to Change

27 Face Degree to which a measure of a construct or other variable appears to measure a given construct in the opinion of clients, other respondents, and other users of the measure

28 Content Degree to which questions, behaviors, or other types of content represent a given construct comprehensively (e.g., the full range of relevant content is represented, and irrelevant content is not) Outcome as conceptualized Irrelevant elements - Relevant elements Outcome as operationalized

29 Criterion Degree to which scores on a measure can predict performance or status on another measure that serves as a standard (i.e., the criterion, sometimes called a “gold standard”)

30 Concurrent-Criterion Degree to which scores on a measure can predict a contemporaneous criterion Usually concurrent validity evidence is collected when we want to replace an existing measure with a simpler, cheaper, or less invasive one

31 Predictive-Criterion Degree to which scores on a measure can predict a criterion measured at a future point in time Usually predictive validity evidence is collected when we want to use results from a measure (e.g., ACT or SAT scores), to find out what might happen in the future (e.g., successful graduation from college), in order to take some course of action in the present (e.g., admit a student to college)

32 Construct Degree to which scores on a measure can be interpreted as representing a given construct, as evidenced by theoretically predicted patterns of associations with: Measures of related variables Measures of unrelated variable Group differences Changes over time

33 Convergent Degree to which scores derived from a measure of a construct are correlated in the predicted way with other measures of the same or related constructs or variables Depression Anxiety Lost Work Days Negative Self Appraisal Stressors

34 Discriminant Degree to which scores derived from a measure of a construct are uncorrelated with, or otherwise distinct from, theoretically dissimilar or unrelated constructs or other variables Anxiety LiterarcyIntelligenceHeight

35 Sensitivity to Change Degree to which a measure detects genuine change in the variable measured

36 Unifying Concept Construct Validity FaceContentCriterionConvergentDiscriminant Sensitivity to change

37 Relationship Between Reliability and Validity Reliability prerequisite for validity Lower reliability leads to lower validity Validity implies reliability

38 Decisions, Decisions… Who, Where, When, How often to collect outcome data

39 Who Client Practitioner Relevant others Independent evaluators

40 Where and When Developmental psychology is “the science of the strange behavior of children in strange situations with strange adults for the briefest possible periods of time.” Bronfenbrenner, 1977, p. 513 Representative samples Subset of observations (e.g., people, situations, times) that has characteristics similar to observations of the population from which the sample was selected

41 Representative? Population Sample Population Sample Population Sample

42 How Often Regular, frequent, pre-designated intervals Often enough to detect significant changes in the problem, but not so often that it becomes problematic

43 Engage and Prepare Clients Be certain the client understands and accepts the value and purpose of monitoring progress Discuss confidentiality Present measures with confidence Don’t ask for information the client can’t provide

44 Engage and Prepare Clients (cont’d) Be sure the client is prepared Be careful how you respond to information Use the information that is collected

45 Practical and Contributes to Favorable Outcomes Reliability and validity are necessary, but not sufficient characteristics of outcome measures Cost, benefit, efficiency, and acceptability also are important Accurate Practical


Download ppt "Foundations of Evidence-Based Outcome Measurement."

Similar presentations


Ads by Google