Presentation is loading. Please wait.

Presentation is loading. Please wait.

Jessica Lambeth, MPT jessicamlambeth@yahoo.com The Basics: Evidence-Based Practice for Physical Therapists and Physical Therapist Assistants, Online Training.

Similar presentations


Presentation on theme: "Jessica Lambeth, MPT jessicamlambeth@yahoo.com The Basics: Evidence-Based Practice for Physical Therapists and Physical Therapist Assistants, Online Training."— Presentation transcript:

1 Jessica Lambeth, MPT jessicamlambeth@yahoo.com
The Basics: Evidence-Based Practice for Physical Therapists and Physical Therapist Assistants, Online Training Module Jessica Lambeth, MPT

2 Module Purpose The purpose of this online training module is to share the basics of evidence-based practice (EBP). This module focuses on general concepts of EBP, with clinical scenarios related to school-based physical therapy.

3 Module Purpose, continued
The main emphasis of this module is not to promote certain treatments or tests and measures or “tell you” what the recent literature reveals about pediatric physical therapy practices. As will be discussed in the coming slides, the “research answer” alone is not the correct answer. In addition, EBP emphasizes the necessity of learning to perform searches and evaluate the material independently, related to clinicians’ and patients’ current circumstances. While initially time consuming, the results are worthwhile.

4 Module Purpose, continued
For those without a recent EBP background or training, this module will not result in immediate efficiency in literature searches, statistical interpretation, etc. Hopefully it will, however, encourage more questions and motivation for EPB. It will also enable the NC DPI to focus educational sessions on meeting needs in the area of EBP and to gauge the current knowledge of school-based PTs and PTAs. Your feedback and comments will help to plan future learning opportunities and resources.

5 Pretest Please stop here and complete the pretest if you have not already done so. Please also keep track of your time, as directed in the instructions. Also, “I don’t know” answers have been included to ensure that we receive appropriate feedback. If you really do not know the answer, in the pretest or the posttest, please indicate that in your response. This will greatly assist in providing necessary learning opportunities in the future, as well as help us to evaluate the effectiveness of the online module format and material.

6 References: The content of this EBP module is largely based on curriculum from courses within the transitional Doctor of Physical Therapy program at Rocky Mountain University of Health Professions in Provo, Utah. (web address:

7 References, continued:
Much of the material within this online training module can be found in the following texts: Guyatt G, Rennie D. Users' Guides to the Medical Literature- Essentials of Evidence-Based Clinical Practice. Chicago: AMA Press; 2002. Jaeschke R, Singer J, Guyatt GH. Measurement of health status: ascertaining the minimal clinically important difference. Control Clin Trials 1989;10: Portney LG, Watkins MP. Foundations of Clinical Research: Applications to Practice. 2nd ed. Upper Saddle River, NJ: Prentice-Hall Inc, 2000. Rothstein JM. Autonomous Practice or Autonomous Ignorance? (Editorial) Physical Therapy 81(10), October 2001.

8 References, continued:
Much of the material within this online training module can be found in the following texts: Straus SE, Richardson WS, Glasziou P, Haynes RB. Evidence-based Medicine: How to Practice and Teach EBM. 3rd ed. Edinburgh: Churchill Livingstone; 2005. All clipart came from Microsoft Office: Accessed 05/09/2009.

9 Module Outline 1) What is EBP? Slides 9-22
2) Statistics Review: Basic Research, Slides 3) How to Search for EBP, Slides 4) How to Interpret Research Related to EBP, Slides

10 Section 1 Outline What is Evidence-Based Practice?
Evidence-Based Practice: General Introduction, 10-12 Survey Results (NC school-based PTs view of EBP),13-18 Guiding Steps to Practice EBP,19-20 Two Fundamental Principles of EBP, 21 Best Research Evidence, 22

11 Evidence-Based Practice:
General Introduction EBP is the integration of the best research evidence, clinical expertise, and the patient’s values and circumstances Best Research Evidence: valid and clinically relevant research with a focus on patient-centered clinical research Clinical Expertise: use of clinical skills and experiences Patient’s Values and Circumstances: the patient’s unique preferences, concerns, and expectations in his or her setting (Straus et al, 2005)

12 Evidence-Based Practice:
General Introduction EBP is Not: Focused only on research studies Only to be used or understood by professionals who routinely participate in research studies A discouragement from trying new treatment  There may be little or no research on a particular topic, or studies with small sample sizes may have lacked the power to demonstrate statistical significance (as later explained in the statistics section)

13 Evidence-Based Practice:
General Introduction “Because RCTs are so difficult, we will always have areas that lack evidence, we will need to find other credible research approaches to supply evidence. Keep in mind that an absence of evidence is different from negative evidence. An absence of evidence is not an excuse to ignore the growing body of data available to guide practice.” (RCT = randomized clinical or controlled trials) (Rothstein 2001)

14 Survey Results Survey responses from North Carolina School-Based PTs/PTAs about EBP: Survey distributed at the NC Exceptional Children’s Conference, Nov 2008 41 Respondents ( for detailed results) 73% had participated in a conference or course on the use of EBP in the last 5 years Of those that participated in a course, respondent data suggests the course changed their view of EBP, but their use and practice of EBP was changed to a lesser degree.

15 Survey Results, Continued…
Highest frequency response as to why we should use EBP  positive impact on our clinical practice 39 of 41 respondents agreed that EBP is relevant and necessary for PTs in the school system (2 left that question blank)  Highest frequency responses as to why it is relevant and necessary: A focus on EBP results in improved clinical practice and provides validation and justification for our role as school-based PTs/PTAs

16 Survey Results, Continued…
When respondents were asked if they were comfortable searching for and using EBP: Yes: 17 (41%), No: 18 (44%), No Response: 6 (15%) Internet was the most frequently used source for searching EBP, but continuing education courses ranked highest for preference.

17 Survey Results, Continued…
A majority of the respondents were comfortable:  Determining the level of evidence and interpreting the authors’ conclusions A majority of the respondents were not comfortable:  Interpreting statistics, even though statistical knowledge is often helpful to evaluate conclusions drawn by the author

18 Survey Results, Continued…
The primary barriers to searching and using EBP, as listed by NC school PTs/PTAs: Time Access Uncertain how to search for EBP Factors that enhance the search and use of EBP, as listed by NC school PTs/PTAs listed: Additional time Education on the appropriate use of EBP Education in how to access EBP resources

19 Survey Results, Continued…
Questions generated from the survey: How to efficiently and effectively increase the knowledge and practice of EBP by NC school PTs/PTAs? How to address barriers to using and searching for EBP? How to enhance use and access of EBP in school-based practice?

20 Guiding Steps to Practice EBP
Analyze what we know and what we do not know, in relation to improving our clinical practice. Form answerable questions to address any gaps in our knowledge. Search for and find the best research evidence to address our questions. Critically appraise the information, based on its validity, impact or size of effect, and applicability. (Straus et al, 2005)

21 Guiding Steps to Practice EBP
Integrate information gathered from the best research evidence with clinical expertise and the patient’s values and circumstances Evaluate the effectiveness of any intervention taken based on steps 1-4, and the effectiveness and efficiency of the process (Straus et al, 2005)

22 Two Fundamental Principles of EBP
“Evidence alone is never sufficient to make a clinical decision” (page 8) Consider risks and benefits, costs, inconvenience, alternative treatment strategies, patient preferences/values and circumstances. “EBM posits a hierarchy of evidence to guide clinical decision making” (page 8) Not all research is equal in terms of relevance and statistical support, however, that does not mean lower level evidence is not worthwhile.  (Guyatt and Rennie use the term Evidence-Based Medicine, EBM)  (Guyatt and Rennie, 2002)

23 Best Research Evidence
The three sections that follow will focus on the “best research evidence” branch of Straus’ three components of EBM. (Straus et al, 2005) Best research evidence is not more important than the other two branches; it is prominent in this module because knowledge concerning clinical expertise and patient values and expectations will vary from situation to situation. Section 2 (Statistics Review, Basic Research) will provide the background information necessary to perform effective searches and interpret the best research evidence (Sections 3 and 4).

24 Section 2 Outline Statistics Review, Basic Research
Types of Research, 25-33 Hierarchy of Evidence, 34-35 Variables, 36-44 Measurement Validity Types, 45-48 Statistics, 49-67 Sensitivity and Specificity Positive and Negative Predictive Value Positive and Negative Likelihood Receiver Operating Characteristic (ROC) Curve Responsiveness to Change Effect Size versus MCID

25 Section 2 Outline, continued…
Statistics Review, Basic Research, continued Measurement Reliability, 68-72 Descriptive Statistics,73-83 Frequency and Shape of Distribution Central Tendency Measures Measures of Variability Inferential Statistics, Probability Sampling Error Confidence Intervals Hypothesis Testing Power

26 Types of Research Nonexperimental (Observational)
Descriptive or Exploratory No control or manipulation of variables Examines populations and relationships Experimental Researcher manipulates/controls variable(s) Comparison of interventions or groups, examines cause and effect (Portney and Watkins, 2000)

27 Types of Research Portney and Watkins suggest viewing various designs as a continuum of research with a descriptive, exploratory, or experimental purpose. Certain designs may include elements of more than one classification. (Portney and Watkins,2000)

28 Descriptive Research Examples:
Case Study: Description of one or more patients, may document unusual conditions or response to intervention Developmental Research: Examines patterns of growth and change, or documents natural history of a disease or condition Normative Research: Establishes typical values for specific variables (Portney and Watkins, 2000)

29 Descriptive Research, continued…
Examples: Qualitative Research: Collection of subjective, narrative information (rather than quantitative, numerical data) in an effort to better understand experiences Evaluation Research: Assessment of a program or policy, often by the collection of descriptive information through surveys or questionnaires (Portney and Watkins, 2000)

30 Exploratory Research Examples:
Correlational Methods: Examines relationships between areas of interest, may be used to predict or suggest, but cannot offer cause and effect Cohort and Case-Control Studies: Used often in epidemiological research to describe and predict risks for certain conditions Methodological Studies: Used to evaluate the validity and reliability of measuring instruments Historical Research: Use of archives or other records to reconstruct the past to generate questions or suggest relationships of historical interest (Portney and Watkins, 2000)

31 Experimental Research
Example Randomized Clinical or Controlled Trial (RCT): In general, a clinical treatment, or experimental condition, is compared to a control condition, often a placebo but in some cases an alternative treatment, where subjects are randomly assigned to a group. (Portney and Watkins, 2000)

32 Experimental Research, continued…
Examples: Single-Subject Design: Variation of RCT, study of an individual over time with repeated measurement and determined design phases (Portney and Watkins, 2000) In an N=1 RCT, a single individual receives alternating treatment and placebo or alternative treatment, with the patient and the assessor blinded to intervention allocation. Objective or subjective measures are then recorded during the allocation periods. (Guyatt and Rennie, 2002)

33 Experimental Research, continued…
Examples: Sequential Clinical Trial: Variation of RCT, technique that allows for the continuous analysis of data as it becomes available, does not require a fixed sample Quasi-Experimental Research: Comparative research in which subjects cannot be randomly assigned to a group, or control groups cannot be used. Lower level of evidence than RCTs. (Portney and Watkins 2000)

34 Experimental Research, continued
Examples: Systematic Review: Combination of several studies with the same or similar variables, in which the studies are summarized and analyzed (Guyatt and Rennie, 2002) Meta-analysis: Statistical combination of the data from several studies with the same or similar variables, to determine an overall outcome (Portney and Watkins, 2000; Guyatt and Rennie, 2002)

35 Hierarchy of Evidence for Treatment Decisions:
Greatest (Top) to Least (Bottom) N of 1 randomized controlled trial Systematic review of randomized trials* Single randomized trial Systematic review of observational studies addressing patient-important outcomes Single observational study addressing patient-important outcomes Physiological studies (studies of blood pressure, cardiac output, exercise capacity, bone density, and so forth) Unsystematic clinical observations *A meta-analysis is often considered higher than a systematic review (Guyatt and Rennie, 2002)

36 Hierarchy of Evidence Ideally, evidence from individual studies would be compiled or synthesized into systematic reviews, with that information succinctly consolidated into easily and quickly read synopses. All relevant information would be integrated and linked to a specific patient’s circumstance. The medical search literature is still far from this, but working towards that goal. Efforts include clinical prediction guidelines and APTA’s emphasis on EBP.   (Straus et al, 2005)

37 Variables Variables: Characteristic that can be manipulated or observed Types of Variables Independent or Dependent Measurement Scales/Levels Classification is useful for communication, so that readers are aware of the author’s hypothesis of what situation or intervention (independent variable) will predict or cause a given outcome (dependent variable) (Portney and Watkins, 2000)

38 Variables: Independent or Dependent
Independent Variable: A variable that is manipulated or controlled by the researcher, presumed to cause or determine another (dependent) variable Dependent Variable: A response variable that is assumed to depend on or be caused by another (independent) variable (Portney and Watkins, 2000)

39 Variables: Measurement Scales
Useful to convey information to the reader about the type of variables observed Necessary to determine what statistical analysis approach should be used to examine relationships between variables From lowest to highest level of measurement, the scales are nominal, ordinal, interval, and ratio (Portney and Watkins, 2000)

40 Variables: Measurement Scales
Nominal Scales (Classification Scale) Data, with no quantitative value, are organized into categories Categorizes are based on some criterion Categories are mutually exclusive and exhaustive (each piece of data will be assigned to only one category) Only permissible mathematical operation is counting (such as the number of items within each category) Examples: Gender, Blood Type, Side of Hemiplegic Involvement (Portney and Watkins, 2000)

41 Variables: Measurement Scales
Ordinal Scales Data are organized into categories, which are rank-ordered on the basis of a defined characteristic or property Categories exhibit a “greater than-less than” relationship with each other and intervals between categories may not be consistent and may not be known (Portney and Watkins, 2000)

42 Variables: Measurement Scales
Ordinal Scales, continued If categories are labeled with a numerical value, the number does not represent a quantity, but only a relative position within a distribution (for example, manual muscle test grades of 0-5) Not appropriate to use arithmetic operations Examples: Pain Scales, Reported Sensation, Military Rank, Amount of Assistance Required (Independent, Minimal…) (Portney and Watkins, 2000)

43 Variables: Measurement Scales
Interval Scales Data are organized into categories, which are rank-ordered with known and equal intervals between units of measurement Not related to a true zero Data can be added or subtracted, but actual quantities and ratios cannot be interpreted, due to lack of a true zero Examples: Intelligence testing scores, temperature in degrees centigrade or Fahrenheit, calendar years in AD or BC (Portney and Watkins, 2000)

44 Variables: Measurement Scales
Ratio Scales Interval score with an absolute zero point (so negative numbers are not possible) All mathematical and statistical operations are permissible Examples: time, distance, age, weight (Portney and Watkins, 2000)

45 Variables: Clinical Example
A study investigates how a strengthening program impacts a child’s ability to independently walk. In this case, the strengthening program is the independent variable and the ability to independently walk is the dependent variable. Amount of assistance required (if ranked maximal, moderate, minimal, independently, not based on weight put on a crutch or other quantitative testing) would be an example of ordinal data. Studies often have more than one independent or dependent variable

46 Measurement Validity Measurement Validity examines the “extent to which an instrument measures what it is intended to measure” (Portney and Watkins, 2000) For example, how accurate is a test or instrument at discriminating, evaluating, or predicting certain items?

47 Measurement Validity Validity of Diagnostic Tests
Based on the ability for a test to accurately determine the presence or absence of a condition Compare the test’s results to known results, such as a gold standard. For example, a test determining balance difficulties likely to result in falls could be compared against the number of falls an individual actually experiences within a certain time frame. A clinical test for a torn ACL could be compared against an MRI. (Portney and Watkins, 2000)

48 Measurement Validity: Types
Face Validity: Examines if an instrument appears to measure what it is supposed to measure (weakest form of measurement validity) Content Validity: Examines if the items within an instrument adequately comprise the entire content of a given domain reported to be measured by the instrument Construct Validity: Examines if an instrument can measure an abstract concept (Portney and Watkins, 2000)

49 Measurement Validity: Types
Criterion-related Validity: Examines if the outcomes of the instrument can be used as a substitute measure for an established gold standard test.  Concurrent Validity: Examination of Criterion-related Validity, when the instrument being examined and the gold standard are compared at the same time  Predictive Validity: Examination of Criterion-related Validity, when the outcome of the instrument being examined can be used to predict a future outcome determined by a gold standard (Portney and Watkins, 2000)

50 Measurement Validity: Statistics
Ways to Evaluate Usefulness of Clinical Screening or Diagnostic Tools Sensitivity and Specificity Positive and Negative Predictive Value Positive and Negative Likelihood Ratios Receiver Operating Characteristic (ROC) Curve The above mentioned statistical procedures are often used when researchers are introducing (and validating) the test. Hopefully the values from these operations can be found tool’s testing manual or in articles evaluating the tool’s validity within certain populations or settings.

51 Measurement Validity: Statistics
Diagnostic Reference Table Condition Present Absent Test Result Positive Negative True Positive a False Positive b False Negative c True Negative d (Guyatt and Rennie, 2002; Portney and Watkins, 2000; Straus et al, 2005)

52 Measurement Validity: Statistics
Table Labels: (a) True Positive: The subject matter has the condition, and the test accurately identifies the presence of the condition (d) True Negative: The subject matter does not have the condition, and the test accurately identifies the absence of the condition (b) False Positive: The subject matter does not have the condition, and the test incorrectly identifies the presence of the condition (c) False Negative: The subject matter has the condition, and the test incorrectly identifies the absence of the condition (Portney and Watkins, 2000)

53 Measurement Validity: Statistics
Positive test result = the test identifies the condition as being present; Negative result = the test identifies the condition as being absent (This may or may not be accurate when compared to the gold standard). The test’s sensitivity and specificity provide information about the accuracy of the test. (Portney and Watkins, 2000)

54 Measurement Validity: Statistics
Sensitivity The ability of the test to obtain a positive test when the condition is present; the ability to detect a true positive (a) a/(a + c) The proportion that test positive out of those with the condition (Portney and Watkins, 2000)

55 Measurement Validity: Statistics
Specificity The ability of the test to obtain a negative result when the condition is absent, the ability to detect a true negative (d) d/(b + d) The proportion that test negative out of those without the condition Sensitivity and Specificity are often provided in terms of percents, from 0% to 100% (low to high) (Portney and Watkins, 2000)

56 Measurement Validity: Statistics
Helpful Hints to remember Sensitivity and Specificity Sensitivity: SnNout: When a test has a high sensitivity (Sn), a negative result (N), rules out (out) the diagnosis. Specificity: SpPin: When a test has a high specificity (Sp), a positive result (P), rules in (in) the diagnosis (Straus et al, 2005)

57 Measurement Validity: Statistics
Clinical Example Example: You’re choosing between two tests that screen participation in school based on physical abilities. A positive result means that the student’s physical abilities impact his or her participation.

58 Measurement Validity: Statistics
Clinical Example One test has a high sensitivity, but a low specificity. A high sensitivity means that a negative test will effectively rule out students whose physical abilities do not impact participation. However, with a low specificity, there may be many false positives, meaning students may test “positive” for the condition when, in fact, their abilities do not impact participation.

59 Measurement Validity: Statistics
Clinical Example Example: You’re choosing between two tests that evaluate participation in school based on physical abilities. A positive result means that the student’s physical abilities impact his or her participation.

60 Measurement Validity: Statistics
Clinical Example The other test has a low sensitivity, but a high specificity. A high specificity means that a positive result will effectively rule in the condition. However, with a low sensitivity, there may be many false negatives, meaning that students may test “negative” for the condition, when, in fact, their abilities do impact their participation.

61 Measurement Validity: Statistics
Predictive Values Provided in terms of percentages, 0% to 100%, low to high Positive Predictive Value (PV+) Probability that a person with a positive test actually has the condition a/(a + b) High PV+ desired for screening tools, to prevent excessive unnecessary future testing Negative Predictive Value (PV-) Probability that a person with a negative test does not have the condition d/(c + d) (Portney and Watkins, 2000)

62 Measurement Validity: Statistics
Likelihood Ratios Calculated from the Diagnostic Reference Table Requires prior calculation of the pretest probability of the condition in question Easy to use when familiar with the concept, but requires the use of a probability guide chart or a nomogram (chart that contains pretest probability and likelihood ratios, with a ruler connecting those two points to determine posttest probabilities) (Guyatt and Rennie, 2002)

63 Measurement Validity: Statistics
Likelihood Ratios, continued Positive and negative likelihood ratios are calculated Determines the usefulness of a diagnostic test. If a positive or negative result will change the posttest probability of having a condition to alter the clinician and patient’s course of action, it will be useful. If the likelihood ratios of the test do not result in a substantial change of knowledge, the test most likely will not be useful. (Guyatt and Rennie, 2002)

64 Measurement Validity: Statistics
Receiver Operating Characteristic (ROC) Curve Uses sensitivity and specificity information to find the probability of correctly choosing between presence or absence of the condition For example, a test with an area under the ROC curve of 0.80, would result in the correct determination of presence or absence of a condition 80% of the time. (Portney and Watkins, 2000)

65 Measurement Validity: Statistics
Responsiveness to chance statistics evaluate a measurement tool’s ability to detect change over time For example, will administration of the test pre and post intervention reflect a change in status, if one actually occurred? Evaluated by examining the change in scores in a pretest-posttest design, or using effect size (Portney and Watkins, 2000)

66 Measurement Validity: Statistics
Effect Size Effect size (ES) is a measure of the amount of difference. For example, experimental group A increased their score on a coordination measure by an average of 15 points, while experimental group B increased their score an average of 8 points. The ES between groups would be 7 points, considering the groups were homogeneous at the start. (Portney and Watkins, 2000)

67 Measurement Validity: Statistics
Effect Size, continued… An effect size index is a converted effect size, a standardized measure of change, so that change scores across different outcome measures can be compared. ES is often displayed as a correlation coefficient, r Portney and Watkins note that considerations of ES vary based on the interpreting clinician, but review Cohen’s suggestions of scores <0.4 as small (treatment had a small effect), 0.5 as moderate, and >0.8 as large (Portney and Watkins, 2000)

68 Measurement Validity: Statistics
Effect Size versus Minimal Clinically Important Difference In addition to numerical data revealed and statistical significance, the clinician should also consider what amount of change is clinically meaningful, such as, how great a gain in strength or endurance will result in a change in function? This is often referred to as the minimal clinically important difference (MCID). (Jaeschke et al 1989)

69 Measurement Reliability: Statistics
Reliability examines a measurement’s consistency and freedom from error Can be thought of as reproducibility or dependability Estimate of how observed scores vary from the actual scores (Portney and Watkins, 2000)

70 Measurement Reliability: Statistics
Reliability Coefficient Ratio of reliability (many different types with various symbol representation) Range between 0.00 to 1.00; 0.00 = no reliability; 1.00 = perfect reliability Reflection of variance, a measure of the differences among scores within a sample (Portney and Watkins, 2000)

71 Measurement Reliability: Statistics
Correlation Comparison of the degree of association between two variables or sets of data Used as a basis for many reliability coefficients (Portney and Watkins, 2000)

72 Measurement Reliability: Statistics
Test-Retest Reliability Examines the consistency of the results of repeated test administration Traditional Analysis Pearson product-moment coefficient of correlation (for interval or ratio data) Spearman rho (for ordinal data) Current, sometimes considered preferred, Analysis Intraclass correlation coefficient (Portney and Watkins, 2000)

73 Measurement Reliability: Statistics
Rater Reliability Intrarater reliability Reliability of data collection from one individual over two or more trials Interrater reliability Reliability of data collection between two or more raters (Portney and Watkins, 2000)

74 Descriptive Statistics
Descriptive statistics are used to describe sample characteristics. A sample is a subset of a population chosen for study. Since data often cannot be collected from an entire population, the data chosen from a selected sample is intended to be representative or an estimate of the population data. Distribution: Total set of scores (from a sample) for a particular variable, given the symbol, n (Portney and Watkins, 2000)

75 Descriptive Statistics
Frequency and Shapes of Distribution Frequency distribution: The number of times each value, from the variable data, occurred Drawing a graph of frequency distributions can result in shapes that characterize the distributions Some graphs are asymmetrical, others are symmetrical A symmetrical graph with a bell-shaped distribution is referred to as a normal distribution. A skewed distribution presents asymmetrically (Portney and Watkins, 2000)

76 Descriptive Statistics
Normal Distributions, when graphed according to frequency, present in the shape of a bell with the majority of scores falling in the middle and progressively fewer scores at either end. It has special properties in statistics. (Portney and Watkins, 2000)

77 Descriptive Statistics: Central Tendency Measures
Used to quantitatively summarize a group’s characteristics. Mode: The score that occurs most frequently Median: The middle score in numerically ordered group of data. If there are an even number of scores, the median is the midpoint between the two middle scores Mean: The sum of a set of scores divided by the number of scores, n. Often referred to as the “average” of a data set. (Portney and Watkins, 2000)

78 Descriptive Statistics: Measures of Variability
Variability is the dispersion of scores. Variability is affected (qualified) by five characteristics: Range Percentiles Variance Standard deviation Coefficient of variation  (Portney and Watkins, 2000)

79 Descriptive Statistics: Measures of Variability
Variability, continued… Range: Difference between the highest and lowest scores in a distribution Percentiles: Used to describe a score’s position within a distribution, distribution data is often divided into quartiles, or four equal parts Variance: (Too in-depth to describe the statistical background for this module purpose). Reflects variation within a set of scores, in square units. Symbolized in sample data by s  (Portney and Watkins, 2000)

80 Descriptive Statistics: Measures of Variability
Variability, continued Standard Deviation: Representative of the variability of scores surrounding the mean, in original units of measurement. Square root of variance. The larger the standard deviation, the more spread out the variable’s scores are around a mean. For example, data set (A) 8,9,10,11,12 and the data set (B) 4,5,10,15,16 both have a mean of However the standard deviation for set A is 1.58 while the standard deviation of set B is 5.52. Symbolized in sample data by s2. (Portney and Watkins, 2000)

81 Descriptive Statistics: Measures of Variability
Variability, continued… Coefficient of Variation: The ratio of the standard deviation to the mean. (Portney and Watkins, 2000)

82 Descriptive Statistics:
Distributions Normal Versus Skewed (Non-normal) Distribution These theoretical shapes of distribution help determine what statistical formulas or measures should be used The characteristics of normally distributed data are constant and predictable. For statistical purposes, the normally distributed curve is often divided into proportional areas, each equal to one standard deviation. Data should be examined for “goodness-of-fit” to see if the sample approximates the normal distribution. (Portney and Watkins, 2000)

83 Descriptive Statistics:
Distributions Normal Distribution Statistics 1st standard deviation on either side of the average contains 34.13% of the data total of 68.62% of the data will be between +1 and -1 standard deviation Between 1st and 2nd standard deviation contains 13.59% of the data total of 95.45% of the data will be between +2 and -2 standard deviations [(13.59 times 2) + (34.13 times 2)] (Portney and Watkins, 2000)

84 Descriptive Statistics:
Distributions Normal Distribution Statistics, continued Between 2nd and 3rd standard deviation contains 2.14% of the data total of 99.73% of the data will be between +3 and -3 standard deviations [(13.59 times 2) + (34.13 times 2) + (2.14 times 2)] (Portney and Watkins, 2000)

85 Inferential Statistics
Estimate population characteristics from sample data Used often when testing theories about the effects of experimental treatments Requires that assumptions are made about how well the sample represents the larger population Assumptions are based on the statistical concepts of probability and sampling error It is important the sample be representative of the population, so that the results of interventions on samples can be applied to the entire population of individuals with those characteristics. (Portney and Watkins, 2000)

86 Inferential Statistics
Probability The likelihood that an event will occur, given all the possible outcomes. Used often in prediction. Given the symbol p Probability may range from p = 1 (certain the event will occur, 100% probability) to p = 0 (certain that the event will not occur, 0% probability) (Portney and Watkins, 2000)

87 Inferential Statistics
Probability Probability, continued… Reflective of should happen in the long run, not necessarily what will happen on a given trial. For example, if a treatment has a 60% chance of success, then 60% of people will likely be successfully treated. That does not mean the treatment will be 60% successful in an individual, likely it will either be a unsuccessful or successful for that individual. (Portney and Watkins, 2000)

88 Inferential Statistics: Probability
Clinical Example Probability statistics can be applied to the distribution of scores Example, for a normally distributed data set: Average long jump for a certain group of children is 35 inches with a standard deviation of 4 inches. Suppose you want to know the probability that a child will jump within one standard deviation (from 31 to 39 inches)? You know that within one standard deviation on either side of the mean is 68.2%, so that is the probability that a child will jump within one standard deviation of the mean.

89 Inferential Statistics: Probability
Clinical Example Example, continued… If you wanted to know the probability that a child will jump more than one standard deviation above the mean (greater than 39 inches), you can refer to the data and calculate 15.86%. Charts and graphs are available to calculate the data in between the standard deviations.

90 Inferential Statistics:
Sampling Error The difference between sample values and population values The lower the sampling error, the greater the confidence that the sample values are similar to the population values. To estimate sampling error, the standard error of the mean statistic is used (too complex to explain statistical basis in this format) The larger the sample size, n, the smaller the standard error of the mean (Portney and Watkins, 2000)

91 Inferential Statistics
Confidence Intervals Confidence Interval (CI) Range of scores with specific boundaries, or confidence limits, that should contain the population mean The boundaries of CIs are based on the sample mean and its standard error Degree of confidence is expressed as a percentage Often, researchers use 95% as a boundary, which is just slightly less than 2 standard deviations on either side of the mean (Portney and Watkins, 2000)

92 Inferential Statistics, Confidence Interval Clinical Example
A physical therapy treatment program resulted in the ability for 40 children with a certain disorder to walk an additional 8 independent steps, on average, within a certain set of parameters. The 95% CI for this data was ± 2 steps. Therefore, we can be 95% certain that the population mean, or average for all children with this disorder, is between 6 and 10 extra independent steps. Said another way, if there were an additional 100 children with the same condition, 95 of them would likely have an average that was between an additional 6 to 10 additional independent steps following the physical therapy treatment.

93 Inferential Statistics
Hypothesis Testing Used to decide if effects from an intervention are due to chance or the result of the intervention Results in one of two decisions: To accept or reject the null hypothesis (Portney and Watkins, 2000)

94 Inferential Statistics
Hypothesis Testing Hypothesis Testing, continued… Statistical Hypothesis (also known as the null hypothesis): Any observed difference (as in pretreatment to post-treatment or treatment compared to a placebo) is due to chance. When the null hypothesis is rejected, the researcher concludes that the effect of treatment is not likely due to chance. (Portney and Watkins, 2000)

95 Inferential Statistics
Hypothesis Testing Hypothesis Testing, continued… Alternative Hypothesis: Any observed difference is not due to chance. Often the researcher is trying to support the alternative hypothesis, as when trying to prove that one particular treatment is better. Sometimes, however, the researcher may be trying to prove that certain interventions are equal. (Portney and Watkins 2000)

96 Inferential Statistics Hypothesis Testing, Errors
Errors in Hypothesis Testing Decision to accept or reject the null hypothesis is based on the results of the statistical procedures on collected data from samples. Decisions are based on sample data, so it is possible that the results obtained are not accurate of population data. There is a chance for error, that the researcher may incorrectly accept or reject the null hypothesis (Portney and Watkins 2000)

97 Inferential Statistics Hypothesis Testing, Errors
Type I error (α): Rejecting the null hypothesis when it is true (for example, deciding that the difference seen between a treatment group and a control group is due to the effect of the treatment, when, in fact, the difference is due to chance). A commonly used standard is α= 0.05, or the researchers accept a 5% chance of making a Type I error Statistical tests completed with the sample data are used to calculate p, the probability that an observed difference did occur by chance. (Portney and Watkins, 2000)

98 Inferential Statistics Hypothesis Testing, p and α
Hypothesis Testing: Relationship between p and α If p is greater than the chosen α, then the researchers chose not to reject the null hypothesis For example, in a placebo versus treatment study, the researchers cannot conclude that the experimental treatment had a different effect then the placebo. (Portney and Watkins, 2000)

99 Inferential Statistics Hypothesis Testing, p and α
Hypothesis Testing: Relationship between p and α, continued… If p is less than the chosen α, then the researchers chose to reject the null hypothesis For example, in a placebo versus treatment study, the researchers conclude that the experimental treatment had a different effect then the placebo. (Portney and Watkins, 2000)

100 Inferential Statistics Hypothesis Testing, p and α
Hypothesis Testing: Relationship between p and α, continued… Confidence intervals surrounding the p value can be calculated, hopefully these are included in data analysis section of the research study. The CIs between two groups should not (rare exceptions) overlap, if a statistically significant difference is found. (Portney and Watkins, 2000)

101 Inferential Statistics Hypothesis, CIs versus MCID
Hypothesis Testing : CIs (slide 90) versus MCID (slide 67) When the null hypothesis is rejected, the researchers conclude that the experimental treatment (versus a placebo, for example) had a statistically significant effect. The clinician should examine the effect size and the MCID to ensure that the change is clinically and functionally relevant. (Guyatt and Rennie, 2002)

102 Inferential Statistics Hypothesis, CIs versus MCID
When a null hypothesis is not rejected, the researchers may conclude that the experimental treatment did not have an effect. However, the researchers should pay close attention to the confidence intervals. If the CI does not include the MCID, then the trial is most likely negative. If the CI includes the MCID, then the possibility that the experimental treatment may have a positive effect cannot be ruled out. The researchers may wish to run a power analysis, explained in later slides. (Guyatt and Rennie, 2002)

103 Inferential Statistics, Hypothesis Testing
Clinical Example Children with similar abilities/diagnosis (homogenous sample) are randomly assigned to two different groups: Group A receives a physical therapy designed to improve gross motor skills, and Group B completes typical daily activities (but don’t worry, due to ethical concerns children in Group B will receive the same treatment as those in Group A at the end of the study period).

104 Inferential Statistics, Hypothesis Testing
Clinical Example 1, continued… The outcome measure will be a tool that tests gross motor activities with a final, numerical outcome score. The authors hypothesis that the experimental group will show statistically significant gains (experimental hypothesis) and that the null hypothesis (there is no difference between groups at the end of the intervention period) will be rejected.

105 Inferential Statistics, Hypothesis Testing
Clinical Example 1, continued… Initially, group A and B have similar average pre-test scores (if not, that can be statistically corrected). Now suppose that Group A (experimental group, receiving additional PT services) increases an average of 9 points from pretest to posttest, with a CI of ±1, 8 to 10. Group B increases an average of 0.5 points, with a CI of ±2, -1.5 to 3.5.

106 Inferential Statistics, Hypothesis Testing
Clinical Example 1, continued… These confidence intervals do not overlap, which corresponds with the statistical analysis performed that p<0.05 (the predetermined α), and there is a statistically significant difference between groups. The null hypothesis is rejected.

107 Inferential Statistics, Hypothesis Testing
Clinical Example 1, continued… Prior to the experiment, the authors agreed that the MCID was 6. Both the statistically determined (p<0.05 ) and clinically determined criteria in support of the experimental group’s treatment were met (ES of experimental group,9, was greater than the MCID, 6).

108 Inferential Statistics, Hypothesis Testing Clinical Example 2
Initially, group A and B have similar average pre-test scores (if not, that can be statistically corrected). Now, change the scenario from previous slides. Suppose that Group A (experimental group, receiving additional PT services) increases an average of 5 points from pretest to posttest, with a CI of ±2, 3 to 7. Group B increases an average of 0.5 point,s CI ±2, -1.5 to 3.5.

109 Inferential Statistics, Hypothesis Testing
Clinical Example 2, continued… These confidence intervals overlap, and the statistical analysis performed revealed that p>0.05 (the predetermined α). There is not a statistically significant difference between groups. The null hypothesis is not rejected.

110 Inferential Statistics, Hypothesis Testing
Clinical Example 2, continued… However, prior to the experiment, the authors agreed that the MCID was 6. Since the CI range of Group A includes that MCID, the beneficial effects of the treatment from Group A over Group B cannot be ruled out. The researchers may want to run a statistical power analysis, to determine if the study was underpowered, and therefore unlikely to show a difference of the treatment.

111 Inferential Statistics Hypothesis Testing, Errors
Errors in Hypothesis Testing, Type II Type II error (β): Not rejecting the null hypothesis when it is false (for example, determining that differences are due to chance, when they are, in fact, due to the experimental treatment) (Portney and Watkins, 2000)

112 Inferential Statistics Hypothesis Testing, Errors
Errors in Hypothesis Testing, Type II The complement of Type II error is the statistical power of the test (1-β) Power: the probability that a test will lead to rejection of the null hypothesis, or the probability of obtaining statistical significance if the differences are not due to chance Many researchers use a standard of β=0.20, or a power of 80%, as reasonable protection against Type II error (Portney and Watkins,2000)

113 Inferential Statistics: Power
Determinants of Statistical Power Even though the researchers may not reject the null hypothesis, it does not always mean that an experimental treatment is not effective. The power of the study may have been too low or small to detect a significant difference. (Guyatt and Rennie 2002, Portney and Watkins, 2000)

114 Inferential Statistics: Power
Determinants of Statistical Power Levels set of α and β Variance Sample Size (n) Effect Size (Difference between two treatments or variables, Treatments with large changes or correlations are more likely to produce significant outcomes) Increases of effect size, sample size, and alpha all increase power, while decreases in variance increases power. (Portney and Watkins, 2000)

115 Inferential Statistics: Tests
Parametric Vs. Nonparametric Tests Parametric statistics are statistics used to estimate population parameters Primary assumptions of parametric tests: Samples randomly drawn from populations with normal distributions Variances in samples are roughly equal Data are measured on interval or ratio scales (Portney and Watkins, 2000)

116 Inferential Statistics
Tests Parametric Vs. Nonparametric Tests Nonparametric tests are generally not as powerful, and researchers may choose to use parametric tests, despite not meeting all generally held assumptions (such as use of parametric test in a study with ordinal data) (Portney and Watkins, 2000)

117 Formation of Clinical Questions used to search for EBP, 117-121
 Section 3 Outline How to Search for EBP Formation of Clinical Questions used to search for EBP, Evidence Search: Sources of Information, Internet/World Wide Web Textbooks Specific Journal Subscriptions Internet Sources for Medical Information The Guide Search Strategies,

118 Formation of Clinical Questions
Used to Search for EBP Background Questions: general knowledge about a condition/area of interest Foreground Questions: specific knowledge used to inform clinical decisions or actions. Often what researches use when investigating a particular treatment, prognosis, outcome measure, etc. Clinicians require the use of background and foreground knowledge, with proportions that may vary over time and depend on knowledge of a particular situation. (Straus et al, 2005)

119 Formation of Clinical Questions
Used to Search for EBP Background Questions Usually consist of two components: A question root (who, what, where, when, why, how) with a verb A condition, test, treatment, or other health care concern Example: “What is juvenile rheumatoid arthritis?” or “What causes cerebral palsy? (Straus et al, 2005)

120 Formation of Clinical Questions
Used to Search for EBP Foreground Questions Usually consist of four components (PICO) Patient and/or problem Intervention (or exposure) Comparison, if relevant Clinical Outcomes, including time if relevant (Straus et al, 2005)

121 Formation of Clinical Questions
Used to Search for EBP Foreground Questions, continued… Example, “For teenage girls with juvenile rheumatoid arthritis, is treatment X or treatment Y more effective in decreasing pain in joints of the hand and wrist?” or “For children with spastic diplegic cerebral palsy age 3-6 years old, which outcome measure (can compare A or B or search in general) most accurately assesses functional mobility?”

122 Formation of Clinical Questions
Used to Search for EBP Question Type: Categorization is useful for writing the question and statistical analysis Therapy: Evaluates the effects of various treatments or interventions Harm (not frequently investigate in PT literature): Evaluates the effects of various treatments or modalities on function, morbidity, mortality Diagnosis: Evaluates a tool or test’s ability to distinguish among certain conditions Prognosis: Evaluates the course of a certain condition (Guyatt and Rennie, 2002)

123 Evidence Search: Sources of Information
Where to find the research evidence, once the foreground question is developed? Internet/World Wide Web Textbooks Specific Journal Subscriptions Internet Sources for Medical Information

124 Evidence Search: Sources of Information
Internet/World Wide Web: This category is related to general search engines, not specific medical information journal search sources or health-related texts and journals available on the web. Provides a rapid and abundant source of information Be careful, not all sites and sources of information meet EBP guidelines (Guyatt and Rennie, 2002)

125 Evidence Search: Sources of Information
Internet/World Wide Web, continued… Consider the reputability of the source and the information found. Examples: MD Consult (fee), Google Scholar (free for the search) Some articles found may be free, others may have a fee. I personally have found Google Scholar very helpful. Its own search page is displayed when you enter “Google Scholar” into Google, (Guyatt and Rennie, 2002)

126 Evidence Search: Sources of Information
Textbooks: Often more useful for answering general background questions Books in print (hard-copy) may exclude valuable recent information, due to the time taken to compile, edit, and publish texts If utilizing a text, use one that is updated frequently and well referenced, so you can access more details if necessary (Guyatt and Rennie, 2002)

127 Evidence Search: Sources of Information
Textbooks, continued… Many texts that are internet-based are becoming more evidence-based and may be an increasing source of information for foreground questions in the future. Examples of internet-based texts (Guyatt and Rennie recommend UpToDate and Clinical Evidence as becoming more evidence-based): emedicine (free) UpToDate (fee) Clinical Evidence (fee) (Guyatt and Rennie, 2002)

128 Evidence Search: Sources of Information
Specific Journal Subscriptions: Browsing a specific journal may be beneficial if specific to your field of practice However, it mostly likely will leave out many applicable articles published in other journals Browsing through full text journals to find a specific article with appropriate quality and relevance often takes considerable time (Straus et al, 2005)

129 Evidence Search: Sources of Information
Commonly Used Internet Sources for Medical Information: Prefiltered Resources, Medline, PEDro, EBSCOhost/CINHAL These internet-accessible search engines allow you enter search terms to find applicable research, most often related to published journals. Most cites will provide an abstract (general summary) of any article found for free, but often a fee is charged for full text access unless you have access to university subscriptions. Many PT programs offer access to their resources for their Clinical Instructors.

130 Evidence Search: Sources of Information
Prefiltered Resources: Contain only those studies, most often systematic reviews, considered high methodological quality. Benefits: Easy to search small database with high quality studies. Drawbacks: not comprehensive Examples: Best Evidence, Cochrane Library (both require a subscription fee) (Guyatt and Rennie, 2002)

131 Evidence Search: Sources of Information
Medline: Online search engine with comprehensive coverage of medical journals, can be intimidating due to large size and knowledge necessary to perform an efficient search. Unfiltered, containing information from prefiltered sources (such as the Cochrane Library) plus additional expansive coverage from a wide variety of types of studies and journal publications (Guyatt and Rennie 2002)

132 Evidence Search: Sources of Information
Medline, continued … Medline, the search engine, not necessarily the articles within Medline, is available free to the public through PubMed ( Commercial vendor access (most often through university or health science center subscription): OVID, Knowledge Finder, Silver Platter Articles through PubMed Central, are free. (Guyatt and Rennie, 2002)

133 Evidence Search: Sources of Information
PEDro (Physiotherapy Evidence Database): Per the website, “[PEDro] has been developed to give rapid access to bibliographic details and abstracts of randomized controlled trials, systematic reviews and evidence-based clinical practice guidelines in physiotherapy.” ( Not-for-profit organization supported by many national physiotherapy associations, including the APTA

134 Evidence Search: Sources of Information
PEDro (Physiotherapy Evidence Database): Helpful Aspects of PEDro: systematic reviews are listed first, followed by articles ranked according to EBM hierarchy (with the scores explained based on the studies methodological characteristics) Free access at: no direct full-text articles. When applicable, the website provides potential links to full-text articles

135 Evidence Search: Sources of Information
EBSCOhost: Provider of CINAHL® (Cumulative Index to Nursing and Allied Health Literature) – CINAHL ® also available through several other sources Fee for use (Often as a university or health science subscription): Similar search strategies to other medical information sources

136 Evidence Search: Sources of Information
APTA’s Open Door: Available with APTA membership (login to look under areas of interest, research subheading) Access to research journal collections including ProQuest, Medline, Cochrane Library, CINAHL® (mostly contains bibliographic records, not full text journals)

137 Evidence Search: Sources of Information
APTA’s Open Door, continued… Suggestions, with links, for free full text (including BioMed Central, Directory of Open Access Journals, Public Library of Science, PubMed Central, and more) Tutorials and searching tips for EBP, along with guidelines for finding full-text articles Information on PT journals and samplings of current research in specific fields

138 Evidence Search: Sources of Information
APTA’s Hooked on Evidence: Available with APTA membership (login in to look under areas of interest, research subheading) Database of article extractions relevant to physical therapists: Peer reviewed information about the methodological quality and level evidence of the included articles

139 Evidence Search: Sources of Information
The Guide to Physical Therapist Practice (Text): Provides PTs with comprehensive descriptions of scope of practice Details preferred practice patterns Indications for specific tests and measures and interventions APTA. Guide to Physical Therapist Practice. 2nd ed. American Physical Therapy Association; 2001.

140 Search Strategies Searching for the Evidence within
Medical Information Sources: It is beyond the scope of this module to provide in-depth training on searching for literature. There are many online training courses, courses offered at health science center libraries, and thick packets of information on such searches.

141 Search Strategies Searching for the Evidence within Medical Information Sources, continued… It is not included in the time estimate or quiz information, but if interested in performing more searches and you would like greater instruction, consider a course or an online tutorial. Many on the internet are free, as the one by PubMed, Efficiency in searching requires time spent learning, experimenting, and practicing!

142 Search Strategies PubMed Search ( As a brief, non-inclusive overview, the following slides provide some basic information on searches. PubMed Database was chosen as an example since it is a free and comprehensive site. This is not intended to result in the ability to perform efficient searches, but rather to give an example, the very tip of the iceberg, of the variety of functions that can be performed on such a database.

143 Search Strategies PubMed Search:
The first step in searching is to form your foreground question and consider which information is most important to search Consider the specificity of your search. There is less information on pediatrics than adults, less information on rare disorders than common ones, etc. If you have too many results, try to be more specific. If too few, be more general.

144 Search Strategies PubMed Search
Go to the PubMed website ( and try entering a basic, general term of interest into the open field: Search, PubMed for (open field – enter term here) Consider the results of your search: Was there too little or too much information, was it relevant to what your interest, do any of the articles appear helpful to you?

145 Search Strategies PubMed Search
Limits: You can limit the responses in several categories. On the results page or on the search page, there is a row of tabs close to the top of the page. Click on “Limits,” the tab farthest to the left. This allows the searcher to limit the results in certain categories such as English language, for studies specifically on children, type of study, etc.

146 Search Strategies PubMed Search
Basic operators commonly included in search terms: * (truncation symbol) : Retrieves various spellings; Example: the term strength* will retrieve articles related to strength, strengthening, strengthen, etc OR : Expands retrieval of various topics; Example: Spina Bifida OR Meningocele will return articles listed under either heading

147 Search Strategies PubMed Search
Basic operators commonly included in search terms, continued… AND : Limits search; Example: Cerebral Palsy AND Strength* will only return articles that are related to both terms (for example, many of those retrieved articles will have a major or minor focus on strengthening programs in individuals with CP)

148 Search Strategies PubMed Search
Basic operators commonly included in search terms, continued… [tw] (text word) : Retrieves articles with a specific word in the title or abstract; Example: [tw]orthotic will retrieve articles with orthotic in the title or possibly the abstract. Be careful though, it might leave out articles with terms such as AFOs, DAFOs, orthoses, etc. That is when MeSH headings (slide 149) are useful.

149 Search Strategies PubMed Search
Basic operators commonly included in search terms, continued… There are many additional terms/commands that can be added to a basic search, which again is beyond the scope of this module. This is only to make you aware that terms can be added to designate a type of study (such as treatment, prognosis, diagnosis, etc), methodological qualities of the study (specificity, sensitivity, validity, reliability, responsiveness to change, etc), and many more.

150 Search Strategies PubMed Search, MeSH
MeSH (Medical Subject Heading) terms provide a way to search for a specific topic that may use different terminology for the same concept. Tutorials are available on PubMed. MeSH database: Try this search! On the main PubMed page, look at the blue column on the left. Under “PubMed Services” click on “MeSH Database”

151 Search Strategies PubMed Search, MeSH
At the top of the page, try a search in the open field. “Search MeSH for” (open field – enter topic of your choice) Checking on the numbered box with your subheading will allow the search to include all related “Entry Terms” listed below the topic. To search for this in PubMed, click on the tab towards the top of the page next to “Display” currently with “Summary” or “Full” in the field. Scroll down. “PubMed Links” will include a general, more comprehensive search, while “PubMed – Major Topic Links” will include articles more specific to your topic.

152 Search Strategies PubMed Search, Clinical Queries
Clinical Queries database: Try this search! On the main PubMed page, look at the blue column on the left. Under “PubMed Services” click on “Clinical Queries” Database. This page provides specialized searches for clinicians, the search can be performed for a specific topic, related to a type of study. For example, the searcher could investigate cerebral palsy, related to etiology, diagnosis, therapy, prognosis, or clinical prediction guides.

153 Search Strategies PubMed Search, Clinical Queries
Clinical Queries database, continued… It is also a place to search for systematic reviews (Don’t be disappointed if you cannot find very many for rare conditions, especially in pediatrics. More studies are being done, but there have not been as many performed yet in those fields) You can perform a narrow, specific search or a board sensitive search. Pick out a topic, category, and scope, and try one!

154 Section 4 Outline How to Interpret Research Related to EBP
Interpretation of Evidence, 154 For therapy/intervention, For diagnosis,

155 Interpretation of Evidence
Consider the type of study and where the study falls on the hierarchy of evidence. When evaluating the study, ask three questions: Are the Results Valid, What Are the Results, How Can I Apply the Results to Patient Care? Don’t forget to incorporate clinical expertise and patient values and preferences after evaluating any research material. As examples, the following slides will detail interpretation of evidence for therapy and diagnosis studies.   (Guyatt and Rennie, 2002)

156 Interpretation of Evidence For Therapy/Intervention: Validity
Are the results valid? (Slides 148 to 156) Did the experimental (or treatment) groups begin the study with a similar prognosis? If not, the end results may be due to differences in the groups versus differences due to intervention Were patients randomized into the groups? If the participants were not randomly assigned to a group, this decreases the validity of the study, as one group’s conditions may have impacted the outcomes. (Guyatt and Rennie, 2002)

157 Interpretation of Evidence For Therapy/Intervention: Validity
Was randomization concealed (blinded or masked)? If randomization was not concealed, this decreases the validity of the study. Those enrolling the participants may choose, unintentionally or intentionally, certain participants for certain groups if randomization is not concealed. (Guyatt and Rennie, 2002)

158 Interpretation of Evidence For Therapy/Intervention: Validity
Were patients analyzed in the groups to which they were randomized? It may be tempting to move patients between groups. For example, if a patient in an experimental group did not participate in the strengthening activity or take the medication prescribed for that group, it may be tempting to move that person to the control group. However, any cross-over or dropping out decreases the validity of the results.   (Guyatt and Rennie, 2002)

159 Interpretation of Evidence For Therapy/Intervention: Validity
Were patients analyzed in the groups to which they were randomized? Reasons for non-compliance or failure to finish treatment may be related to prognosis, and switching groups negates the unbiased comparison of randomization. Researchers are encouraged to follow an “intention-to-treat” analysis where all participants are analyzed in groups to which they were originally assigned. If drop-outs occurred, statistical methods are present to compensate. (Guyatt and Rennie, 2002)

160 Interpretation of Evidence For Therapy/Intervention: Validity
Were patients in the treatment group and control groups similar with respect to known prognostic factors? If not, this decreases the validity of the study, as groups’ prognostic differences may impact the outcome. Statistical testing can determine if groups are similar, and can also compensate for potential differences in final analyses. The magnitude of initial differences should be examined. (Guyatt and Rennie, 2002)

161 Interpretation of Evidence For Therapy/Intervention: Validity
Did experimental and control groups retain a similar prognosis after the study started? Were patients aware of group allocation? If so, this decreases the validity of the study. Often in PT research, it is difficult to blind a participant to the fact that they received a certain treatment. However, patients’ perceptions of received treatment may impact results and responses.  (Guyatt and Rennie, 2002)

162 Interpretation of Evidence For Therapy/Intervention: Validity
Were clinicians aware of group allocation? If so, this decreases the validity of the study. The researchers may treat patients in different groups differently, aside from the treatment, such as spending more time with a certain participants. This may bias the results. (Guyatt and Rennie, 2002)

163 Interpretation of Evidence For Therapy/Intervention: Validity
Were outcome assessors aware of group allocation? If so, this decreases the validity of the study. If the researcher knows the group assignment of the enrolled participant, it may unintentionally impact the results, especially if the outcome measures are more subjective. (Guyatt and Rennie, 2002)

164 Interpretation of Evidence For Therapy/Intervention: Validity
Was follow-up complete? If not, this decreases the validity of the study. When there is loss of follow-up, the researcher may consider a “worse-case” scenario, such as the experimental treatment that was supposed to improve function decreased function while those in the control group improved. Statistical analysis can determine how results may have been different. (Guyatt and Rennie, 2002)

165 Interpretation of Evidence For Therapy/Intervention: Results
What are the results? How large was the treatment effect? Usually involves examination of statistical analysis. Was there a statistically significant difference? How large was that difference or effect size? How precise was the estimate of the treatment effect? What do the statistics, such as confidence intervals and minimal clinically important differences, reveal about the treatment effect? (Refer to statistics knowledge for these questions) (Guyatt and Rennie, 2002)

166 Interpretation of Evidence For Therapy/Intervention: Application
How can I apply the results to patient care? Were the study patients similar to my patient? Would your patient have qualified to be included in the study? Can you extrapolate any useful information if the patient would not have qualified? Were all clinically important outcomes considered? Are the outcomes of interest areas of importance in your practice? (Guyatt and Rennie, 2002)

167 Interpretation of Evidence For Therapy/Intervention: Application
Are the likely treatment benefits worth the potential harm and cost? The clinician and patient would need to contemplate the impact of any potentially adverse effects, as well as the amount of time, effort, and additional costs of the treatment. (Guyatt and Rennie, 2002)

168 Interpretation of Evidence
For Diagnosis This information does not apply only to a medical diagnoses, it can also be used with physical therapy tests and measures that evaluate a multitude of conditions. For example, tools that evaluate: risk for falls, developmentally appropriate gross motor skills, sensory processing abilities impact on performance

169 Interpretation of Evidence For Diagnosis: Validity
Are the Results Valid? (slides ) Did clinicians face diagnostic uncertainty? If not, this decreases the validity of the results. The tool is only helpful if it can distinguish among those with uncertainty in diagnosis. For example, if testing a tool determining fall risk, it would not be beneficial to study the tool with young adults with no history of falls and no risks for falls. The tool would need to be investigated in individuals with the potential to be at risk for falls. (Guyatt and Rennie, 2002)

170 Interpretation of Evidence For Diagnosis: Validity
Was there a blind comparison with an independent gold standard applied similarly to the treatment group and the control group? If not, this decreases the validity of the results. When the comparison is not blinded, there is a potential for bias, especially if the tools contain more subjective material. The gold standard, also known as a reference standard, should be separate from the test being evaluated. (Guyatt and Rennie, 2002)

171 Interpretation of Evidence For Diagnosis: Validity
Was there a blind comparison with an independent gold standard applied similarly to the treatment group and the control group? continued… If two different test batteries are being compared, for example, they should not contain the same items. Finally, it is important to consider the reputability of the gold standard. In some cases, there might not be a universally accepted gold standard of comparison. The clinician should consider if the reference standard is acceptable. (Guyatt and Rennie, 2002)

172 Interpretation of Evidence For Diagnosis: Validity
Did the results of the test being evaluated influence the decision to perform the reference standard? If so, this decreases the validity of the results. The reference standard should be applied in every case, regardless of the results of the test being evaluated.  (Guyatt and Rennie, 2002)

173 Interpretation of Evidence For Diagnosis: Results
What are the results? Consider the statistical and clinical implications What likelihood ratios were associated with the range of possible test results? The researcher could also examine other related statistics, such as sensitivity and specificity, predictive values, likelihood ratios, and the ROC curve. The researcher can evaluate the psychometric properties of the diagnostic test: its validity, reliability, and responsiveness to change (as discussed in previous slides). (Guyatt and Rennie, 2002)

174 Interpretation of Evidence For Diagnosis: Application
How can I apply the results to patient care? Will the reproducibility of the test result and its interpretation be satisfactory in my clinical setting? It is important that the test produce reliable results that can be effectively interpreted. Are the results applicable to the patient in my practice? How similar are the patients in the study to your patient? (Guyatt and Rennie, 2002)

175 Interpretation of Evidence For Diagnosis: Application
How can I apply the results to patient care? Will the results change my management strategy? Will your treatment of the patient change based on the test results? Will patients be better off as a result of the test? What are the potential risks versus benefits of the test? (Guyatt and Rennie, 2002)

176 Thank You! Please mark your time of completion for this module and complete the posttest (same as the pretest, just print another copy). You are free also to look back to the slides to answer questions in the posttest. However, if you are unable to answer any question, even upon review of the PowerPoint module, “I don’t know” answers have been included to ensure that we receive appropriate feedback.

177 Thank You! Please also complete the comments section for this online module. This will greatly assist in providing necessary learning opportunities in the future, as well as help us to evaluate the effectiveness of the online module format and material. We appreciate your time and hope you found this training worthwhile.

178 Thank You! A special thanks to the following: Laurie Ray, MPT
Physical Therapy Consultant for NC DPI Project Mentor Jane Sweeney, PT, PhD, PCS, FAPTA Rocky Mountain University of Health Professions Graduate Program Consultant Jean Bryan Coe, PhD, DPT Project Consultant, Survey Data Stephen Allison, PT, PhD Project Consultant, Resources for Survey Data


Download ppt "Jessica Lambeth, MPT jessicamlambeth@yahoo.com The Basics: Evidence-Based Practice for Physical Therapists and Physical Therapist Assistants, Online Training."

Similar presentations


Ads by Google