Unit IX: Validity and Reliability in nursing research

Slides:



Advertisements
Similar presentations
Standardized Scales.
Advertisements

Chapter 8 Flashcards.
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Topics: Quality of Measurements
The Research Consumer Evaluates Measurement Reliability and Validity
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
Chapter 4 – Reliability Observed Scores and True Scores Error
VALIDITY AND RELIABILITY
Part II Sigma Freud & Descriptive Statistics
Part II Sigma Freud & Descriptive Statistics
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Reliability and Validity of Research Instruments
RESEARCH METHODS Lecture 18
Concept of Measurement
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 5 Making Systematic Observations.
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Research Methods in MIS
Classroom Assessment A Practical Guide for Educators by Craig A
Measurement and Data Quality
Reliability, Validity, & Scaling
Instrumentation.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
WELNS 670: Wellness Research Design Chapter 5: Planning Your Research Design.
1 Chapter 4 – Reliability 1. Observed Scores and True Scores 2. Error 3. How We Deal with Sources of Error: A. Domain sampling – test items B. Time sampling.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
JS Mrunalini Lecturer RAKMHSU Data Collection Considerations: Validity, Reliability, Generalizability, and Ethics.
Psychometrics. Goals of statistics Describe what is happening now –DESCRIPTIVE STATISTICS Determine what is probably happening or what might happen in.
Criteria for selection of a data collection instrument. 1.Practicality of the instrument: -Concerns its cost and appropriateness for the study population.
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Measurement Experiment - effect of IV on DV. Independent Variable (2 or more levels) MANIPULATED a) situational - features in the environment b) task.
Chapter 6 - Standardized Measurement and Assessment
VALIDITY, RELIABILITY & PRACTICALITY Prof. Rosynella Cardozo Prof. Jonathan Magdalena.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
5. Evaluation of measuring tools: reliability Psychometrics. 2011/12. Group A (English)
RELIABILITY AND VALIDITY Dr. Rehab F. Gwada. Control of Measurement Reliabilityvalidity.
Measurement and Scaling Concepts
1 Measurement Error All systematic effects acting to bias recorded results: -- Unclear Questions -- Ambiguous Questions -- Unclear Instructions -- Socially-acceptable.
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 25 Critiquing Assessments Sherrilene Classen, Craig A. Velozo.
Understanding Populations & Samples
Survey Methodology Reliability and Validity
MGMT 588 Research Methods for Business Studies
Lecture 5 Validity and Reliability
Reliability and Validity
Reliability and Validity in Research
Concept of Test Validity
Assessment Theory and Models Part II
RELIABILITY OF QUANTITATIVE & QUALITATIVE RESEARCH TOOLS
CHAPTER 5 MEASUREMENT CONCEPTS © 2007 The McGraw-Hill Companies, Inc.
Understanding Results
Reliability & Validity
Introduction to Measurement
Human Resource Management By Dr. Debashish Sengupta
پرسشنامه کارگاه.
Reliability and Validity of Measurement
Analyzing Reliability and Validity in Outcomes Assessment Part 1
PSY 614 Instructor: Emily Bullock, Ph.D.
Evaluation of measuring tools: reliability
RESEARCH METHODS Lecture 18
Measurement Concepts and scale evaluation
Analyzing Reliability and Validity in Outcomes Assessment
Chapter 8 VALIDITY AND RELIABILITY
Presentation transcript:

Unit IX: Validity and Reliability in nursing research Prepared by: NUR 500 Research team 1ST semester 38/39. H

Scientific Rigor Function of Methods Used: To define the problems and develop an evidence based aims and hypothesis To measure the variables with attention to potential sources of measurement error and bias. To use and interpret statistical and other analysis precisely. To limit generalizations Polit & Beck 2017

OBJECTIVES On completing this Unit, you will be able to: Describe measurement error and its impact on the research process. Identify how item analysis can assist in maintaining rigor of measurement instruments. Describe reliability & validity strategies. Explain the concept of responsiveness and its associated techniques. Clarify sensitivity and specificity.

Item Analysis Not usually reported in the literature unless the study is one that seeking to establishing the psychometric properties of an instrument. An instrument that has to many items will artificially increase the scores used in reliability testing as well as respondent fatigue. Goal: To determine that each item is measuring the concept that it intents to measure. To delete items that are redundant or are measuring another concepts. Technique: Item to Total Correlation Items that score a correlation score >0.70 are redundant of other items on the scale Polit & Beck 2017

Measurement Error An measurement’s results varies as a result of accuracy in the measurement (true scores) and other factors (commonly called error) The degree of deviation between true scores and obtained scores when measuring a characteristic Observed Data = True Score + Error (XO = XT +_ XE) Too much error can lead to misleading conclusions about study findings Polit & Beck 2017

Error of Measurement Error score does not necessarily mean “wrong” score Error component is a composite of other factors that are also being measured by the researcher Error score includes other factors Example: Pain score (how much of the score is from “anxiety” around pain)? Polit & Beck 2017

Types of Error Random Error Inconsistent, random variation Error without qualification Cannot find a pattern in error Systematic Error Consistent error, not random Bias Has a consistent pattern Polit & Beck 2017

Sources of Measurement Error Situational Contaminants Was there disruptions during the measurement? Response-Set Bias Where a respondent always chooses the same answer or one that they believe the investigators want Transitory Personal Factors Such as a headache, or increased stress Polit & Beck 2017

Sources of Measurement Error Administration Variations Are score different at different times of the year? Instrument Clarity Was the instrument written at the correct literacy level? Response Sampling Was the convenience sample biased in some way such as all female? Polit & Beck 2017

Reliability: Data Collection • Degree of consistency with repeated measurements Categories of Reliability: Stability: Used with same people (patients) on separate occasions (over time) and get same answers Internal Consistency: All subparts/items are measuring same general thing Equivalence: Equitable results from two or more instruments or observers Polit & Beck 2017

Reliability: Stability Sometimes called test, re-test reliability Is the agreement of measuring instruments over time. To determine stability, a measure or test is repeated on the same subjects at a future date. Results are compared and correlated with the initial test to give a measure of stability. Typically the Spearman-Brown coefficient. Scores equal to or greater than 0.70 are usually considered sufficient. Polit & Beck 2017

Reliability: Internal Consistency This form of reliability is used to judge the consistency of results across items on the same test. Essentially, you are comparing test items that measure the same construct to determine the tests internal consistency. Statistical Techniques Split-Half: Items are divided into 2 sections, then a correlation between the two sections is determined. Cronbach’s Alpha: The average of all possible split half reliabilities for a set of items By convention, a lenient cut-off of .60 is common in exploratory research; alpha should be at least .70 or higher to retain an item in an "adequate" scale; and many researchers require a cut-off of .80 for a "good scale." Polit & Beck 2017

Reliability: Equivalence Equivalency reliability is the extent to which two observers or items measure identical concepts at an identical level of difficulty. Inter-Rater Reliability Are different observers using the same instrument measuring the same phenomena equivalent? A statistical measure of inter-rater reliability is Cohen’s Kappa Ranges from -1.0 to 1.0 where Large numbers mean better reliability, Values near zero suggest that agreement is attributable to chance, and Values less than zero signify that agreement is even less than that which could be attributed to chance. Instrumental Equivalency Are two (presumably parallel) instruments administered at about the same time equivalent? Polit & Beck 2017

Validity: Data Collection Degree to which a data collection instrument measures what it is supposed to be measuring. Validity isn’t determined by a single statistic, but by a body of research that demonstrates the relationship between the test and the behavior it is intended to measure. Polit & Beck 2017

Levels of Validity There are many types of validity: Face validity Content validity Construct validity Criterion-related validity Polit & Beck 2017

Validity: Face Validity Face validity is concerned with how a measure or procedure appears. Does it seem like a reasonable way to gain the information the researchers are attempting to obtain? Does it seem well designed? Does it seem as though it will work reliably? Unlike content validity, face validity does not depend on established theories for support Polit & Beck 2017

Validity: Content Validity Content validity evidence involves the degree to which the content of the test matches a content domain associated with the construct. Ask experts, based on judgment Adequacy of the “content” area Do my questions adequately get to the area of interest? Yes No Maybe Content Validity Index Statistical measure of agreement among the experts. Polit & Beck 2017

Validity: Construct Validity Validity of a test or a measurement tool that is established by demonstrating its ability to identify or measure the variables or constructs that it proposes to identify or measure. The judgment is based on the accumulation of statistical findings, usually correlations, from numerous studies using the instrument being evaluated. Polit & Beck 2017

Validity: Criterion-related Validity Most common strategy: Factor Analysis Types Exploratory Confirmatory How many factors did the analysis reveal? What variance was explained by all of the factors? Usually want to explain more than 70% of the variance Polit & Beck 2017

Validity: Criterion-related Validity A measure of how well one variable or set of variables predicts an outcome based on information from other variables. Criterion of comparison must be valid itself! Types Concurrent (Known Groups) A measurement/instrument is given to two divergent groups. If the measurement is valid, the scores should diverge. Predictive A measurements ability to predict scores on another measurement that is related or purports to measure the same or similar construct Polit & Beck 2017

Responsiveness Polit & Beck 2017 Controversial as to whether or not this is separate from or just one type of validity. Looking for the effect size change of the instrument across time. For some investigators, it is also an assessment of an instruments: Ceiling Effect A ceiling effect occurs when test items aren't challenging enough for a group of individuals. Thus, the test score will not increase for a subsample of people who may have clinically improved because they have already reached the highest score that can be achieved on that test. Floor Effect The floor effect is when data cannot take on a value lower than some particular number. Thus, it represents a subsample for whom clinical decline may not register as a change in score, even if there is worsening of function/behavior etc. Polit & Beck 2017

Reliability & Validity Not totally independent of each other An instrument that is not reliable cannot possibly be valid erratic, inconsistent, inaccurate However, an instrument can be reliable and not valid Polit & Beck 2017

Sensitivity & Specificity Assess the properties of a diagnostic instrument. Sensitivity and specificity describe how well the test discriminates between patients with and without disease. Sensitivity is the proportion of patients with disease who test positive. Specificity is the proportion of patients without disease who test negative Polit & Beck 2017

Reliability and Validity in Qualitative Studies Traditional Criteria for Judging Quantitative Research Alternative Criteria for Judging Qualitative Research Internal validity External validity Credibility Transferability Dependability Conformability Polit & Beck 2017

Thank you