RELIABILITY AND VALIDITY Dr. Rehab F. Gwada. Control of Measurement Reliabilityvalidity.

Slides:



Advertisements
Similar presentations
Questionnaire Development
Advertisements

Reliability and Validity
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Cal State Northridge Psy 427 Andrew Ainsworth PhD
Survey Methodology Reliability and Validity EPID 626 Lecture 12.
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
VALIDITY AND RELIABILITY
1Reliability Introduction to Communication Research School of Communication Studies James Madison University Dr. Michael Smilowitz.
Professor Gary Merlo Westfield State College
Research Methodology Lecture No : 11 (Goodness Of Measures)
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Reliability and Validity of Research Instruments
RESEARCH METHODS Lecture 18
Concept of Measurement
Beginning the Research Design
Psych 231: Research Methods in Psychology
Personality, 9e Jerry M. Burger
Research Methods in MIS
Chapter 7 Evaluating What a Test Really Measures
Reliability and Validity. Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? How do we judge.
Test Validity S-005. Validity of measurement Reliability refers to consistency –Are we getting something stable over time? –Internally consistent? Validity.
Measurement and Data Quality
Reliability, Validity, & Scaling
Variation, Validity, & Variables Lesson 3. Research Methods & Statistics n Integral relationship l Must consider both during planning n Research Methods.
Measurement in Exercise and Sport Psychology Research EPHE 348.
Instrumentation.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Copyright © 2008 by Nelson, a division of Thomson Canada Limited Chapter 11 Part 3 Measurement Concepts MEASUREMENT.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
Final Study Guide Research Design. Experimental Research.
Principles of Test Construction
The Basics of Experimentation Ch7 – Reliability and Validity.
Research: Conceptualization and Measurement Conceptualization Steps in measuring a variable Operational definitions Confounding Criteria for measurement.
Experiment Basics: Variables Psych 231: Research Methods in Psychology.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
Nurhayati, M.Pd Indraprasta University Jakarta.  Validity : Does it measure what it is supposed to measure?  Reliability: How the representative is.
Reliability and Validity Themes in Psychology. Reliability Reliability of measurement instrument: the extent to which it gives consistent measurements.
Validity and Reliability in Instrumentation : Research I: Basics Dr. Leonard February 24, 2010.
VALIDITY, RELIABILITY & PRACTICALITY Prof. Rosynella Cardozo Prof. Jonathan Magdalena.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
Measurement and Scaling Concepts
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 25 Critiquing Assessments Sherrilene Classen, Craig A. Velozo.
Some Terminology experiment vs. correlational study IV vs. DV descriptive vs. inferential statistics sample vs. population statistic vs. parameter H 0.
Survey Methodology Reliability and Validity
MGMT 588 Research Methods for Business Studies
Reliability and Validity
Ch. 5 Measurement Concepts.
Reliability and Validity
Reliability and Validity in Research
Concept of Test Validity
Test Validity.
Understanding Results
Human Resource Management By Dr. Debashish Sengupta
Week 3 Class Discussion.
پرسشنامه کارگاه.
5. Reliability and Validity
VALIDITY Ceren Çınar.
Unit IX: Validity and Reliability in nursing research
RESEARCH METHODS Lecture 18
Statistical Reasoning December 8, 2015 Chapter 6.2
Chapter 4: Designing Studies
Reliability and Validity
Presentation transcript:

RELIABILITY AND VALIDITY Dr. Rehab F. Gwada

Control of Measurement Reliabilityvalidity

Control of Measurement  In selecting a measuring instrument, the research is faced with two basic questions:  Does the instrument measure a variable consistently?  Is the instrument a true measure of the variable?  The first question seeks a reliability index while the second question raises the issue of validity.  The instrument must be reliable to be valid.

Reliability  Reliability(consistency)= the extent to which an instrument consistently measures what it is supposed to ReliabilityInter-rater Internal consistency Intra-rater reliability Parallel Forms Reliability

1-Intra-rater reliability  The degree to which same raters/observers give consistent estimates of the same measurements over time.  Stability and consistency over time as it is used to assess measure from one time to another.  The same groups, used the same measurement at two different times.  Consistency of patient/client results over time.  To monitor changes following treat.  A single examiner can replicate the results

2-Inter-rater Reliability The degree to which different raters/observers give consistent estimates of the same measurements.  Stability and consistency across raters/examiners.  There are many ways in which studies on inter-tester reliability can be performed:  1-Therapist and then a second therapist may measure the same patient  2-Therapists make their measurements and judgments simultaneously.

2-Inter-rater reliability  This is important because clients often move between therapy services, for example from an acute ward to a rehabilitation ward, from in-patient service to a day hospital or outpatient service, or from an intermediate care/rapid-response service to longer-term support by a community team. So a person might be given the same assessment on a number of occasions but each time a different therapist administers the test.

3-Parallel form reliability Parallel/Alternate Forms Method - refers to the administration of two alternate forms of the same measurement device and then comparing the scores.  Both forms are administered to the same person and the scores are correlated. If the two produce the same results, the instrument is considered reliable.  A good example is the SAT. There are two versions that measure Verbal and Math skills. Two forms for measuring Math should be highly correlated and that would document reliability.

3-Parallel form reliability  In parallel forms reliability you first have to create two parallel forms. One way to accomplish this is to create a large set of questions that address the same construct and then randomly divide the questions into two sets(instruments). You administer both instruments to the same sample of people.

4-Internal consistency reliability (IC)  In internal consistency reliability estimation we use our single measurement instrument administered to a group of people on one occasion to estimate reliability.  In effect we judge the reliability of the instrument by estimating how well the items that reflect the same construct yield similar result

4-Internal consistency reliability (IC)  Used to assess the degree of consistency of results across items within a test.  The degree of homogeneity among the items in a scale or measure The IC used to calculate the association: The higher (IC), the test is homogenous construct, The lower (IC), the test has heterogenic factors.

What is the Validity? Valid=faithful= true Validity: It is the measuring of what the instrument intended to measuring.

Relationship Between Reliability & Validity A measure that is not reliable cannot be valid  Thus, reliability is a necessary condition for validity A measure that is reliable is not necessarily valid  Thus, reliability is not a sufficient condition for validity

 In the first one, you are hitting the target consistently, but you are missing the centre of the target.  That is, you are consistently and systematically measuring the wrong value for all respondents.  In this case your measure is reliable, but not valid (that is, it's you are consistent but wrong!!). Reliable but not Valid

 The second, shows hits that are randomly spread across the target.  You seldom hit the centre of the target but, on average, you are getting the right answer for the group (but not very well for individuals).  In this case, you get a valid group estimate, but you are inconsistent.  Here, you can clearly see that reliability is directly related to the variability of your measure. Fairly Valid but not very Reliable

 The third scenario shows a case where your hits are spread across the target and you are consistently missing the centre.  In this case your measure is neither reliable nor valid. Neither Valid nor Reliable

 Finally, we see the "Robin Hood “ scenario -- you consistently hit the centre of the target.  In this case your measure is both reliable and valid. Valid & Reliable

Types of Measurement Validity ValidityConstruct Translation Criterion Validity

types of validity Criterion Validity  If an instrument has demonstrated a close relationship to another instrument (criterion) when measuring some known quantity or quality,the instrument is said valid.  The criterion is an instrument which is well established, accepted, or considered the best instrument of its kind. often called a ‘gold standard ‘.  For example Manual muscle test and the dynamometer

Types of Measurement Validity Translation validity contentFace

Translation validity Face validity  The simplest and least scientific definition of validity  It is demonstrated when a measure superficially appears to measure what it claims to measure.  Based on subjective judgment and difficult to quantify.  In simple terms, face validity is whether a test seems to measure what it is intended to measure (Asher, 1996).  Hydrostatic weighing, and lower extremity function scale

Translation validity  Content Validity The extent that, the test items actually represent the kinds of material (i.e., content) they are supposed to represent.  Law (1997) defines content validity as ‘ the comprehensiveness of an assessment and its inclusion of items that fully represent the attribute being measured.

Content Validity  Asher (1996) notes that content validity is descriptive rather than statistically determined. Because of this, it is sometimes considered a weaker form of validity compared to other types of validity.

Content Validity For example, if you want to measure balance and have 10 different items related to aspects of balance, you would need to examine each item separately to see if it really did relate to the domain of balance (Lewis and Bottomley, 1994). For example, lower extremity functional scale, one would expect a set of activities that cover all aspects of lower extremity function 1-Walking a mile. 2-Ability to climb stairs

types of validity Construct Validity  Refers to the extend to which a test is measuring the underlying theoretical constructs.  A test designed to measure depression must only measure that particular construct, not closely related ideas such as anxiety or stress.

Construct Validity Construct validation involves forming theories about the domain of interest and then assessing the extent to which the measure under investigation provides results that are consistent with the theories. For example: several theories applied to the construct validation of functional scale measure in low back pain

EXTERNAL VALIDITY  External validity is about generalization: To what extent can an effect in research, be generalized to populations, settings, treatment variables, and measurement variables?  External validity is usually split into two distinct types, population validity and ecological validity and they are both essential elements in judging the strength of an experimental design.

INTERNAL VALIDITY  The extent to which the results demonstrate that a causal relationship exists between the independent and dependent variables.  If – effect on dependant variable only due to variation in the independent variable(s), which need Controlling of extraneous variables.  then – internal validity achieved

Question?