Reliability and Validity

Slides:



Advertisements
Similar presentations
Agenda Levels of measurement Measurement reliability Measurement validity Some examples Need for Cognition Horn-honking.
Advertisements

Chapter 8 Flashcards.
Reliability and Validity
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Conceptualization and Measurement
Cal State Northridge Psy 427 Andrew Ainsworth PhD
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
Chapter 5 Measurement, Reliability and Validity.
Measurement Reliability and Validity
RESEARCH METHODS Lecture 18
Beginning the Research Design
Research Methods in MIS
Classroom Assessment A Practical Guide for Educators by Craig A
Reliability and Validity. Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? How do we judge.
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Measurement Concepts & Interpretation. Scores on tests can be interpreted: By comparing a client to a peer in the norm group to determine how different.
Technical Issues Two concerns Validity Reliability
Measurement and Data Quality
Reliability and Validity what is measured and how well.
Bryman: Social Research Methods, 4 th edition What is a concept? Concepts are: Building blocks of theory Labels that we give to elements of the social.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand.
Final Study Guide Research Design. Experimental Research.
The Basics of Experimentation Ch7 – Reliability and Validity.
Reliability & Validity
Validity Is the Test Appropriate, Useful, and Meaningful?
Measurement Validity.
CHAPTER OVERVIEW The Measurement Process Levels of Measurement Reliability and Validity: Why They Are Very, Very Important A Conceptual Definition of Reliability.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Validity: Introduction. Reliability and Validity Reliability Low High Validity Low High.
Measurement Theory in Marketing Research. Measurement What is measurement?  Assignment of numerals to objects to represent quantities of attributes Don’t.
Criteria for selection of a data collection instrument. 1.Practicality of the instrument: -Concerns its cost and appropriateness for the study population.
Validity and Reliability in Instrumentation : Research I: Basics Dr. Leonard February 24, 2010.
Measurement Experiment - effect of IV on DV. Independent Variable (2 or more levels) MANIPULATED a) situational - features in the environment b) task.
Chapter 6 - Standardized Measurement and Assessment
1 Announcement Movie topics up a couple of days –Discuss Chapter 4 on Feb. 4 th –[ch.3 is on central tendency: mean, median, mode]
Survey Design Class 02.  It is a true measure  Measurement Validity is the degree of fit between a construct and indicators of it.  It refers to how.
Dr. Jeffrey Oescher 27 January 2014 Technical Issues  Two technical issues  Validity  Reliability.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
© 2009 Pearson Prentice Hall, Salkind. Chapter 5 Measurement, Reliability and Validity.
RELIABILITY AND VALIDITY Dr. Rehab F. Gwada. Control of Measurement Reliabilityvalidity.
Measurement and Scaling Concepts
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Survey Methodology Reliability and Validity
MGMT 588 Research Methods for Business Studies
Chapter 2 Theoretical statement:
Ch. 5 Measurement Concepts.
VALIDITY by Barli Tambunan/
Lecture 5 Validity and Reliability
Reliability and Validity
MEASUREMENT: RELIABILITY AND VALIDITY
Assessment Theory and Models Part II
Survey Research.
Reliability & Validity
Test Validity.
CHAPTER 5 MEASUREMENT CONCEPTS © 2007 The McGraw-Hill Companies, Inc.
Tests and Measurements: Reliability
Journalism 614: Reliability and Validity
Human Resource Management By Dr. Debashish Sengupta
Week 3 Class Discussion.
پرسشنامه کارگاه.
Reliability and Validity of Measurement
RESEARCH METHODS Lecture 18
Social Research Methods
Reliability & Validity
Reliability & Validity
Methodology Week 5.
Cal State Northridge Psy 427 Andrew Ainsworth PhD
Presentation transcript:

Reliability and Validity

Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? Reliability – consistency of measurement Validity – confidence in measures and design

Reliability and Validity Reliability focuses on measurement Validity also extends to: Precision in the design of the study – ability to isolate causal agents while controlling other factors (Internal Validity) Ability to generalized from the unique and idiosyncratic settings, procedures and participants to other populations and conditions (External Validity)

Reliability Consistency of Measurement Estimates of Reliability Reproducibility over time Consistency between different coders/observers Consistency among multiple indicators Estimates of Reliability Statistical coefficients that tell use how consistently we measured something

Measurement Validity Are we really measuring concept we defined? Is it a valid way to measure the concept? Many different approaches to validation Judgmental as well as empirical aspects

Key to Reliability and Validity Concept explication Thorough meaning analysis Conceptual definition: Defining what a concept means Operational definition: Spelling out how we are going to measure concept

Four Aspects of Reliability: 1. Stability 2. Reproducibility 3. Homogeneity 4. Accuracy

1. Stability Consistency across time repeating a measure at a later time to examine the consistency Compare time 1 and time 2

2. Reproducibility Consistency between observers Equivalent application of measuring device Do observers reach the same conclusion? If we don’t get the same results, what are we measuring? Lack of reliability can compromise validity

3. Homogeneity Consistency between different measures of the same concept Different items used to tap a given concept show similar results – ex. open-ended and closed-ended questions

4. Accuracy Lack of mistakes in measurement Increased by clear, defined procedures Reduce complications that lead to errors Observers must have sufficient: Training Motivation Concentration

Increasing Reliability General: Training coders/interviewers/lab personnel More careful concept explication (definitions) Specification of procedures/rules Reduce subjectivity (room for interpretation) Survey measurement: Increase the number of items in scale Weeding out bad items from “item pool” Content analysis coding: Improve definition of content categories Eliminate bad coders

Indicators of Reliability Test-retest Make measurements more than once and see if they yield the same result Split-half If you have multiple measures of a concept, split items into two scales, which should then be correlated Cronbach’s Alpha or Mean Item-total Correlation

Reliability and Validity Reliability is a necessary condition for validity If it is not reliable it cannot be valid Reliability is NOT a sufficient condition for validity If it is reliable it may not necessarily be valid Example: Bathroom scale, old springs

Not Reliable or Valid

Reliable but not Valid

Reliable and Valid

Types of Validity 1. Face validity 2. Content validity 3. Pragmatic (criterion) validity A. Concurrent validity B. Predictive validity 4. Construct validity A. Testing of hypotheses B. Convergent validity C. Discriminant validity

Face Validity Subjective judgment of experts about: “what’s there” Do the measures make sense? Compare each item to conceptual definition Do it represent the concept in question? If not, it should be dropped Is the measure valid “on its face”

Content Validity Subjective judgment of experts about: “what is not there” Start with conceptual definition of each dimension: Is it represented by indicators at the operational level? Are some over or underrepresented? If current indicators are insufficient: develop and add more indicators Example--Civic Participation questions: Did you vote in the last election? Do you belong to any civic groups? Have you ever attended a city council meeting? What about “protest participation” or “online organizing”?

Pragmatic Validity Empirical evidence used to test validity Compare measure to other indicators 1. Concurrent validity Does a measure predict simultaneous criterion? Validating new measure by comparing to existing measure E.g., Does new intelligence test correlate with established test 2. Predictive validity Does a measure predict future criterion? E.g., SAT scores: Do they predict college GPA?

Construct Validity Encompasses other elements of validity Do measurements: A. Represent all dimensions of the concept B. Distinguish concept from other similar concepts Tied to meaning analysis of the concept Specifies the dimensions and indicators to be tested Assessing construct validity A. Testing hypotheses B. Convergent validity C. Discriminant validity

A. Testing Hypotheses When measurements are put into practice: Are hypotheses that are theoretically derived, supported by observations? If not, there is a problem with: A. Theory B. Research design (internal validity) C. Measurement (construct validity?) In seeking to examine construct validity: Examine theoretical linkages of the concept to others Must identify antecedent and consequences What leads to the concept? What are the effects of the concept?

B. Convergent Validity Measuring a concept with different methods If different methods yield the same results: than convergent validity is supported E.g., Survey items measuring Participation: Voting Donating to money to candidates Signing petitions Writing letters to the editor Civic group memberships Volunteer activities

C. Discriminant (Divergent) Validity Measuring a concept to discriminate that concept from other closely related concepts E.g., Measuring Maternalism and Paternalism as distinct concepts

Dimensions of Validity for Research Design Internal Validity of research design Validity of sampling, measurement, procedures External Given the research design, how valid are Inferences made from the conclusions Implications for real world

Internal and External Validity in Experimental Design Internal validity: Did the experimental treatment make a difference? Or is there an internal design flaw that invalidates the results? External validity: Are the results generalizable? Generalizable to: What populations? What situations? Without internal validity, there is no external validity 1