Ch 6 Validity of Instrument

Slides:



Advertisements
Similar presentations
Cal State Northridge Psy 427 Andrew Ainsworth PhD
Advertisements

The Research Consumer Evaluates Measurement Reliability and Validity
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
VALIDITY AND RELIABILITY
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Chapter 4A Validity and Test Development. Basic Concepts of Validity Validity must be built into the test from the outset rather than being limited to.
CH. 9 MEASUREMENT: SCALING, RELIABILITY, VALIDITY
Measurement. Scales of Measurement Stanley S. Stevens’ Five Criteria for Four Scales Nominal Scales –1. numbers are assigned to objects according to rules.
Reliability and Validity of Research Instruments
Reliability, Validity, Trustworthiness If a research says it must be right, then it must be right,… right??
RESEARCH METHODS Lecture 18
Chapter 4 Validity.
VALIDITY.
MEASUREMENT. Measurement “If you can’t measure it, you can’t manage it.” Bob Donath, Consultant.
McGraw-Hill/Irwin Business Research Methods, 10eCopyright © 2008 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 11 Measurement.
Concept of Measurement
Beginning the Research Design
Validity Does test measure what it says it does? Is the test useful? Can a test be reliable, but not valid? Can a test be valid, but not reliable?
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Chapter 7 Evaluating What a Test Really Measures
Classroom Assessment A Practical Guide for Educators by Craig A
Rosnow, Beginning Behavioral Research, 5/e. Copyright 2005 by Prentice Hall Ch. 6: Reliability and Validity in Measurement and Research.
Understanding Validity for Teachers
Reliability, Validity, & Scaling
Measurement and Scaling
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
Reliability & Validity
Validity Is the Test Appropriate, Useful, and Meaningful?
6. Evaluation of measuring tools: validity Psychometrics. 2012/13. Group A (English)
Measurement Validity.
Research: Conceptualization and Measurement Conceptualization Steps in measuring a variable Operational definitions Confounding Criteria for measurement.
Research: Conceptualization and Measurement Conceptualization Steps in measuring a variable Operational definitions Confounding Criteria for measurement.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Chapter 4 Validity Robert J. Drummond and Karyn Dayle Jones Assessment Procedures for Counselors and Helping Professionals, 6 th edition Copyright ©2006.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Week 4 Slides. Conscientiousness was most highly voted for construct We will also give other measures – protestant work ethic and turnover intentions.
Ch 9 Internal and External Validity. Validity  The quality of the instruments used in the research study  Will the reader believe what they are readying.
12-1 Chapter 12 Measurement Learning Objectives Understand... distinction between measuring objects, properties, and indicants of properties similarities.
RESEARCH METHODS IN INDUSTRIAL PSYCHOLOGY & ORGANIZATION Pertemuan Matakuliah: D Sosiologi dan Psikologi Industri Tahun: Sep-2009.
Chapter 6 - Standardized Measurement and Assessment
Lesson 3 Measurement and Scaling. Case: “What is performance?” brandesign.co.za.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Ch. 5 Measurement Concepts.
Lecture 5 Validity and Reliability
QUESTIONNAIRE DESIGN AND VALIDATION
Reliability and Validity in Research
Concept of Test Validity
Evaluation of measuring tools: validity
Journalism 614: Reliability and Validity
Reliability & Validity
Week 3 Class Discussion.
پرسشنامه کارگاه.
PSY 614 Instructor: Emily Bullock Yowell, Ph.D.
Chapter Eight: Quantitative Methods
Analyzing Reliability and Validity in Outcomes Assessment Part 1
Measurement 變數測量 (構念的操作化)
第 10 章 測量 This chapter introduces the concept of measurement, the four scale types, the major sources of error and criteria for evaluating good measurement.
RESEARCH METHODS Lecture 18
Measurement 變數測量 (構念的操作化)
Chapter 11 Measurement This chapter introduces the concept of measurement, the four scale types, the major sources of error and criteria for evaluating.
Chapter 11 Measurement This chapter introduces the concept of measurement, the four scale types, the major sources of error and criteria for evaluating.
Analyzing Reliability and Validity in Outcomes Assessment
Cal State Northridge Psy 427 Andrew Ainsworth PhD
Presentation transcript:

Ch 6 Validity of Instrument

Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Validity Determinants Content Criterion Construct

There are three major forms of validity: content, construct, and criterion. Students need to know that they are in control of the level of validity of their own measurement. Content validity refers to the extent to which measurement scales provide adequate coverage of the investigative questions. If the instrument contains a representative sample of the universe of subject matter of interest, then content validity is good. To evaluate content validity, one must first agree on what elements constitute adequate coverage. To determine content validity, one may use one’s own judgment and the judgment of a panel of experts. Criterion-related validity reflects the success of measures used for prediction or estimation. There are two types of criterion-related validity: concurrent and predictive. These differ only on the time perspective. An attitude scale that correctly forecasts the outcome of a purchase decision has predictive validity. An observational method that correctly categorizes families by current income class has concurrent validity. Criterion validity is discussed further on the following slide.

Construct validity is a measurement scale that demonstrates both convergent validity and discriminant validity. In attempting to evaluate construct validity, one considers both the theory and measurement instrument being used. For instance, suppose we wanted to measure the effect of trust in relationship marketing. We would begin by correlating results obtained from our measure with those obtained from an established measure of trust. To the extent that the results were correlated, we would have indications of convergent validity. We could then correlate our results with the results of known measures of similar, but different measures such as empathy and reciprocity. To the extent that the results are not correlated, we can say we have shown discriminant validity.

Definition of Validity The appropriateness of the inferences or interpretations of an assessment's results

How should one select a criterion?

Validity Determinants Content Construct

Construct validity is a measurement scale that demonstrates both convergent validity and discriminant validity. In attempting to evaluate construct validity, one considers both the theory and measurement instrument being used. For instance, suppose we wanted to measure the effect of trust in relationship marketing. We would begin by correlating results obtained from our measure with those obtained from an established measure of trust. To the extent that the results were correlated, we would have indications of convergent validity. We could then correlate our results with the results of known measures of similar, but different measures such as empathy and reciprocity. To the extent that the results are not correlated, we can say we have shown discriminant validity. This example is expanded upon in the following slide.

Construct Related Validity Assessment scores are valid for making inferences regarding a psychological characteristic or theory of human behavior. Example: Matrix

Example: Matrix Content Journal Content 1 Content 2 Content 3 Article 1 Article 2

Case example: Marketing Contents Journal Price Product Promotion Kotler, P. (1997). Marketing management: Analysis, planning and control (9th ed. ). Englewood Cliff, NJ: Prentice-Hall. P 12 P 23 P 32 Louis, E. B., & David, L. K. (1974). Contemporary marketing. Hinsdale, IL: Dryden Press. P 61

Construct Validity is determined through: Intervention Studies --- Pre-test / post-test changes based on treatment Differential-Population Studies --- Determine how different populations perform on measure Related-Measures Studies --- Compares scores to other measures valid for measuring the same construct Literature Review

Increasing Construct Validity New measure of trust Known measure of trust Empathy Credibility

Again, for a measure to illustrate construct validity, it must have both convergent validity and discriminant validity. Continuing with the trust in relationship marketing example, to show convergent validity, we must show that our measure correlates with a known and trusted measure. To show discriminant validity, our measure must be able to discriminate from other similar, but different measures. This means we would begin by conducting a pilot test to gather data using both the new and known measures of trust and the similar variables such as empathy and credibility. We would then run a correlation analysis to determine if the measures are correlated. To the extent that the new and known measures of trust are correlated, we have demonstrated convergent validity. To the extent that the new trust measure and the measures of empathy and credibility are not correlated, we have shown discriminant validity.

What is a content domain?

Content Validity Does the instrument items represent the defined domain of content? 1. Samples the content domain 2. Elicits student behavior that demonstrates content domain

Determining Content Validity During assessment development 1. Define the content domain 2. Define the components of the content domain 3. Write questions that reflect the content domain 4. Determine the weight of the components of the content domain 5. Stratified sample from content domain

Determining Content Validity (Continued) After instrument construction 1. Panel of experts examinee assessment to determine if content domain is adequately represented 2. Have panel members individually review each test item 3. Calculate the percentage of the experts who agree to the appropriateness of each question for the content domain (content and relevance) 4. Compute an overall index that reflects the test’s content coverage

Increasing Content Validity Literature Search Etc. Expert Interviews Question Database Group Interviews

Content validity refers to the extent to which measurement scales provide adequate coverage of the investigative questions. If the instrument contains a representative sample of the universe of subject matter of interest, then content validity is good. To evaluate content validity, one must first agree on what elements constitute adequate coverage. To determine content validity, one may use one’s own judgment and the judgment of a panel of experts. Using the example of trust in relationship marketing, what would need to be included as measures of trust? Ask the students for their own ideas. To extend the questions included and to check for representativeness, students could check the literature on trust, conduct interviews with experts, conduct group interviews, check a database of questions, and so on.

Validity Determinants Content Criterion Construct

Criterion-related validity reflects the success of measures used for prediction or estimation. There are two types of criterion-related validity: concurrent and predictive. These differ only on the time perspective. An attitude scale that correctly forecasts the outcome of a purchase decision has predictive validity. An observational method that correctly categorizes families by current income class has concurrent validity. Criterion validity is discussed further on the following slide.

Judging Criterion Validity Relevance Criterion Freedom from bias Reliability Availability

The researcher must ensure that the validity criterion used is itself valid. Any criterion measure must be judged in terms of the four qualities named in the slide. A criterion is relevant if it is defined and scored in the terms the researchers judge to be the proper measures. Freedom from bias is attained when the criterion gives each unit of interest an opportunity to score well. A reliable criterion is stable or reproducible. Finally, the information specified by the criterion must be available.

Criterion-Related Validity Demonstrates that the test scores relate to a specific criteria of performance Two types: --- Predictive Validity --- Concurrent Validity

Predictive Criterion-Related Validity Assessment’s scores are valid for the prediction of future behavior. --- Subjects take assessment --- Researcher waits a significant amount of time, then collects data on the subjects performance on the specified criteria

Concurrent Criterion Related Validity Assessment’s scores are valid for indicating current behavior --- Subjects take assessment --- Subjects scores are compared to data reflecting their current proficiency of the specified criteria --- The scores from both measures are correlated to determine if the assessment agrees with the other measure.

Understanding Validity and Reliability

Exhibit 11-6 Exhibit 11-6 illustrates reliability and validity by using an archer’s bow and target as an analogy. High reliability means that repeated arrows shot from the same bow would hit the target in essentially the same place. If we had a bow with high validity as well, then every arrow would hit the bull’s eye. If reliability is low, arrows would be more scattered. High validity means that the bow would shoot true every time. It would not pull right or send an arrow careening into the woods. Arrows shot from a high-validity bow will be clustered around a central point even when they are dispersed by reduced reliability.