Presentation is loading. Please wait.

Presentation is loading. Please wait.

Ch 6 Validity of Instrument

Similar presentations


Presentation on theme: "Ch 6 Validity of Instrument"— Presentation transcript:

1 Ch 6 Validity of Instrument

2 Outline of Validity Definition Types of Validity 1. Construct Validity
2. Criterion-Related Validity 3. Content Validity

3 Validity Determinants
Content Criterion Construct

4 There are three major forms of validity: content, construct, and criterion. Students need to know that they are in control of the level of validity of their own measurement. Content validity refers to the extent to which measurement scales provide adequate coverage of the investigative questions. If the instrument contains a representative sample of the universe of subject matter of interest, then content validity is good. To evaluate content validity, one must first agree on what elements constitute adequate coverage. To determine content validity, one may use one’s own judgment and the judgment of a panel of experts. Criterion-related validity reflects the success of measures used for prediction or estimation. There are two types of criterion-related validity: concurrent and predictive. These differ only on the time perspective. An attitude scale that correctly forecasts the outcome of a purchase decision has predictive validity. An observational method that correctly categorizes families by current income class has concurrent validity. Criterion validity is discussed further on the following slide.

5 Construct validity is a measurement scale that demonstrates both convergent validity and discriminant validity. In attempting to evaluate construct validity, one considers both the theory and measurement instrument being used. For instance, suppose we wanted to measure the effect of trust in relationship marketing. We would begin by correlating results obtained from our measure with those obtained from an established measure of trust. To the extent that the results were correlated, we would have indications of convergent validity. We could then correlate our results with the results of known measures of similar, but different measures such as empathy and reciprocity. To the extent that the results are not correlated, we can say we have shown discriminant validity.

6 Definition of Validity
The appropriateness of the inferences or interpretations of an assessment's results

7 How should one select a criterion?

8 Validity Determinants
Content Construct

9 Construct validity is a measurement scale that demonstrates both convergent validity and discriminant validity. In attempting to evaluate construct validity, one considers both the theory and measurement instrument being used. For instance, suppose we wanted to measure the effect of trust in relationship marketing. We would begin by correlating results obtained from our measure with those obtained from an established measure of trust. To the extent that the results were correlated, we would have indications of convergent validity. We could then correlate our results with the results of known measures of similar, but different measures such as empathy and reciprocity. To the extent that the results are not correlated, we can say we have shown discriminant validity. This example is expanded upon in the following slide.

10 Construct Related Validity
Assessment scores are valid for making inferences regarding a psychological characteristic or theory of human behavior. Example: Matrix

11 Example: Matrix Content Journal Content 1 Content 2 Content 3
Article 1 Article 2

12 Case example: Marketing
Contents Journal Price Product Promotion Kotler, P. (1997). Marketing management: Analysis, planning and control (9th ed. ). Englewood Cliff, NJ: Prentice-Hall. P 12 P 23 P 32 Louis, E. B., & David, L. K. (1974). Contemporary marketing. Hinsdale, IL: Dryden Press. P 61

13 Construct Validity is determined through:
Intervention Studies --- Pre-test / post-test changes based on treatment Differential-Population Studies --- Determine how different populations perform on measure Related-Measures Studies --- Compares scores to other measures valid for measuring the same construct Literature Review

14 Increasing Construct Validity
New measure of trust Known measure of trust Empathy Credibility

15 Again, for a measure to illustrate construct validity, it must have both convergent validity and discriminant validity. Continuing with the trust in relationship marketing example, to show convergent validity, we must show that our measure correlates with a known and trusted measure. To show discriminant validity, our measure must be able to discriminate from other similar, but different measures. This means we would begin by conducting a pilot test to gather data using both the new and known measures of trust and the similar variables such as empathy and credibility. We would then run a correlation analysis to determine if the measures are correlated. To the extent that the new and known measures of trust are correlated, we have demonstrated convergent validity. To the extent that the new trust measure and the measures of empathy and credibility are not correlated, we have shown discriminant validity.

16 What is a content domain?

17 Content Validity Does the instrument items represent the defined domain of content? 1. Samples the content domain 2. Elicits student behavior that demonstrates content domain

18 Determining Content Validity
During assessment development 1. Define the content domain 2. Define the components of the content domain 3. Write questions that reflect the content domain 4. Determine the weight of the components of the content domain 5. Stratified sample from content domain

19 Determining Content Validity (Continued)
After instrument construction 1. Panel of experts examinee assessment to determine if content domain is adequately represented 2. Have panel members individually review each test item 3. Calculate the percentage of the experts who agree to the appropriateness of each question for the content domain (content and relevance) 4. Compute an overall index that reflects the test’s content coverage

20 Increasing Content Validity
Literature Search Etc. Expert Interviews Question Database Group Interviews

21 Content validity refers to the extent to which measurement scales provide adequate coverage of the investigative questions. If the instrument contains a representative sample of the universe of subject matter of interest, then content validity is good. To evaluate content validity, one must first agree on what elements constitute adequate coverage. To determine content validity, one may use one’s own judgment and the judgment of a panel of experts. Using the example of trust in relationship marketing, what would need to be included as measures of trust? Ask the students for their own ideas. To extend the questions included and to check for representativeness, students could check the literature on trust, conduct interviews with experts, conduct group interviews, check a database of questions, and so on.

22 Validity Determinants
Content Criterion Construct

23 Criterion-related validity reflects the success of measures used for prediction or estimation.
There are two types of criterion-related validity: concurrent and predictive. These differ only on the time perspective. An attitude scale that correctly forecasts the outcome of a purchase decision has predictive validity. An observational method that correctly categorizes families by current income class has concurrent validity. Criterion validity is discussed further on the following slide.

24 Judging Criterion Validity
Relevance Criterion Freedom from bias Reliability Availability

25 The researcher must ensure that the validity criterion used is itself valid. Any criterion measure must be judged in terms of the four qualities named in the slide. A criterion is relevant if it is defined and scored in the terms the researchers judge to be the proper measures. Freedom from bias is attained when the criterion gives each unit of interest an opportunity to score well. A reliable criterion is stable or reproducible. Finally, the information specified by the criterion must be available.

26 Criterion-Related Validity
Demonstrates that the test scores relate to a specific criteria of performance Two types: --- Predictive Validity --- Concurrent Validity

27 Predictive Criterion-Related Validity
Assessment’s scores are valid for the prediction of future behavior. --- Subjects take assessment --- Researcher waits a significant amount of time, then collects data on the subjects performance on the specified criteria

28 Concurrent Criterion Related Validity
Assessment’s scores are valid for indicating current behavior --- Subjects take assessment --- Subjects scores are compared to data reflecting their current proficiency of the specified criteria --- The scores from both measures are correlated to determine if the assessment agrees with the other measure.

29 Understanding Validity and Reliability

30 Exhibit 11-6 Exhibit 11-6 illustrates reliability and validity by using an archer’s bow and target as an analogy. High reliability means that repeated arrows shot from the same bow would hit the target in essentially the same place. If we had a bow with high validity as well, then every arrow would hit the bull’s eye. If reliability is low, arrows would be more scattered. High validity means that the bow would shoot true every time. It would not pull right or send an arrow careening into the woods. Arrows shot from a high-validity bow will be clustered around a central point even when they are dispersed by reduced reliability.


Download ppt "Ch 6 Validity of Instrument"

Similar presentations


Ads by Google