Presentation is loading. Please wait.

Presentation is loading. Please wait.

© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.

Similar presentations


Presentation on theme: "© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight."— Presentation transcript:

1 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight

2 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight

3 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity Validity has been defined as referring to the appropriateness, correctness, meaningfulness, and usefulness of the specific inferences researchers make based on the data they collect. Validity has been defined as referring to the appropriateness, correctness, meaningfulness, and usefulness of the specific inferences researchers make based on the data they collect. It is the most important idea to consider when preparing or selecting an instrument. It is the most important idea to consider when preparing or selecting an instrument. Validation is the process of collecting and analyzing evidence to support such inferences. Validation is the process of collecting and analyzing evidence to support such inferences.

4 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Evidence of Validity There are 3 types of evidence a researcher might collect There are 3 types of evidence a researcher might collect Content-related evidence of validity Content-related evidence of validity Content and format of the instrument Content and format of the instrument Criterion-related evidence of validity Criterion-related evidence of validity Relationship between scores obtained using the instrument and scores obtained Relationship between scores obtained using the instrument and scores obtained Construct-related evidence of validity Construct-related evidence of validity Psychological construct being measured by the instrument Psychological construct being measured by the instrument

5 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Illustration of Types of Evidence of Validity (Figure 8.1)

6 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Content-related Evidence A key element is the adequacy of the sampling of the domain it is supposed to represent. A key element is the adequacy of the sampling of the domain it is supposed to represent. The other aspect of content validation is the format of the instrument. The other aspect of content validation is the format of the instrument. Attempts to obtain evidence that the items measure what they are supposed to measure typify the process of content-related evidence. Attempts to obtain evidence that the items measure what they are supposed to measure typify the process of content-related evidence.

7 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Criterion-related Evidence A criterion is a second test presumed to measure the same variable. A criterion is a second test presumed to measure the same variable. There are two forms of criterion-related validity: There are two forms of criterion-related validity: 1) Predictive validity: time interval elapses between administering the instrument and obtaining criterion scores 2) Concurrent validity: instrument data and criterion data are gathered and compared at the same time A Correlation Coefficient (r) indicates the degree of relationship that exists between the scores of individuals obtained by two instruments. A Correlation Coefficient (r) indicates the degree of relationship that exists between the scores of individuals obtained by two instruments.

8 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Construct-related Evidence Considered the broadest of the three categories. Considered the broadest of the three categories. There is no single piece of evidence that satisfies construct-related validity. There is no single piece of evidence that satisfies construct-related validity. Researchers attempt to collect a variety of types of evidence, including both content-related and criterion-related evidence. Researchers attempt to collect a variety of types of evidence, including both content-related and criterion-related evidence. The more evidence researchers have from different sources, the more confident they become about the interpretation of the instrument. The more evidence researchers have from different sources, the more confident they become about the interpretation of the instrument.

9 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Reliability Refers to the consistency of scores or answers provided by an instrument. Refers to the consistency of scores or answers provided by an instrument. Scores obtained can be considered reliable but not valid. Scores obtained can be considered reliable but not valid. An instrument should be reliable and valid (Figure 8.2), depending on the context in which an instrument is used. An instrument should be reliable and valid (Figure 8.2), depending on the context in which an instrument is used.

10 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Reliability and Validity (Figure 8.2)

11 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Reliability of Measurement (Figure 8.3)

12 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Errors of Measurement Because errors of measurement are always present to some degree, variation in test scores are common. Because errors of measurement are always present to some degree, variation in test scores are common. This is due to: This is due to: Differences in motivation Differences in motivation Energy Energy Anxiety Anxiety Different testing situation Different testing situation

13 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Reliability Coefficient Expresses a relationship between scores of the same instrument at two different times or parts of the instrument. Expresses a relationship between scores of the same instrument at two different times or parts of the instrument. The 3 best known methods are: The 3 best known methods are: Test-retest Test-retest Equivalent forms method Equivalent forms method Internal consistency method Internal consistency method

14 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Test-Retest Method Involves administering the same test twice to the same group after a certain time interval has elapsed. Involves administering the same test twice to the same group after a certain time interval has elapsed. A reliability coefficient is calculated to indicate the relationship between the two sets of scores. A reliability coefficient is calculated to indicate the relationship between the two sets of scores. Reliability coefficients are affected by the lapse of time between the administrations of the test. Reliability coefficients are affected by the lapse of time between the administrations of the test. An appropriate time interval should be selected. An appropriate time interval should be selected. In Educational Research, scores collected over a two-month period is considered sufficient evidence of test-retest reliability. In Educational Research, scores collected over a two-month period is considered sufficient evidence of test-retest reliability.

15 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Equivalent-Forms Method Two different but equivalent (alternate or parallel) forms of an instrument are administered to the same group during the same time period. Two different but equivalent (alternate or parallel) forms of an instrument are administered to the same group during the same time period. A reliability coefficient is then calculated between the two sets of scores. A reliability coefficient is then calculated between the two sets of scores. It is possible to combine the test-retest and equivalent-forms methods by giving two different forms of testing with a time interval between the two administrations. It is possible to combine the test-retest and equivalent-forms methods by giving two different forms of testing with a time interval between the two administrations.

16 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Internal-Consistency Methods There are several internal-consistency methods that require only one administration of an instrument. There are several internal-consistency methods that require only one administration of an instrument. Split-half Procedure: involves scoring two halves of a test separately for each subject and calculating the correlation coefficient between the two scores. Split-half Procedure: involves scoring two halves of a test separately for each subject and calculating the correlation coefficient between the two scores. Kuder-Richardson Approaches: (KR20 and KR21) requires 3 pieces of information: Kuder-Richardson Approaches: (KR20 and KR21) requires 3 pieces of information: Number of items on the test Number of items on the test The mean The mean The standard deviation The standard deviation Considered the most frequent method for determining internal consistency Considered the most frequent method for determining internal consistency Alpha Coefficient: a general form of the KR20 used to calculate the reliability of items that are not scored right vs. wrong. Alpha Coefficient: a general form of the KR20 used to calculate the reliability of items that are not scored right vs. wrong.

17 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Standard Error of Measurement An index that shows the extent to which a measurement would vary under changed circumstances. An index that shows the extent to which a measurement would vary under changed circumstances. There are many possible standard errors for scores given. There are many possible standard errors for scores given. Also known as measurement error, a range of scores that show the amount of error which can be expected. (Appendix D) Also known as measurement error, a range of scores that show the amount of error which can be expected. (Appendix D)

18 © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Scoring Agreement Scoring agreement requires a demonstration that independent scorers can achieve satisfactory agreement in their scoring. Scoring agreement requires a demonstration that independent scorers can achieve satisfactory agreement in their scoring. Instruments that use direct observations are highly vulnerable to observer differences. Instruments that use direct observations are highly vulnerable to observer differences. What is desired is a correlation of at least.90 among scorers as an acceptable level of agreement. What is desired is a correlation of at least.90 among scorers as an acceptable level of agreement.


Download ppt "© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight."

Similar presentations


Ads by Google