Reliability: Introduction. Reliability Session Definitions & Basic Concepts of Reliability Theoretical Approaches Empirical Assessments of Reliability.

Slides:



Advertisements
Similar presentations
Chapter 8 Flashcards.
Advertisements

Consistency in testing
Topics: Quality of Measurements
RELIABILITY Reliability refers to the consistency of a test or measurement. Reliability studies Test-retest reliability Equipment and/or procedures Intra-
Some (Simplified) Steps for Creating a Personality Questionnaire Generate an item pool Administer the items to a sample of people Assess the uni-dimensionality.
Reliability Definition: The stability or consistency of a test. Assumption: True score = obtained score +/- error Domain Sampling Model Item Domain Test.
The Department of Psychology
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 5 Reliability.
1Reliability Introduction to Communication Research School of Communication Studies James Madison University Dr. Michael Smilowitz.
 A description of the ways a research will observe and measure a variable, so called because it specifies the operations that will be taken into account.
Reliability Analysis. Overview of Reliability What is Reliability? Ways to Measure Reliability Interpreting Test-Retest and Parallel Forms Measuring and.
Methods for Estimating Reliability
Reliability: Internal Consistency By Lynn Woolever AED 615 October 23, 2006.
Testing 05 Reliability.
Part II Knowing How to Assess Chapter 5 Minimizing Error p115 Review of Appl 644 – Measurement Theory – Reliability – Validity Assessment is broader term.
Reliability n Consistent n Dependable n Replicable n Stable.
Reliability Analysis. Overview of Reliability What is Reliability? Ways to Measure Reliability Interpreting Test-Retest and Parallel Forms Measuring and.
Reliability n Consistent n Dependable n Replicable n Stable.
RELIABILITY consistency or reproducibility of a test score (or measurement)
Reliability and Validity
LECTURE 5 TRUE SCORE THEORY. True Score Theory OBJECTIVES: - know basic model, assumptions - know definition of reliability, relation to TST - be able.
Conny’s Office Hours will now be by APPOINTMENT ONLY. Please her at if you would like to meet with.
Measurement: Reliability and Validity For a measure to be useful, it must be both reliable and valid Reliable = consistent in producing the same results.
Session 3 Normal Distribution Scores Reliability.
Chapter 9 Hypothesis Testing II. Chapter Outline  Introduction  Hypothesis Testing with Sample Means (Large Samples)  Hypothesis Testing with Sample.
Research Methods in MIS
Validity and Reliability EAF 410 July 9, Validity b Degree to which evidence supports inferences made b Appropriate b Meaningful b Useful.
Reliability of Selection Measures. Reliability Defined The degree of dependability, consistency, or stability of scores on measures used in selection.
Classical Test Theory By ____________________. What is CCT?
Multivariate Methods EPSY 5245 Michael C. Rodriguez.
Reliability, Validity, & Scaling
MEASUREMENT MODELS. BASIC EQUATION x =  + e x = observed score  = true (latent) score: represents the score that would be obtained over many independent.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Data Analysis. Quantitative data: Reliability & Validity Reliability: the degree of consistency with which it measures the attribute it is supposed to.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Copyright © 2008 by Nelson, a division of Thomson Canada Limited Chapter 11 Part 3 Measurement Concepts MEASUREMENT.
Reliability Chapter 3. Classical Test Theory Every observed score is a combination of true score plus error. Obs. = T + E.
Reliability Chapter 3.  Every observed score is a combination of true score and error Obs. = T + E  Reliability = Classical Test Theory.
Reliability: Introduction. Reliability Session 1.Definitions & Basic Concepts of Reliability 2.Theoretical Approaches 3.Empirical Assessments of Reliability.
Creating Assessments AKA how to write a test. Creating Assessments All good assessments have three key features: All good assessments have three key features:
Reliability & Validity
Tests and Measurements Intersession 2006.
Reliability & Agreement DeShon Internal Consistency Reliability Parallel forms reliability Parallel forms reliability Split-Half reliability Split-Half.
Independent vs Dependent Variables PRESUMED CAUSE REFERRED TO AS INDEPENDENT VARIABLE (SMOKING). PRESUMED EFFECT IS DEPENDENT VARIABLE (LUNG CANCER). SEEK.
All Hands Meeting 2005 The Family of Reliability Coefficients Gregory G. Brown VASDHS/UCSD.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Class # 3 Copyright © 2005 by R. Halstead. All rights reserved.
Class 9 Dependent Variables, Instructions/Literature Review
Multitrait Scaling and IRT: Part I Ron D. Hays, Ph.D. Questionnaire Design and Testing.
Reliability n Consistent n Dependable n Replicable n Stable.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Reliability: Introduction. Reliability Session 1.Definitions & Basic Concepts of Reliability 2.Theoretical Approaches 3.Empirical Assessments of Reliability.
Measurement Experiment - effect of IV on DV. Independent Variable (2 or more levels) MANIPULATED a) situational - features in the environment b) task.
2. Main Test Theories: The Classical Test Theory (CTT) Psychometrics. 2011/12. Group A (English)
Multitrait Scaling and IRT: Part I Ron D. Hays, Ph.D. Questionnaire.
Measuring Research Variables
Lesson 2 Main Test Theories: The Classical Test Theory (CTT)
Lesson 5.1 Evaluation of the measurement instrument: reliability I.
5. Evaluation of measuring tools: reliability Psychometrics. 2011/12. Group A (English)
Chapter 2 Norms and Reliability. The essential objective of test standardization is to determine the distribution of raw scores in the norm group so that.
Classical Test Theory Psych DeShon. Big Picture To make good decisions, you must know how much error is in the data upon which the decisions are.
Measurement and Scaling Concepts
Class 9 Dependent Variables, Instructions/Literature Review Class 9 Dependent Variables, Instructions/Literature Review Chapters 13 Spring 2016.
Questions What are the sources of error in measurement?
Classical Test Theory Margaret Wu.
Reliability & Validity
Part II Knowing How to Assess Chapter 5 Minimizing Error
By ____________________
Review Questions III Compare and contrast the components of an individual score for a between-subject design (Completely Randomized Design) and a Randomized-Block.
Measurement Concepts and scale evaluation
Psy 425 Tests & Measurements
Presentation transcript:

Reliability: Introduction

Reliability Session Definitions & Basic Concepts of Reliability Theoretical Approaches Empirical Assessments of Reliability Interpreting Coefficients Test Refinement and Reliability

Conceptions of Reliability “He’s often late!” “My car won’t start!”   S.E.M.

Components of Measurement Measured Value = True Value + Systematic Error (Bias) + Random Error The usefulness of a measure depends on the ratio of the true value to any error variance that it produces

Classical Test Theory Random error vs. systematic, or bias Classical theory assumptions: –Error independent of score –Mean of errors = 0 –Observed score = true score + error –Random errors tend to cancel out if sufficient observations made

Internal Consistency Logic: problem of change over time Alternate forms Split-half Kuder-Richardson & alpha Internal consistency Item-total correlations Number of items & reduction in error term Spearman-Brown formula # items reliability

Sources of Variance: Which to Include in Estimating Reliability? Error Patients Observers Time Measurement instrument

Reliability Subject Variability Subject variability + Measurement Error Reliability = Subject Variability Subject Var. + Observer Variability + Meas’t Error or,

Generalizability Theory Separates sources of variability: –Observer inconsistency –Between observers –Subject change over time Quantifies these Helps to show how to optimize design (and administration) of test given these performance characteristics.

Reliability versus Sensitivity of a Measurement Metaphor of the combs

Statistics to use: ICC vs. Pearson r ICC = 1.0; r = 1.0 r = 1.0; ICC < 1.0

What is the Reliability when: Every student is rated “above average” Physician A rates every BP as 5 mm Hg higher than physician B The measure is applied to a different population The observers change The patients do, in reality, improve over time