Explaining Cronbach’s Alpha

Slides:



Advertisements
Similar presentations
Questionnaire Development
Advertisements

FACULTY DEVELOPMENT PROFESSIONAL SERIES OFFICE OF MEDICAL EDUCATION TULANE UNIVERSITY SCHOOL OF MEDICINE Using Statistics to Evaluate Multiple Choice.
Psychology Practical (Year 2) PS2001 Correlation and other topics.
MEASUREMENT: RELIABILITY Lu Ann Aday, Ph.D. The University of Texas School of Public Health.
Consistency in testing
Topics: Quality of Measurements
Reliability and Validity checks S-005. Checking on reliability of the data we collect  Compare over time (test-retest)  Item analysis  Internal consistency.
Reliability Definition: The stability or consistency of a test. Assumption: True score = obtained score +/- error Domain Sampling Model Item Domain Test.
© McGraw-Hill Higher Education. All rights reserved. Chapter 3 Reliability and Objectivity.
The Department of Psychology
Chapter 4 – Reliability Observed Scores and True Scores Error
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 5 Reliability.
1Reliability Introduction to Communication Research School of Communication Studies James Madison University Dr. Michael Smilowitz.
 A description of the ways a research will observe and measure a variable, so called because it specifies the operations that will be taken into account.
Reliability Analysis. Overview of Reliability What is Reliability? Ways to Measure Reliability Interpreting Test-Retest and Parallel Forms Measuring and.
Reliability for Teachers Kansas State Department of Education ASSESSMENT LITERACY PROJECT1 Reliability = Consistency.
Item Analysis: A Crash Course Lou Ann Cooper, PhD Master Educator Fellowship Program January 10, 2008.
Reliability, the Properties of Random Errors, and Composite Scores.
Statistics of EBO 2010 Examination EBO General Assembly Sunday June 21st, 2010 (Tallin, Estonia) Danny G.P. Mathysen MSc. Biomedical Sciences EBOD Assessment.
-生醫統計期末報告- Reliability 學生 : 劉佩昀 學號 : 授課老師 : 蔡章仁.
Reliability n Consistent n Dependable n Replicable n Stable.
Reliability Analysis. Overview of Reliability What is Reliability? Ways to Measure Reliability Interpreting Test-Retest and Parallel Forms Measuring and.
Reliability and Validity
When Measurement Models and Factor Models Conflict: Maximizing Internal Consistency James M. Graham, Ph.D. Western Washington University ABSTRACT: The.
A quick introduction to the analysis of questionnaire data John Richardson.
Item Analysis Prof. Trevor Gibbs. Item Analysis After you have set your assessment: How can you be sure that the test items are appropriate?—Not too easy.
Multiple Choice Test Item Analysis Facilitator: Sophia Scott.
Research Methods in MIS
Reliability of Selection Measures. Reliability Defined The degree of dependability, consistency, or stability of scores on measures used in selection.
Social Science Research Design and Statistics, 2/e Alfred P. Rovai, Jason D. Baker, and Michael K. Ponton Internal Consistency Reliability Analysis PowerPoint.
Objectives The student will be able to: find the variance of a data set. find the standard deviation of a data set. SOL: A
Measures of Central Tendency
McMillan Educational Research: Fundamentals for the Consumer, 6e © 2012 Pearson Education, Inc. All rights reserved. Educational Research: Fundamentals.
Reliability Lesson Six
Reliability: Introduction. Reliability Session 1.Definitions & Basic Concepts of Reliability 2.Theoretical Approaches 3.Empirical Assessments of Reliability.
Validity and Reliability THESIS. Validity u Construct Validity u Content Validity u Criterion-related Validity u Face Validity.
Research Methodology Lecture No :24. Recap Lecture In the last lecture we discussed about: Frequencies Bar charts and pie charts Histogram Stem and leaf.
1 Chapter 4 – Reliability 1. Observed Scores and True Scores 2. Error 3. How We Deal with Sources of Error: A. Domain sampling – test items B. Time sampling.
Counseling Research: Quantitative, Qualitative, and Mixed Methods, 1e © 2010 Pearson Education, Inc. All rights reserved. Basic Statistical Concepts Sang.
Reliability & Agreement DeShon Internal Consistency Reliability Parallel forms reliability Parallel forms reliability Split-Half reliability Split-Half.
Objectives The student will be able to: find the variance of a data set. find the standard deviation of a data set.
Chapter 16 Data Analysis: Testing for Associations.
RELIABILITY Prepared by Marina Gvozdeva, Elena Onoprienko, Yulia Polshina, Nadezhda Shablikova.
Reliability n Consistent n Dependable n Replicable n Stable.
Chapter 16: Correlation. So far… We’ve focused on hypothesis testing Is the relationship we observe between x and y in our sample true generally (i.e.
Objectives The student will be able to:
Reliability: Introduction. Reliability Session 1.Definitions & Basic Concepts of Reliability 2.Theoretical Approaches 3.Empirical Assessments of Reliability.
Reliability: Introduction. Reliability Session Definitions & Basic Concepts of Reliability Theoretical Approaches Empirical Assessments of Reliability.
Reliability a measure is reliable if it gives the same information every time it is used. reliability is assessed by a number – typically a correlation.
Reliability When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing. Degree to which scores are free of.
Applied Quantitative Analysis and Practices LECTURE#17 By Dr. Osman Sadiq Paracha.
Introduction Dispersion 1 Central Tendency alone does not explain the observations fully as it does reveal the degree of spread or variability of individual.
5. Evaluation of measuring tools: reliability Psychometrics. 2011/12. Group A (English)
Reliability. Basics of test score theory Each person has a true score that would be obtained if there were no errors in measurement. However, measuring.
Normal Distribution Students will be able to: find the variance of a data set. find the standard deviation of a data set. use normal distribution curve.
Professor Jim Tognolini
Reliability Analysis.
Objectives The student will be able to:
Classical Test Theory Margaret Wu.
Reliability & Validity
Calculating Reliability of Quantitative Measures
Objectives The student will be able to:
Learning Targets I can: find the variance of a data set.
Reliability.
Reliability Analysis.
Standard Deviation Standard Deviation summarizes the amount each value deviates from the mean. SD tells us how spread out the data items are in our data.
Psychological Measurement: Reliability and the Properties of Random Errors The last two lectures were concerned with some basics of psychological measurement:
Objectives The student will be able to:
Objectives The student will be able to:
Chapter 8 VALIDITY AND RELIABILITY
Presentation transcript:

Explaining Cronbach’s Alpha Kirk Allen Graduate Research Assistant kcallen@ou.edu University of Oklahoma Dept. of Industrial Engineering

What is alpha and why should we care? Cronbach’s alpha is the most commonly used measure of reliability (i.e., internal consistency). It was originally derived by Kuder & Richardson (1937) for dichotomously scored data (0 or 1) and later generalized by Cronbach (1951) to account for any scoring method. People know that a high alpha is good, but it is important to have a deeper knowledge to use it properly. That is the purpose of this presentation.

Other types of reliability Test/Re-Test The same test is taken twice. Equivalent Forms Different tests covering the same topics Can be accomplished by splitting a test into halves

Cronbach’s basic equation for alpha n = number of questions Vi = variance of scores on each question Vtest = total variance of overall scores (not %’s) on the entire test

How alpha works Vi = pi * (1-pi) Vi varies from 0 to 0.25 pi = percentage of class who answers correctly This formula can be derived from the standard definition of variance. Vi varies from 0 to 0.25 pi 1-pi Vi 1 0.25 0.75 0.1875 0.5

How alpha works Vtest is the most important part of alpha If Vtest is large, it can be seen that alpha will be large also: Large Vtest  Small Ratio ΣVi/Vtest  Subtract this small ratio from 1  high alpha

High alpha is good. High alpha is caused by high variance. But why is high variance good? High variance means you have a wide spread of scores, which means students are easier to differentiate. If a test has a low variance, the scores for the class are close together. Unless the students truly are close in ability, the test is not useful.

What makes a question “Good” or “Bad” in terms of alpha? SPSS and SAS will report “alpha if item deleted”, which shows how alpha would change if that one question was not on the test. Low “alpha if item deleted” means a question is good because deleting that question would lower the overall alpha. In a test such as the SCI (34 items), no one question will have a large deviation from the overall alpha. Usually at most 0.03 in either direction

What causes a question to be “Bad”? Questions with high “alpha if deleted” tend to have low inter-item correlations (Pearson’s r).

What causes low or negative inter-item correlations? When a question tends to be answered correctly by students who have low overall scores on the test, but the question is missed by people with high overall scores. The “wrong” people are getting the question correct. Quantified by the “gap” between correct and incorrect students Correct students: average score 15.0 Incorrect students: average score 12.5 Gap = 15.0 – 12.5 = 2.5

Possible causes (based on focus group comments) If a question is “bad”, this means it is not conforming with the rest of the test to measure the same basic factor (e.g., statistics knowledge). The question is not “internally consistent” with the rest of the test. Possible causes (based on focus group comments) Students are guessing (e.g., question is too hard). Students use test-taking tricks (e.g., correct answer looks different from incorrect answers). Question requires a skill that is different from the rest of the questions (e.g., memory recall of a definition).

How does test length “inflate” alpha? For example, consider doubling the test length: Vtest will increase by a power of 4 because variance involves a squared term. However, ΣVi will only double because each Vi is just a number between 0 and 0.25. Since Vtest increases faster than ΣVi (recall that high Vtest is good), then alpha will increase by virtue of lengthening the test.

References Kuder & Richardson, 1937, “The Theory of the Estimation of Test Reliability” (Psychometrika v. 2 no. 3) Cronbach, 1951, “Coefficient Alpha and the Internal Structure of Tests” (Psychometrika v. 16 no. 3) Cortina, 1993, “What is coefficient alpha? An examination of theory and applications” (J. of Applied Psych. v. 78 no. 1 p. 98-104) Streiner, 2003, “Starting at the Beginning: An Introduction to Coefficient Alpha and Internal Consistency” (J. of Personality Assessment v. 80 no. 1 p. 99-103)