PSY 614 Instructor: Emily Bullock, Ph.D.

Slides:



Advertisements
Similar presentations
Questionnaire Development
Advertisements

Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Reliability IOP 301-T Mr. Rajesh Gunesh Reliability  Reliability means repeatability or consistency  A measure is considered reliable if it would give.
MEASUREMENT: RELIABILITY Lu Ann Aday, Ph.D. The University of Texas School of Public Health.
Conceptualization and Measurement
Topics: Quality of Measurements
Reliability Definition: The stability or consistency of a test. Assumption: True score = obtained score +/- error Domain Sampling Model Item Domain Test.
© McGraw-Hill Higher Education. All rights reserved. Chapter 3 Reliability and Objectivity.
Chapter 5 Reliability Robert J. Drummond and Karyn Dayle Jones Assessment Procedures for Counselors and Helping Professionals, 6 th edition Copyright ©2006.
The Department of Psychology
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
Chapter 4 – Reliability Observed Scores and True Scores Error
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 5 Reliability.
 A description of the ways a research will observe and measure a variable, so called because it specifies the operations that will be taken into account.
Reliability Analysis. Overview of Reliability What is Reliability? Ways to Measure Reliability Interpreting Test-Retest and Parallel Forms Measuring and.
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
Reliability and Validity of Research Instruments
Reliability n Consistent n Dependable n Replicable n Stable.
Reliability Analysis. Overview of Reliability What is Reliability? Ways to Measure Reliability Interpreting Test-Retest and Parallel Forms Measuring and.
Lesson Seven Reliability. Contents  Definition of reliability Definition of reliability  Indication of reliability: Reliability coefficient Reliability.
Research Methods in MIS
Chapter 7 Correlational Research Gay, Mills, and Airasian
Reliability of Selection Measures. Reliability Defined The degree of dependability, consistency, or stability of scores on measures used in selection.
Technical Issues Two concerns Validity Reliability
Measurement and Data Quality
Reliability and Validity what is measured and how well.
Instrumentation.
Foundations of Educational Measurement
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
McMillan Educational Research: Fundamentals for the Consumer, 6e © 2012 Pearson Education, Inc. All rights reserved. Educational Research: Fundamentals.
Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand.
Technical Adequacy Session One Part Three.
1 Chapter 4 – Reliability 1. Observed Scores and True Scores 2. Error 3. How We Deal with Sources of Error: A. Domain sampling – test items B. Time sampling.
Counseling Research: Quantitative, Qualitative, and Mixed Methods, 1e © 2010 Pearson Education, Inc. All rights reserved. Basic Statistical Concepts Sang.
Tests and Measurements Intersession 2006.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Measurement MANA 4328 Dr. Jeanne Michalski
1 LANGUAE TEST RELIABILITY. 2 What Is Reliability? Refer to a quality of test scores, and has to do with the consistency of measures across different.
Reliability n Consistent n Dependable n Replicable n Stable.
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Chapter 6 - Standardized Measurement and Assessment
Reliability When a Measurement Procedure yields consistent scores when the phenomenon being measured is not changing. Degree to which scores are free of.
Reliability EDUC 307. Reliability  How consistent is our measurement?  the reliability of assessments tells the consistency of observations.  Two or.
TEST SCORES INTERPRETATION - is a process of assigning meaning and usefulness to the scores obtained from classroom test. - This is necessary because.
Language Assessment Lecture 7 Validity & Reliability Instructor: Dr. Tung-hsien He
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 5 What is a Good Test?
Assessing Student Performance Characteristics of Good Assessment Instruments (c) 2007 McGraw-Hill Higher Education. All rights reserved.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
5. Evaluation of measuring tools: reliability Psychometrics. 2011/12. Group A (English)
Reliability. Basics of test score theory Each person has a true score that would be obtained if there were no errors in measurement. However, measuring.
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Professor Jim Tognolini
Lecture 5 Validity and Reliability
Assessment Theory and Models Part II
RELIABILITY OF QUANTITATIVE & QUALITATIVE RESEARCH TOOLS
Classical Test Theory Margaret Wu.
Reliability & Validity
پرسشنامه کارگاه.
Calculating Reliability of Quantitative Measures
Reliability and Validity of Measurement
Evaluation of measuring tools: reliability
RESEARCH METHODS Lecture 18
MANA 5341 Dr. George Benson Measurement MANA 5341 Dr. George Benson 1.
By ____________________
The first test of validity
How can one measure intelligence?
Measurement Concepts and scale evaluation
Presentation transcript:

PSY 614 Instructor: Emily Bullock, Ph.D. Reliability PSY 614 Instructor: Emily Bullock, Ph.D.

Reliability Outline Sources of Error Types of Reliability Factors effecting reliability Nature of the Test

Sources of Error Observed Score = True Score + Error Test Construction-item or content sampling Test Administration-test environment, test taker variables, examiner related variables Test Scoring and Interpretation

Types of Reliability Test Retest Parallel-Forms and Alternate-Forms Internal Consistency Inter-Rater

Types of Reliability- Test Retest Test-Retest Reliability: estimate of reliability obtained by correlating pairs of scores from the same people on two different administrations of the same test Only appropriately used to measure reliability of stable constructs (ex. personality) Consider intervening variables due to time Interval time must be reported with test-retest

Types of Reliability- Parallel Forms and Alternate Forms True Parallel form: indicates that each form of the test produces equivalent means and variances of test scores Alternate forms: does not meet the above standard but attempts at two versions of a test that have been constructed to be equivalent with respect to content and difficulty Often termed the coefficient of equivalence

Types of Reliability- Internal Consistency Split-Half Reliability: obtained by correlating two pairs of scores obtained from equivalent halves of a single test administered once Steps of Split-Half Administer the entire test to participants Divide the test item scores into equivalent halves Compute a correlation (Pearson’s r) between the two halves Adjust the half-test reliability using Spearman-Brown formula How do you split a test?

Types of Reliability- Internal Consistency Inter-item consistency: degree of correlation among all the items on a scale Only appropriate for homogeneous scales or tests

Types of Reliability- Inter-Rater Inter-Rater or Inter-scorer reliability: degree of agreement or consistency that exists between two or more scorers

Purpose of the Reliability Coefficient Types of Reliability # of Test Sessions # of Test Forms Sources of Error Variance Stat. Procedure Test-retest 2 1 Administration Pearson’s r Alternate forms 1 or 2 Test construction & Administration Internal consistency Test Construction Pearson’s r with various corrections Inter-rater Scoring and Interpretation r, or kappa coefficient -

Factors Effecting Reliability Some legitimate factors contribute to variability of test scores (see Murphy & Davidshofer p. 118) Characteristics of the people taking the test What is your group of test takers like? Characteristics of the test itself Is the test worth making longer? The intended uses of the test scores Are your differences in test takers big or small? The method used to estimate reliability How long do you plan to make decisions based on this test?

What is good reliability? Commonly used standard: .75-.80 standard for decision-making tests .70 standard for research instruments Charter (2003) Article: Less than .70 unacceptable .70-.80 fair .80 to .90 good Greater than .90 excellent Text: .80 high reliability .70 moderate reliability .60 unacceptable reliability *This information is most applicable to internal consistency reliability

The Nature of the Test Homogeneous or Heterogeneous test items Characteristics, ability, or trait being measured is presume to be dynamic or static Range of test scores is or is not restricted Test is a speed or a power test Test is or is not criterion-referenced

Using Reliability and Error in Interpretation Standard Error of Measurement an index of the extent to which one individual’s scores vary over tests presumed to be parallel Indicates how much variability can be expected in test scores as a result of measurement error Can be used to form confidence intervals or bandwidths around scores