The ABC’s of Pattern Scoring Dr. Cornelia Orr. Slide 2 Vocabulary Measurement – Psychometrics is a type of measurement Classical test theory Item Response.

Slides:



Advertisements
Similar presentations
Psychometrics to Support RtI Assessment Design Michael C. Rodriguez University of Minnesota February 2010.
Advertisements

Test Development.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 January 23, 2012.
Item Response Theory in a Multi-level Framework Saralyn Miller Meg Oliphint EDU 7309.
The measurement model: what does it mean and what you can do with it? Presented by Michael Nering, Ph. D.
1 Scaling of the Cognitive Data and Use of Student Performance Estimates Guide to the PISA Data Analysis ManualPISA Data Analysis Manual.
Topic 4B Test Construction.
Reliability Definition: The stability or consistency of a test. Assumption: True score = obtained score +/- error Domain Sampling Model Item Domain Test.
Item Response Theory in Health Measurement
Introduction to Item Response Theory
AN OVERVIEW OF THE FAMILY OF RASCH MODELS Elena Kardanova
Item Analysis What makes a question good??? Answer options?
Item PersonI1I2I3 A441 B 323 C 232 D 112 Item I1I2I3 A(h)110 B(h)110 C(l)011 D(l)000 Item Variance: Rank ordering of individuals. P*Q for dichotomous items.
Item Response Theory. Shortcomings of Classical True Score Model Sample dependence Limitation to the specific test situation. Dependence on the parallel.
z-Scores What is a z-Score? How Are z-Scores Useful? Distributions of z-Scores Standard Normal Curve.
Measurement Joseph Stevens, Ph.D. ©  Measurement Process of assigning quantitative or qualitative descriptions to some attribute Operational Definitions.
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
Standardized Test Scores Common Representations for Parents and Students.
Classical Test Theory By ____________________. What is CCT?
Item Analysis: Classical and Beyond SCROLLA Symposium Measurement Theory and Item Analysis Modified for EPE/EDP 711 by Kelly Bradley on January 8, 2013.
Identification of Misfit Item Using IRT Models Dr Muhammad Naveed Khalid.
Item Response Theory Psych 818 DeShon. IRT ● Typically used for 0,1 data (yes, no; correct, incorrect) – Set of probabilistic models that… – Describes.
Item Response Theory. What’s wrong with the old approach? Classical test theory –Sample dependent –Parallel test form issue Comparing examinee scores.
Measurement 102 Steven Viger Lead Psychometrician
1 Item Analysis - Outline 1. Types of test items A. Selected response items B. Constructed response items 2. Parts of test items 3. Guidelines for writing.
Introduction to plausible values National Research Coordinators Meeting Madrid, February 2010.
Modern Test Theory Item Response Theory (IRT). Limitations of classical test theory An examinee’s ability is defined in terms of a particular test The.
Technical Adequacy Session One Part Three.
Out with the Old, In with the New: NYS Assessments “Primer” Basics to Keep in Mind & Strategies to Enhance Student Achievement Maria Fallacaro, MORIC
UNIT IV ITEM ANALYSIS IN TEST DEVELOPMENT CHAP 14: ITEM ANALYSIS CHAP 15: INTRODUCTION TO ITEM RESPONSE THEORY CHAP 16: DETECTING ITEM BIAS 1.
Test item analysis: When are statistics a good thing? Andrew Martin Purdue Pesticide Programs.
智慧型系統實驗室 iLab 南台資訊工程 1 Evaluation for the Test Quality of Dynamic Question Generation by Particle Swarm Optimization for Adaptive Testing Department of.
Measuring Human Intelligence with Artificial Intelligence Adaptive Item Generation Sangyoon Yi Susan E. Embretson.
Educational Psychology: Theory and Practice Chapter 14 Standardized Tests This multimedia product and its contents are protected under copyright law. The.
Test Scaling and Value-Added Measurement Dale Ballou Vanderbilt University April, 2008.
University of Georgia – Chemistry Department JExam - A Method to Measure Outcomes Assessment Charles H. Atwood, Kimberly D. Schurmeier, and Carrie G. Shepler.
A COMPARISON METHOD OF EQUATING CLASSIC AND ITEM RESPONSE THEORY (IRT): A CASE OF IRANIAN STUDY IN THE UNIVERSITY ENTRANCE EXAM Ali Moghadamzadeh, Keyvan.
1 Item Analysis - Outline 1. Types of test items A. Selected response items B. Constructed response items 2. Parts of test items 3. Guidelines for writing.
Differential Item Functioning. Anatomy of the name DIFFERENTIAL –Differential Calculus? –Comparing two groups ITEM –Focus on ONE item at a time –Not the.
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
The ABC’s of Pattern Scoring
SOCW 671: #5 Measurement Levels, Reliability, Validity, & Classic Measurement Theory.
Multitrait Scaling and IRT: Part I Ron D. Hays, Ph.D. Questionnaire Design and Testing.
Item Factor Analysis Item Response Theory Beaujean Chapter 6.
NATIONAL CONFERENCE ON STUDENT ASSESSMENT JUNE 22, 2011 ORLANDO, FL.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Latent regression models. Where does the probability come from? Why isn’t the model deterministic. Each item tests something unique – We are interested.
Item Response Theory in Health Measurement
Item Analysis: Classical and Beyond SCROLLA Symposium Measurement Theory and Item Analysis Heriot Watt University 12th February 2003.
The Design of Statistical Specifications for a Test Mark D. Reckase Michigan State University.
2. Main Test Theories: The Classical Test Theory (CTT) Psychometrics. 2011/12. Group A (English)
Item Response Theory Dan Mungas, Ph.D. Department of Neurology University of California, Davis.
Multitrait Scaling and IRT: Part I Ron D. Hays, Ph.D. Questionnaire.
Item Response Theory and Computerized Adaptive Testing Hands-on Workshop, day 2 John Rust, Iva Cek,
Lesson 2 Main Test Theories: The Classical Test Theory (CTT)
Chapter 2 Norms and Reliability. The essential objective of test standardization is to determine the distribution of raw scores in the norm group so that.
IRT Equating Kolen & Brennan, 2004 & 2014 EPSY
The Impact of Item Response Theory in Educational Assessment: A Practical Point of View Cees A.W. Glas University of Twente, The Netherlands University.
Using Item Response Theory to Track Longitudinal Course Changes
Classical Test Theory Margaret Wu.
The Impact of Item Response Theory in Educational Assessment: A Practical Point of View Cees A.W. Glas University of Twente, The Netherlands University.
Item Analysis: Classical and Beyond
By ____________________
Mohamed Dirir, Norma Sinclair, and Erin Strauts
UNIT IV ITEM ANALYSIS IN TEST DEVELOPMENT
Item Analysis: Classical and Beyond
Evaluating Multi-item Scales
Multitrait Scaling and IRT: Part I
Item Analysis: Classical and Beyond
Presentation transcript:

The ABC’s of Pattern Scoring Dr. Cornelia Orr

Slide 2 Vocabulary Measurement – Psychometrics is a type of measurement Classical test theory Item Response Theory – IRT (AKA logistic trait theory) 1, 2, & 3-parameter IRT models Pattern Scoring

Slide 3 General & Specialized Measurement Assign numbers to objects or events Ex. – hurricanes, earthquakes, time, stock market, height, weight Psychometrics Assigning numbers to psychological characteristics Ex. – achievement personality, IQ, opinion, interests

Slide 4 Different Theories of Psychometrics Classical Test Theory Item discrimination values Item difficulty values (p-values) Guessing (penalty) Number correct scoring Item Response Theory a)Item discrimination values b)Item difficulty values c)Guessing (pseudo- guessing) values Pattern scoring Similar constructs – Different derivations

Slide 5 Different Methods of Scoring Number-Correct Scoring Simple Mathematics Raw scores (# of points) –Mean, SD, SEM, % correct Number right scale Score conversions –Scale scores, percentile ranks, etc. Pattern Scoring Complex Mathematics Maximum likelihood estimates –Item statistics, student’s answer pattern, SEM Theta scale (mean=0, standard dev=1) Score conversions –Scale scores, percentile ranks, etc.

Slide 6 Comparison: Number Correct and Pattern Scoring Similarities The relationship of derived scores is the same, e.g., –High correlation, (0.95) of number right scores and scale scores –Scale score has the same percentile rank for both methods Differences Methods of deriving scores The number of scale scores possible –Number right = limited to the number of items –IRT = unlimited or is limited by the scale (ex )

Slide 7 Choosing the Scoring Method Which model? Simple vs. Complex? Best estimates? Advantages/Disadvantages? Ex. – Why do the same number correct get different scale scores? Ex. – Flat screen TV – how do they do that?

Slide 8 Advantages of IRT and Pattern Scoring Better estimates of an examinee’s ability –the score that is most likely, given the student’s responses to the questions on the test (maximum likelihood scoring) More information about students and items are used More reliability than number right scoring Less measurement error (SEM)

Slide 9 Disadvantages of IRT and Pattern Scoring Technical - Complex Mathematics – –Difficult to understand –Difficult to explain Not common – Not like my experience. Perceived as “Hocus Pocus”

Slide 10 Item Characteristic Curve (ICC)

Slide 11 Examples Effect Of Item Difficulty No Type a b c 1 1 MC MC MC MC MC Response Patterns (1=correct) Pattern SEM SS Answering more difficult items (b-parameter) can result in higher scores.

Slide 12 Examples 5 Items (Effects of Item Discrimination) No Type a b c 1 MC MC MC MC MC Response patterns (1=correct) Pattern SEM SS Answering more discriminating items (a-parameter) can result in higher scores.