Item Analysis Ursula Waln, Director of Student Learning Assessment

Slides:



Advertisements
Similar presentations
Item Analysis.
Advertisements

Test Development.
FACULTY DEVELOPMENT PROFESSIONAL SERIES OFFICE OF MEDICAL EDUCATION TULANE UNIVERSITY SCHOOL OF MEDICINE Using Statistics to Evaluate Multiple Choice.
Item Analysis: A Crash Course Lou Ann Cooper, PhD Master Educator Fellowship Program January 10, 2008.
How to Norm Rubrics Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.
The Effects of Achievement Priming on Expectations and Performance Kathryn Raso Team 14 PSY 321.
Measuring Student Learning March 10, 2015 Cathy Sanders Director of Assessment.
Test Construction Processes 1- Determining the function and the form 2- Planning( Content: table of specification) 3- Preparing( Knowledge and experience)
MATH ASSESSMENT TEST OCMA May, HISTORY OF MAT Test originally developed in late 60’s.
Item Analysis What makes a question good??? Answer options?
Item Analysis Prof. Trevor Gibbs. Item Analysis After you have set your assessment: How can you be sure that the test items are appropriate?—Not too easy.
Business Math, Eighth Edition Cleaves/Hobbs © 2009 Pearson Education, Inc. Upper Saddle River, NJ All Rights Reserved 7.1 Measures of Central Tendency.
Multiple Choice Test Item Analysis Facilitator: Sophia Scott.
1 1 Slide © 2003 South-Western/Thomson Learning TM Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
1 Business Math Chapter 7: Business Statistics. Cleaves/Hobbs: Business Math, 7e Copyright 2005 by Pearson Education, Inc. Upper Saddle River, NJ
SENSE 2013 Findings for College of Southern Idaho.
Designing and evaluating good multiple choice items Jack B. Monpas-Huber, Ph.D. Director of Assessment & Student Information.
Migration from ExamSystem II to REMARK Examination Scanning Software.
Part #3 © 2014 Rollant Concepts, Inc.2 Assembling a Test #
Are there “Hidden Variables” in Students’ Initial Knowledge State Which Correlate with Learning Gains? David E. Meltzer Department of Physics and Astronomy.
Test item analysis: When are statistics a good thing? Andrew Martin Purdue Pesticide Programs.
Introduction Neuropsychological Symptoms Scale The Neuropsychological Symptoms Scale (NSS; Dean, 2010) was designed for use in the clinical interview to.
The Genetics Concept Assessment: a new concept inventory for genetics Michelle K. Smith, William B. Wood, and Jennifer K. Knight Science Education Initiative.
Chapter 7 Item Analysis In constructing a new test (or shortening or lengthening an existing one), the final set of items is usually identified through.
110/10/2015Slide 1 The homework problems on comparing central tendency and variability extend our focus on central tendency and variability to a comparison.
Instrumentation (cont.) February 28 Note: Measurement Plan Due Next Week.
Research & Statistics Looking for Conclusions. Statistics Mathematics is used to organize, summarize, and interpret mathematical data 2 types of statistics.
Techniques to improve test items and instruction
Group 2: 1. Miss. Duong Sochivy 2. Miss. Im Samphy 3. Miss. Lay Sreyleap 4. Miss. Seng Puthy 1 ROYAL UNIVERSITY OF PHNOM PENH INSTITUTE OF FOREIGN LANGUAGES.
Key Concepts, continued To determine the probability of an outcome using continuous data, we use the proportion of the area under the normal curve associated.
Assessment in College Teaching Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College.
NRTs and CRTs Group members: Camila, Ariel, Annie, William.
Reliability & Validity
Research Methodology Lecture No :24. Recap Lecture In the last lecture we discussed about: Frequencies Bar charts and pie charts Histogram Stem and leaf.
Lab 5: Item Analyses. Quick Notes Load the files for Lab 5 from course website –
NC NAEP Project Module 2 - Activity 3 Multi-step NAEP Items: Exploring Distractor Answers Elementary Module 2, Activity 3.
Administering, Analyzing, and Improving the Written Test
How to Perform Simple Manual Item Analysis Dr. Belal Hijji, RN, PhD January 18, 2012.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Class # 3 Copyright © 2005 by R. Halstead. All rights reserved.
Grading and Analysis Report For Clinical Portfolio 1.
8 Strategies for the Multiple Choice Portion of the AP Literature and Composition Exam.
Basic Rules for Test Taking Success Adelita G. Cantu, PhD, RN
RELIABILITY AND VALIDITY OF ASSESSMENT
Assessment and Testing
Introduction to Item Analysis Objectives: To begin to understand how to identify items that should be improved or eliminated.
1 Children First Intensive 2008 Grade 5 Social Studies Analyzing Outcomes for ESO Network 14 March 25, 2009 Social Studies Conference, PS/MS 3 Deena Abu-Lughod,
Overview and interpretation
TEST SCORES INTERPRETATION - is a process of assigning meaning and usefulness to the scores obtained from classroom test. - This is necessary because.
Dan Thompson Oklahoma State University Center for Health Science Evaluating Assessments: Utilizing ExamSoft’s item-analysis to better understand student.
Psychometrics: Exam Analysis David Hope
Dept. of Community Medicine, PDU Government Medical College,
Copyright © Springer Publishing Company, LLC. All Rights Reserved. DEVELOPING AND USING TESTS – Chapter 11 –
Items analysis Introduction Items can adopt different formats and assess cognitive variables (skills, performance, etc.) where there are right and.
COMMON TEST TECHNIQUES FROM TESTING FOR LANGUAGE TEACHERs.
Ch. 5 Measurement Concepts.
ARDHIAN SUSENO CHOIRUL RISA PRADANA P.
Assessment Instruments and Rubrics Workshop Series
Data Analysis and Standard Setting
assessing scale reliability
Classroom Analytics.
Classroom Assessment Validity And Bias in Assessment.
Peer Evaluation of Teammates
Test Development Test conceptualization Test construction Test tryout
Using Evidence to Improve NCLEX® Pass Rates
Dept. of Community Medicine, PDU Government Medical College,
Statistical Inference for Managers
Summative Assessment Grade 6 April 2018 Develop Revise Pilot Analyze
Analyzing test data using Excel Gerard Seinhorst
Tests are given for 4 primary reasons.
Presentation transcript:

Item Analysis Ursula Waln, Director of Student Learning Assessment Central New Mexico Community College

Item Analysis Used with Objective Assessment Looks at frequency of correct responses (or behaviors) in connection with overall performance Used to examine item reliability How consistently a question or performance criterion discriminates between high and low performers Can be useful in improving validity of measures Can help instructors decide whether to eliminate certain items from the grade calculations Can reveal specific strengths and gaps in student learning

How Item Analysis Works Groups students by the highest, mid-range, and lowest overall scores and examines item responses by group Assumes that higher-scoring students have a higher probability of getting any given item correct than do lower-scoring students May have studied and/or practiced more and understood the material better May have greater test-taking savvy, less anxiety, etc. Produces a calculation for each item Do it yourself to easily calculate a group difference or discrimination index Use EAC Outcomes (a Blackboard plug-in made available to all CNM faculty by the Nursing program) to generate a point-biserial correlation coefficient Gives the instructor a way to analyze performance on each item

One Way to Do Item Analysis by Hand Shared by Linda Suskie at the NMHEAR Conference, 2015 Tally of those in Top 27% who missed item* Tally of those in the Middle 46% who missed item Tally of those in the Lower 27% who missed item* Total % Who Missed Item Group Difference (# in Lower minus # in Top) 1 ||||| ||||| ||||| ||||| || ||||| ||||| ||||| || 34% 17 2 ||||| || ||||| ||||| ||||| ||||| ||||| ||||| ||||| |||| 40% 12 3 ||| | || 5% -1 4   ||||| |||| ||||| ||||| 17% 11 Good discrimination An unreliable question * You can use whatever portion you want for the top and lower groups, but they need to be equal. Using 27% is accepted convention (Truman Kelley, 1939).

Another Way to Do Item Analysis by Hand Rasch Item Discrimination Index (D) N=31 because the upper and lower group each contain 31 students (115 students tested) Item # in Upper Group who answered correctly (#UG) Portion of UG who answered correctly (pUG) # in Lower Group who answered correctly (#LG) Portion of LG who answered correctly (pLG) Discrimination Index (D) D = pUG−pLG or D = #𝑈𝐺−#𝐿𝐺 𝑁 1 31 1.00 (100%) 14 0.45 (45%) 0.55 2 24 0.77 (77%) 12 0.39 (39%) 0.38 3 28 0.90 (90%) 29 0.93 (93%) -0.03 4  31 20 0.65 (65%) 0.35 Good discrimination An unreliable question A discrimination index of 0.4 or greater is generally regarded as high and anything less than 0.2 as low (R.L. Ebel, 1954).

The Same Thing but Less Complicated Rasch Item Discrimination Index (D) N in Upper and Lower Groups is 31 (27% of 115 students) Item # in Upper Group who answered correctly (#UG) # in Lower Group who answered correctly (#LG) Discrimination Index (D) D = #𝑈𝐺−#𝐿𝐺 𝑁 1 31 14 0.55 2 24 12 0.38 3 28 29 -0.03 4  31 20 0.35 This is really easy to do! 31−14 31 = 0.55 24−12 31 = 0.38 28−29 31 = -0.03 31−20 31 = 0.35 It isn’t necessary to calculate the portions of correct responses in each group if you use the formula shown here.

Example of an EAC Outcomes Report Good discrimination An unreliable question A point-biserial correlation is the Pearson correlation between responses to a particular item and scores on the total test (with or without that item). Correlation coefficients range from -1 to 1. This is available to CNM faculty through Blackboard course tools.

Identifying Key Questions A key (a.k.a. signature) question is one that provides information about student learning in relation to a specific instructional objective (or student learning outcome statement). The item analysis methods shown in the preceding slides can help you identify and improve the reliability of key questions. A low level of discrimination may indicate a need to tweak the wording. Improving discrimination value also improves question validity. The more valid an assessment measure, the more useful it is in gauging student learning.

Detailed Multiple-Choice Item Analysis The detailed item analysis method shown on the next slide is for use with key multiple-choice items. This type of analysis can provide clues to the nature of students’ misunderstanding, provided: The item is a valid measure of the instructional objective Incorrect options (distractors) are written to be diagnostic (i.e., to reveal misconceptions or breakdowns in understanding)

Example of a Detailed Item Analysis Item 2 of 4. The correct option is E. (115 students tested)   Item Response Pattern A B C D E Row Total Upper 27% || ||||| ||||| ||||| ||||| ||||| |||| 31 6.5% 16% 77.5% Middle 46% ||| ||||| ||||| |||| | ||||| ||||| ||||| ||||| ||||| ||||| ||| 53 6% 26% 4% 2% 62% Lower 27% ||||| || ||||| ||||| || 23% 39% Grand Total 10 26 7 3 69 115 8.5% 2.5% 60% These results suggest that distractor B might provide the greatest clue about breakdown in students’ understanding, followed by distractor A, then C.