Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 1 Estimation of Item Difficulty Index Based on Item Response Theory.

Slides:



Advertisements
Similar presentations
What is a CAT?. Introduction COMPUTER ADAPTIVE TEST + performance task.
Advertisements

Optimal Analyses for 3  n AB Games in the Worst Case Li-Te Huang and Shun-Shii Lin Dept. of Computer Science & Information Engineering, National Taiwan.
Test Construction Processes 1- Determining the function and the form 2- Planning( Content: table of specification) 3- Preparing( Knowledge and experience)
The Research Problem PE 357. Selecting the problem Can be for research or a literature review -To break the problem down more … -needs to be of interest.
The use of a computerized automated feedback system Trevor Barker Dept. Computer Science.
RISKY SHIFT: INTRODUCTION Week 5 Practical. WEEK 5 PRACTICALRISKY SHIFT WEEK 1 WEEK 2 WEEK 3 WEEK 4 WEEK 5 WEEK 6 WEEK 7 WEEK 8 WEEK 9 WEEK 10 LECTURE.
A Kolmogorov Complexity Approach for Measuring Attack Path Complexity By Nwokedi C. Idika & Bharat Bhargava Presented by Bharat Bhargava.
Item Response Theory. Shortcomings of Classical True Score Model Sample dependence Limitation to the specific test situation. Dependence on the parallel.
© UCLES 2013 Assessing the Fit of IRT Models in Language Testing Muhammad Naveed Khalid Ardeshir Geranpayeh.
Why Scale -- 1 Summarising data –Allows description of developing competence Construct validation –Dealing with many items rotated test forms –check how.
A comparison of online and paper-based test results Sanja Čandrlić, Martina Ašenbrener Katić,
Techniques for Improving Student Learning Outcomes Lynn M. Forsythe Ida M. Jones Deborah J. Kemp Craig School of Business California State University,
Personalized mobile English vocabulary learning system based on item response theory and learning memory cycle Source ︰ Computers & Education Author ︰
Identification of Misfit Item Using IRT Models Dr Muhammad Naveed Khalid.
Reliability and factorial structure of a Portuguese version of the Children’s Hope Scale José Tomás da Silva Maria Paula Paixão Catarina Carvalho dos Santos.
Evaluation and analysis of the application of interactive digital resources in a blended-learning methodology for a computer networks subject F.A. Candelas,
1 Item Analysis - Outline 1. Types of test items A. Selected response items B. Constructed response items 2. Parts of test items 3. Guidelines for writing.
1 How can self-regulated learning be supported in mathematical E-learning environments? Presenters: Wei-Chih Hsu Professor : Ming-Puu Chen Date : 11/10/2008.
Modern Test Theory Item Response Theory (IRT). Limitations of classical test theory An examinee’s ability is defined in terms of a particular test The.
TESTS © LOUIS COHEN, LAWRENCE MANION & KEITH MORRISON.
Technical Adequacy Session One Part Three.
Out with the Old, In with the New: NYS Assessments “Primer” Basics to Keep in Mind & Strategies to Enhance Student Achievement Maria Fallacaro, MORIC
Suggested Guidelines for Thesis/Dissertation Chapters Chapter 1: Introduction  Problem Statement: Why is this research important?  What are the objectives.
Simulation of Computer Adaptive Learning and Improved Algorithm of Pyramidal Testing Taras Lendyuk 1, Sergey Rippa 2, Svitlana Sachenko 1 1 Ternopil National.
Are Some Banks More Lenient in Implementation of Placement Classification Rules?* An Application of Dichotomous Rasch Model to Classification of Credit.
Introduction to scientific ideas scientific method.
智慧型系統實驗室 iLab 南台資訊工程 1 Evaluation for the Test Quality of Dynamic Question Generation by Particle Swarm Optimization for Adaptive Testing Department of.
Ying (“Alison”) Cheng 1 John Behrens 2 Qi Diao 3 1 Lab of Educational and Psychological Measurement Department of Psychology, University of Notre Dame.
Measuring Human Intelligence with Artificial Intelligence Adaptive Item Generation Sangyoon Yi Susan E. Embretson.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. A Coursework Support System for Offering Challenges and Assistance by Analyzing Students’ Web Portfolios.
University of Georgia – Chemistry Department JExam - A Method to Measure Outcomes Assessment Charles H. Atwood, Kimberly D. Schurmeier, and Carrie G. Shepler.
Method Participants. Approximately 64 second (43.8%) and third (56.3%) grade students. 51.6% female and 48.6% male Measures Oral Reading Fluency (ORF)
A COMPARISON METHOD OF EQUATING CLASSIC AND ITEM RESPONSE THEORY (IRT): A CASE OF IRANIAN STUDY IN THE UNIVERSITY ENTRANCE EXAM Ali Moghadamzadeh, Keyvan.
The Effect of Brand Image and Service Quality on Southern Taiwan's Chain Restaurant Consumers' Repurchase Intentions Jessie Liu.
1 Item Analysis - Outline 1. Types of test items A. Selected response items B. Constructed response items 2. Parts of test items 3. Guidelines for writing.
Recommendation for English multiple-choice cloze questions based on expected test scores 2011, International Journal of Knowledge-Based and Intelligent.
Basic concepts in measurement Basic concepts in measurement Believe nothing, no matter where you read it, or who said it, unless it agrees with your own.
Item Response Theory in Health Measurement
INTRODUCTION HYPOTHESES MEASURES RESULTS Correspondence to: at the 27 th Association for Psychological Science Conference,
Intelligent Database Systems Lab N.Y.U.S.T. I. M. 1 Mining concept maps from news stories for measuring civic scientific literacy in media Presenter :
Descriptive and Inferential Statistics Descriptive statistics The science of describing distributions of samples or populations Inferential statistics.
Assessment to assure quality Barry McGaw Professorial Fellow, University of Melbourne Former Chair, Australian Curriculum, Assessment and Reporting Authority.
Self-assessing with adaptive exercises Chye-Foong Yong & Colin A Higgins University of Nottingham.
Two Approaches to Estimation of Classification Accuracy Rate Under Item Response Theory Quinn N. Lathrop and Ying Cheng Assistant Professor Ph.D., University.
Dan Thompson Oklahoma State University Center for Health Science Evaluating Assessments: Utilizing ExamSoft’s item-analysis to better understand student.
Thinking About Psychology: The Science of Mind and Behavior Charles T. Blair-Broeker Randal M. Ernst.
5. Evaluation of measuring tools: reliability Psychometrics. 2011/12. Group A (English)
Taras Lendyuk1, Sergey Rippa2, Svitlana Sachenko1
Author Name 1, Author Name 2, & Author Name 3
Oleh: Beni Setiawan, Wahyu Budi Sabtiawan
Assessing Students' Understanding of the Scientific Process Amy Marion, Department of Biology, New Mexico State University Abstract The primary goal of.
Assessing the Quality of Instructional Materials: Item Response Theory
7th International Conference on Cognitive Science
Adopting The Item Response Theory in Operations Management Research
Atuahene, B.T | Baiden, B.K | Agyekum, K
Preparing for the Verbal Reasoning Measure
Chapter 1: Introduction to Scientific Thinking
Research Methods.
Asch variation studies
Author Name 1, Author Name 2, & Author Name 3
Author Name 1, Author Name 2, & Author Name 3
By ____________________
ARE SMARTER THAN A SIXTH GRADER
MOCK TEST 3 ( ) DETAILED ANALYSIS.
HCI Evaluation Techniques
Psy 425 Tests & Measurements
Author Name 1, Author Name 2, & Author Name 3
Improving the Algorithm for Pyramidal Testing
Improving the Algorithm for Pyramidal Testing
Presentation transcript:

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 1 Estimation of Item Difficulty Index Based on Item Response Theory for Computerized Adaptive Testing Authors : Shu-Chen Cheng, Guan-Yu Chen

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology Introduction 2. Literature Reviews 3. Methods 4. Experiments and Results 5. Conclusions Outline

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 3 3 Computerized Adaptive Testing –Item Response Theory  Advantage: Personalized test, Shorter test length.  Shortcoming: The number of pre-test samples. IRT-1PL: 20 items, 200 testees (Wright & Stone, 1979) IRT-2PL: 30 items, 500 testees (Hulin et al., 1982) IRT-3PL: 60 items, 1000 testees (Hulin et al., 1982) ( There are 1,513 items in our item bank ! ) 1. Introduction (1/2)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 1. Introduction (2/2) Test System = Item Bank + Item Selection  Item Difficulty Index  Answers Abnormal Rate  Dynamic Item Selection Strategy  Particle Swarm Optimization 4

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology Computerized Adaptive Testing 2.2 Item Difficulty Index 2.3 Item Response Theory 2. Literature Reviews

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 6 6 To select the item that its difficulty is most consistent with testee’s ability. To assess testee’s ability immediately. The difficulty of next item is affected by previous answer. 2.1 Computerized Adaptive Testing (1/2)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 7 7 To test for different abilities through dynamitic item selection strategy. –High ability testee  No too easy items. –Low ability testee  No too difficult items. A personalized test. 2.1 Computerized Adaptive Testing (2/2)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 2.2 Item Difficulty Index (1/2) Method 1 : 8 P: Item difficulty. R: The number of correct answers. N: The number of total testees.

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 2.2 Item Difficulty Index (2/2) 9 Method 2 : P: Item difficulty. P H : Correct rate of high score group. P L : Correct rate of low score group. (Generally take 25%, 27%, 33%, etc.)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 10 Item Response Theory (Lord, 1980) –To estimate testee’s ability, aptitude, or location of other continuous psychological interval by the information of their item responses. –Ability location  Item response (Psychometric theory) –In addition to the model of IRT, without any other information to describe the item responses. 2.3 Item Response Theory (1/2)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology Item Response Theory (2/2)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 12 Answers 1)Testees’ ability > Item difficulty index  Most testees are supposed to answer correctly. 2)Testees’ ability < Item difficulty index  Most testees are supposed to answer wrong. 3)Testees’ ability = Item difficulty index  The correct answer rate is 50%. 3. Methods (1/4)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 13 Answers Abnormal –Violations of any one of these above 3 assumptions among answers are answers abnormal.  1st group with wrong answers. (Testee’s ability > Item difficulty)  2nd group with correct answers. (Testee’s ability < Item difficulty)  3rd group, correct answer rate ≠ 0.5. (Testee’s ability = Item difficulty) 3. Methods (2/4)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 14 Answers Abnormal Rate T : The number of correct answers. F : The number of wrong answers. N : The number of total testees. h : 1st group (Testee’s ability > Item difficulty). l : 2nd group (Testee’s ability < Item difficulty). e : 3rd group (Testee’s ability = Item difficulty). 3. Methods (3/4)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology Methods (4/4) 15 Item Difficulty Difficulty j, letbe the smallest. : Item difficulty index of item i. Answers abnormal rate of item i with difficulty j. :

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology System Descriptions 4.2 Experiment Descriptions 4.3 Results and Discussions 4. Experiments and Results

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology System Descriptions (1/3)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology System Descriptions (2/3)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 4.1 System Descriptions (3/3) 19 PSO Dynamic Item Selection Strategy Item Difficulty Knowledge Weights Item Exposure Rate

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology Experiment Descriptions Method: Online test Item Bank: –Items: 1,513 –Initial Difficulty: 0.5 (9 levels, 0.1~0.9) Participants: –Students: 51 –Initial Ability: 0.2 (9 levels, 0.1~0.9) Periods: 6 weeks

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology Results and Discussions (1/3)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology Results and Discussions (2/3)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology Results and Discussions (3/3)

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology Conclusions Each test item is treated as independent, and the item difficulty can be estimated individually. Therefore, the item bank can be expanded easily at any time. The estimation based on the answers abnormal rate proposed in this study can estimate the item difficulty index quickly and reasonably without too many pre- test samples.

Intelligent System Lab. (iLab) Southern Taiwan University of Science and Technology 25 The End ~ Thanks for your attention!