Presented at the 2006 CLEAR Annual Conference September 14-16 Alexandria, Virginia Something from Nothing: Limitations of Diagnostic Information in a CAT.

Slides:



Advertisements
Similar presentations
Knowledge Dietary Managers Association 1 PART II - DMA Certification Exam Blueprint and Exam Development-
Advertisements

SADC Course in Statistics General approaches to sample size determinations (Session 12)
What You Need to Know about the Computer Adaptive NREMT Exam.
Experts Respond to Your Questions 2006 Annual ConferenceAlexandria, Virginia Council on Licensure, Enforcement and Regulation Expect the Unexpected: Are.
What is a CAT?. Introduction COMPUTER ADAPTIVE TEST + performance task.
Copyright © 2012 Pearson Education, Inc. or its affiliate(s). All rights reserved
The trouble with resits … Dr Chris Ricketts Sub-Dean (Teaching Enhancement), Faculty of Technology and Director of Assessment, Peninsula College of Medicine.
Initial Licensure & NCLEX Karen Burke, RN, MS Education Consultant.
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Advanced Topics in Standard Setting. Methodology Implementation Validity of standard setting.
Effective Communication of Exam Results: What Should (or Shouldn't) be Included in the Candidate's Score Report Elizabeth A. Witt, Ph.D. American Board.
“You Can’t Always Get What You Want: Accountability and Limits in Score Reports to Boards” Troy Elliott Association of Social Work Boards 2006 Annual ConferenceAlexandria,
Chapter 7 Sampling Distributions
Software Quality Control Methods. Introduction Quality control methods have received a world wide surge of interest within the past couple of decades.
Statistical Evaluation of Data
Copyright ©2009 Cengage Learning 1.1 Day 3 What is Statistics?
+ A New Stopping Rule for Computerized Adaptive Testing.
Statistics for Managers Using Microsoft® Excel 7th Edition
Unit 3: Sample Size, Sampling Methods, Duration and Frequency of Sampling #3-3-1.
Classroom Assessment A Practical Guide for Educators by Craig A
1 Reducing the duration and cost of assessment with the GAIN: Computer Adaptive Testing.
“Accountability and Reporting in an Environment of Public Expectation” N. Genell Lee, MSN, RN, JD Alabama Board of Nursing 2006 Annual ConferenceAlexandria,
GUIDELINES FOR SETTING A GOOD QUESTION PAPER
Department of Health Professions Practical Nursing Directors Meeting C. N. Ridout, R.N., M.S., RNFA, CNE.
Descriptive Statistics Used to describe the basic features of the data in any quantitative study. Both graphical displays and descriptive summary statistics.
Kaizen–What Can I Do To Improve My Program? F. Jay Breyer, Ph.D. Presented at the 2005 CLEAR Annual Conference September Phoenix,
New York State Education Department Understanding The Process: Science Assessments and the New York State Learning Standards.
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
BY Karen Liu, Ph. D. Indiana State University August 18,
Copyright © 2010 Pearson Education, Inc. or its affiliate(s). All rights reserved. PCAT 101 September 29 and 30, 2010.
ASCP Board of Registry______________ ASCP Certification – The Standard of Excellence September 30, CLEAR Conference A Ten-Year Review of CBT.
1.1 Chapter One What is Statistics?. 1.2 What is Statistics? “Statistics is a way to get information from data.”
Chapter 7 Item Analysis In constructing a new test (or shortening or lengthening an existing one), the final set of items is usually identified through.
 Closing the loop: Providing test developers with performance level descriptors so standard setters can do their job Amanda A. Wolkowitz Alpine Testing.
Sterling Practices in Design & Scoring of Performance-Based Exams #156 F. Jay Breyer Presented at the 2005 CLEAR Annual Conference.
Chapter Thirteen Validation & Editing Coding Machine Cleaning of Data Tabulation & Statistical Analysis Data Entry Overview of the Data Analysis.
Fast Tracking Best Practice Assessment Through Task Force Review Colleen Anderson, Steven Barkley, Casey Marks, PhD & Cheryl Wild, PhD.
Building Individual Professionalism in Construction Since A Valid, Reliable & Fair Direct or Indirect Measure for ACCE Construction Program Student.
Counseling Research: Quantitative, Qualitative, and Mixed Methods, 1e © 2010 Pearson Education, Inc. All rights reserved. Basic Statistical Concepts Sang.
Standardized Testing (1) EDU 330: Educational Psychology Daniel Moos.
Descriptive Statistics Prepared by: Asma Qassim Al-jawarneh Ati Sardarinejad Reem Suliman Dr. Dr. Balakrishnan Muniandy PTPM-USM.
NCLEX ® is a Computerized Adaptive Test (CAT) How Does It Work?
Board of Registration in Nursing. Module: Preparing for Licensed Nursing Practice.
Assessment and Testing
Measures of variability: understanding the complexity of natural phenomena.
HESI RN CAT Exam.
Chapter 6: Analyzing and Interpreting Quantitative Data
NATIONAL CONFERENCE ON STUDENT ASSESSMENT JUNE 22, 2011 ORLANDO, FL.
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Understanding the 2015 Smarter Balanced Assessment Results Assessment Services.
Appropriate Testing Administration
Reliability EDUC 307. Reliability  How consistent is our measurement?  the reliability of assessments tells the consistency of observations.  Two or.
Taking the NCLEX exam – what to expect C. Matthews RN, MSN Spring 2010.
Advanced Dental Admission Test (ADAT)
Jean-Guy Blais Université de Montréal
What is a CAT? What is a CAT?.
Student Growth Measurements and Accountability
Best NCLEX Training Institute in Jalandhar
In-Service Teacher Training
Booklet Design and Equating
ATI facilitating the learning process for students
Using statistics to evaluate your test Gerard Seinhorst
Mohamed Dirir, Norma Sinclair, and Erin Strauts
Analyzing test data using Excel Gerard Seinhorst
Chapter 5THE NCLEX-PN®
Chapter 7: Sampling Distributions
Chapter 5THE NCLEX-PN®
Chapter 5THE NCLEX-PN®
Presentation transcript:

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Something from Nothing: Limitations of Diagnostic Information in a CAT Environment Casey Marks NCSBN 2006 Annual ConferenceAlexandria, Virginia Council on Licensure, Enforcement and Regulation Expect the Unexpected: Are We Clearly Prepared?

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia NCSBN Background Twenty-seven year old, not -for-profit organization consisting of 59 state and territorial boards of nursing Owner and developer of the national nurse licensure examinations (NCLEX-RN® and NCLEX-PN®) 17 NCSBN staff members assigned to NCLEX program operations NCSBN contracts with a testing service, Pearson VUE, to aid in the development and administration of the NCLEX

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia NCLEX® Facts NCLEX is the successor to the SBTPE, the national nurse licensure examination used by various nursing board for more than fifty years NCLEX has been administered via computer exclusively since 1994 when it was converted from Paper and Pencil to CBT administration Approximately 235,000 examinations administered per year, over 2 million examinations administered since 1994 Examinations administered continuously, on-demand in approximately 220 NCSBN approved Pearson Professional Centers around the world Both NCLEX-RN and NCLEX–PN are variable length, computerized adaptive examinations

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia What Is An Adaptive Test? It is tailored to the person taking it. It asks high ability people very few easy items. (They would very likely get them correct.) It asks low ability people very few difficult questions. (They would be guessing.) Everyone finds the test challenging. It can be fixed length or variable. It can be designed for maximum efficiency (50% correct) or something less efficient (perhaps 65% correct).

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia One Ability Is Measured The NCLEX examination makes a single decision, pass or fail. This decision is based on the assessment of the examinee’s ability. NCLEX treats “nursing ability” as a unitary concept. There is only the global ability estimate. Pass-fail decisions are never based on “subtest scores.”

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Item Difficulty To get the maximum amount of examinee information per question, the computer attempts to select questions for which the examinee has a 50/50 chance of answering it correctly. Because an adaptive test targets items to the person’s ability, the difficulty of the item must be known in advance.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Therefore, CAT requires that: The difficulty of each item is known. Item calibrations are stable across the ability continuum. (Item calibrations cannot be contingent on the ability level of the group of people testing.) Item calibrations are predominantly invariant across nursing-irrelevant factors such as gender and ethnicity.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia How CAT Works in NCLEX® It re-estimates the examinee’s ability after every answer. An item selection algorithm selects (from a large bank of items) an item that the examinee should find challenging. The test is efficient because high ability people get few easy items and low ability people get few difficult items.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia NCLEX® is a Variable Length Test The number of items an examinee receives on their test depends upon their ability. The RN test can range from 75 to 265 items (60 to 250 operational items). The PN/VN test can range from 85 to 205 items (60 to 180 operational items).

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia NCLEX® is a Variable Length Test To ensure that the content is adequately covered, no decisions regarding an examinee’s pass-fail status are made until at least 60 operational items have been answered. Although the test estimates the amount of ability an examinee has, ultimately a yes-no decision about the examinee’s competence must be made.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Content Balancing A specific percentage of the operational items must come from each content area. For example, Safety & Infection Control must be 11% (±3%) of each RN test and 10 (±3%) of each PN test. No examination ever deviates from these targeted percentages by more than 3%.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Content Balancing Each examination must conform to the Test Plan specifications. To ensure this, the item selection algorithm first determines what content area deviates the most from the test specifications. An item from that content area is administered next. Within that content area, an item of appropriate difficulty is selected.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Sample NCLEX-RN ®

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Pass-Fail Decisions Beginning with the 60th operational item, the ability estimate is compared to the passing standard. If you are clearly above the standard, you pass and the exam ends. If you are clearly below the standard, then you fail and the exam ends. If your ability estimate is so close to the standard that it's still not clear whether you should pass or not, then the computer continues to ask you questions.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Pretest Items On every test, there are a few pretest items. These items are randomly intermixed with operational questions in the beginning of the test. Pretest items are not used to estimate the examinee’s ability. Examinees cannot distinguish between pretest and operational items.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Test Length Minimum RN test length is 75 items, maximum RN test length is 265 items. (60 – 250 operational + 15 pretest) Minimum PN test length is 85 items,maximum PN test length is 205 items. (60 – 180 operational + 25 pretest)

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Pass Fail Begin Evaluation Fail Minimum Item Fail Chart

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia So why does all this make reporting difficult? Exam built for max information on min item exposure –Not enough items taken to reliability estimate sub scores –Most candidates get relatively the same number of items correct/incorrect –Even when enough info to calculate subscores highly correlated… –So…. What can be done?

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Candidate Performance Reports Are only provided to failing candidates. If the candidate did not answer at least the minimum number of items, no “diagnostic” feedback is given. If at least the minimum number of items were answered, descriptions are given for each of the content areas.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Program Reports Detailed information regarding program performance Can aid programs with: –Curriculum planning –Program evaluation –Trends Data compiled based on testing date for 1st time test takers only

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Program Reports Summary Overview Report Test Plan Report Content Dimension Reports Test Duration/Test Plan Performance Report

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Summary Overview Report Lists jurisdictions where graduates applied to for licensure. Percentage of graduates passing. Rank of program - based on the percentage of graduates passing. Distribution of programs based on national pass rates.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Summarizes the test plan. Summarizes graduates’ performance by test plan category. Graduates are compared to: –Graduates from last year –Graduates from the same jurisdiction –National population of graduates Test Plan Report

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Non-Test Plan Content Dimension Reports 6 frameworks –Nursing process –Categories of human functioning –Categories of health alterations –Wellness/illness continuum –Stages of maturity –Stress, adaptation and coping

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Non-Test Plan Content Dimension Reports Each framework describes how a typical graduate performed as compared to jurisdictional and national populations of graduates. Typical? – the ability of your median graduate.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Non-Test Plan Content Dimensions Nursing Process Categories of Human Functioning Categories of Health Alterations Wellness-Illness Continuum Stages of Maturity Stress, Adaptation, and Coping

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Test Duration/Test Performance Report Average number of questions answered Percent taking maximum questions Percentage taking minimum questions Average test time Reported for passers and failures separately, as well as for the total group.

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Conclusions? Limited information individual reports More detailed group reports for the “greater”good Suggestions?

Presented at the 2006 CLEAR Annual Conference September Alexandria, Virginia Speaker Contact Information Casey Marks,PhD NCSBN 111 E. Wacker Drive, Suite