Cut Points ITE - 695. Section One n What are Cut Points?

Slides:



Advertisements
Similar presentations
Assessing Student Performance
Advertisements

Hypothesis testing Another judgment method of sampling data.
© McGraw-Hill Higher Education. All rights reserved. Chapter 3 Reliability and Objectivity.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
Standardized Tests What They Measure How They Measure.
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Chapter 4 Validity.
REVIEW I Reliability Index of Reliability Theoretical correlation between observed & true scores Standard Error of Measurement Reliability measure Degree.
Becoming Acquainted With Statistical Concepts CHAPTER CHAPTER 12.
QUANTITATIVE DATA ANALYSIS
STANDARDIZED TESTING MEASUREMENT AND EVALUATION  All teaching involves evaluation. We compare information to criteria and ten make judgments.  Measurement.
DATA ANALYSIS I MKT525. Plan of analysis What decision must be made? What are research objectives? What do you have to know to reach those objectives?
Appraisal in Counseling Session 2. Schedule Finish History Finish History Statistical Concepts Statistical Concepts Scales of measurement Scales of measurement.
Introduction to Educational Statistics
Copyright 2001 by Allyn and Bacon Standardized Testing Chapter 14.
Multiple Choice Test Item Analysis Facilitator: Sophia Scott.
Chapter 2 Simple Comparative Experiments
Measuring Human Performance. Introduction n Kirkpatrick (1994) provides a very usable model for measurement across the four levels; Reaction, Learning,
Today Concepts underlying inferential statistics
Research Methods in MIS
Statistics for CS 312. Descriptive vs. inferential statistics Descriptive – used to describe an existing population Inferential – used to draw conclusions.
Introduction to Statistical Quality Control, 4th Edition Chapter 15 Other Acceptance Sampling Techniques.
Chap 11 Engineering Statistics PREP004 – Introduction to Applied Engineering College of Engineering - University of Hail Fall 2009.
Foundations of Educational Measurement
Data Analysis. Quantitative data: Reliability & Validity Reliability: the degree of consistency with which it measures the attribute it is supposed to.
McMillan Educational Research: Fundamentals for the Consumer, 6e © 2012 Pearson Education, Inc. All rights reserved. Educational Research: Fundamentals.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
CRT Dependability Consistency for criterion- referenced decisions.
Analyzing and Interpreting Quantitative Data
Instrumentation (cont.) February 28 Note: Measurement Plan Due Next Week.
Basic Statistics for Engineers. Collection, presentation, interpretation and decision making. Prof. Dudley S. Finch.
Reliability & Validity
Counseling Research: Quantitative, Qualitative, and Mixed Methods, 1e © 2010 Pearson Education, Inc. All rights reserved. Basic Statistical Concepts Sang.
Chapter 21 Basic Statistics.
EDU 8603 Day 6. What do the following numbers mean?
Item specifications and analysis
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Class # 3 Copyright © 2005 by R. Halstead. All rights reserved.
Research Methods. Measures of Central Tendency You will be familiar with measures of central tendency- averages. Mean Median Mode.
Analyzing Statistical Inferences How to Not Know Null.
 Two basic types Descriptive  Describes the nature and properties of the data  Helps to organize and summarize information Inferential  Used in testing.
Three Broad Purposes of Quantitative Research 1. Description 2. Theory Testing 3. Theory Generation.
Chapter 6: Analyzing and Interpreting Quantitative Data
RESEARCH & DATA ANALYSIS
Sampling (cont.) Instrumentation Measurement Plan Due 3/7.
Quality Control: Analysis Of Data Pawan Angra MS Division of Laboratory Systems Public Health Practice Program Office Centers for Disease Control and.
1 Outline 1. Why do we need statistics? 2. Descriptive statistics 3. Inferential statistics 4. Measurement scales 5. Frequency distributions 6. Z scores.
Psychologist use statistics for 2 things
Assessment through Standardized Testing Chapter 15.
REVIEW I Reliability scraps Index of Reliability Theoretical correlation between observed & true scores Standard Error of Measurement Reliability measure.
Standardized Testing. Basic Terminology Evaluation: a judgment Measurement: a number Assessment: procedure to gather information.
Chapter 6 - Standardized Measurement and Assessment
Measurements and Their Analysis. Introduction Note that in this chapter, we are talking about multiple measurements of the same quantity Numerical analysis.
Chapter 3 Selection of Assessment Tools. Council of Exceptional Children’s Professional Standards All special educators should possess a common core of.
Classroom Assessment Chapters 4 and 5 ELED 4050 Summer 2007.
Chapter 6 Becoming Acquainted With Statistical Concepts.
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 5 What is a Good Test?
Chapter 2 Norms and Reliability. The essential objective of test standardization is to determine the distribution of raw scores in the norm group so that.
Appendix I A Refresher on some Statistical Terms and Tests.
CHAPTER 3: Practical Measurement Concepts
Research Methodology Lecture No :25 (Hypothesis Testing – Difference in Groups)
Understanding Results
Chapter 2 Simple Comparative Experiments
Statistical significance & the Normal Curve
Using statistics to evaluate your test Gerard Seinhorst
Chapters 5 Formal Assessment.
Week 11.
REVIEW I Reliability scraps Index of Reliability
Presentation transcript:

Cut Points ITE - 695

Section One n What are Cut Points?

I. Introduction A. The more critical the issue (task) the more critical the cut point (example: programming a machine). 1. Interpretation of readouts. 2. Tolerances in measurement. B. Assumption: Test has both of these: 1. Validity. 2. Reliability. C. Select instrument that best measures action needed (performance vs. explanation).

II Types A. Normative-Referenced Testing (NTR) 1. Significance - Accepted reliability & validity 2. Measurement a. Common Averages: - mode - median - mean

II Types (cont.) b. Variability: - range - quartile deviation - standard deviation 3. Reliability - Historical acceptance

II Types (cont.) B. Criterion-Referenced testing (CRT) 1. Significance a. Testing b. Distribution 2. Measurement a. Judgements b. Variables

II Types (cont.) 3. Reliability a. Criterion not based on normal distribution. b. Data dichotomous, mastery/non-mastery.

NORM REFRENCED TESTING 1. Separate test takers 2. Seek Normal Distribution Curve

NORM REFRENCED TESTING 1.Test items separate test - takers from one another. 2. Normal Distribution Curve.

MEASURES of CENTRAL TENDENCIES n MODE n MEDIAN n MEAN n MEASURES of VARIBILITY or SCATTER – RANGE – DEVIATION (QUARTILE) – DEVIATION (STANDARD)

CRITERION REFERENCED TESTING 1. Test items based on specific objectives. 2. Mastery Curve

Standard normal curve with standard deviations SEE HANDOUT

CRITERION REFRENCED TEST 1. Test Compares to Objectives 2. Mastery Distribution

Norm-Reference Testing GOALS RELIABILITY VALIDITY ADMINISTRATION STANDARD MOTIVATION COMPETITION INSTRUCTIONAL DOMAIN Criterion Referenced Testing Test Achievement Usually High Instruction Dependent Standard Averages-Based Avoidance of Failure Student to Student Low Level Cognitive Test Performance Mastery Usually Unknown Usually High Variable Performance Levels Based Likelihood of Success Student to Criterion Cognitive or Psychomotor

Comparison models ? INPUTPRODUCT (Instruction) (NRT Results) Model For NRT Construction DESIGN TESTINPUTPRODUCT MODIFY? NOYES (Instruction)(CRT Results) (Test, Objectives, or Instruction) Model For CRT Construction

Mastery curve SEE HANDOUT

Frequency distributions with standard deviations of various sizes SEE HANDOUT

Section II Establishing Cut Points Three Primary Procedures

ESTABLISHING CUT-POINT 1. Informed Judgement 2. Conjectural Approach 3. Contrast Group

I. Informed Judgement A. Significance: Separates mastery from non- mastery B. Procedure: 1. Analyze consequences of mid- classification (political, legal, or operational). 2. Gather previous test-taker data. 3. Ask other stakeholders. 4. Make decision.

II Conjecture Method A. Significance: “Angoff-Nedeisky Method” - most useful. B. Procedure: 1. Select three informed judges. 2. Estimate probability of correct response. 3. Chosen cut-off is average of the three judges.

III Contrast Group Method A. Significance: Single strongest technique; should still use human judgement. B. Procedure: 1. Select judges to identify mastery/non-mastery. 2. Select equal groups (15 minimum, 30 optimum). 3. Administer mastery/non-mastery test to both groups. 4. Plot scores on distribution chart. 5. Make critical cut-off where two distributions intersect. 6. Adjust score between highest non-master and lowest master. score.

Establishing A Criterion Cut- Point Mastery Level - (Separates master from non-master) 1. Informed judgement 2. Conceptual Approach 3. Control groups

Establishing A Criterion Cut- Point (cont.) Mastery Level - (Separates master from non master) 1. Informed judgement 2. Conceptual Approach 3. Control groups

Establishing A Criterion Cut- Point (cont.) Mastery Level - (Separates master from non master) 1. Informed judgement 2. Conceptual Approach 3. Control groups

Contrasting group method of cut-off score selection chart. SEE HANDOUT

Section Three: Reliability

I. Types A. Internal Consistency 1. Kuder-Richardson Method. 2. Computer Statistical Package. 3. Problem: Lack of variance. 4. Problem: Excludes items that measure unrelated objectives. B. Test-Retest Score Consistency.

Review Types of Validity: Methods of Establishing Cut-Points 1. Content 1. Informed Judgment 2. Construct 2. Conjecture Method 3. Criterion-related 3. Contrast Group Method Types of Reliability: 1. Test-Retest 2. Internal Consistency 3. Equivalent forms 4. Interrupter reliability

Section Four: Review Questions n Validity cannot exist without reliability. (True or False) n Since CRT relies on judgment rather than normal distribution for scoring, how is reliability assured? n If it becomes necessary for you to establish cut- point for your training program, which of the three methods would you use and why? (Informed judgment, Conjecture method, or Contrast group method)

Norm-Reference Testing GOALS RELIABILITY VALIDITY ADMINISTRATION STANDARD MOTIVATION COMPETITION INSTRUCTIONAL DOMAIN Criterion Referenced Testing Test Achievement Usually High Instruction Dependent Standard Averages-Based Avoidance of Failure Student to Student Low Level Cognitive Test Performance Mastery Usually Unknown Usually High Variable Performance Levels Based Likelihood of Success Student to Criterion Cognitive or Psychomotor