Now that you know what assessment is, you know that it begins with a test. Ch 4.

Slides:



Advertisements
Similar presentations
Chapter 8 Flashcards.
Advertisements

Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Survey Methodology Reliability and Validity EPID 626 Lecture 12.
The Research Consumer Evaluates Measurement Reliability and Validity
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
© McGraw-Hill Higher Education. All rights reserved. Chapter 3 Reliability and Objectivity.
VALIDITY AND RELIABILITY
Research Methodology Lecture No : 11 (Goodness Of Measures)
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
Assessment: Reliability, Validity, and Absence of bias
RESEARCH METHODS Lecture 18
Concept of Measurement
Formative and Summative Evaluations
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 5 Making Systematic Observations.
Introduction to Measurement and Evaluation
Measurement Concepts & Interpretation. Scores on tests can be interpreted: By comparing a client to a peer in the norm group to determine how different.
Methods of Psychology Hypothesis: A tentative statement about how or why something happens. e.g. non experienced teachers use corporal punishment more.
Measurement and Data Quality
Validity and Reliability
Descriptive and Causal Research Designs
Instrumentation.
Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
McMillan Educational Research: Fundamentals for the Consumer, 6e © 2012 Pearson Education, Inc. All rights reserved. Educational Research: Fundamentals.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
CSD 5100 Introduction to Research Methods in CSD Observation and Data Collection in CSD Research Strategies Measurement Issues.
Review: Cognitive Assessments II Ambiguity (extrinsic/intrinsic) Item difficulty/discrimination relationship Questionnaires assess opinions/attitudes Open-/Close-ended.
Research Seminars in IT in Education (MIT6003) Research Methodology I Dr Jacky Pow.
EDU 8603 Day 6. What do the following numbers mean?
SHOWTIME! NATURE AND ADMINISTRATION OF TESTS TEST CHARACTERISTICS RELIABILITY, OBJECTIVITY, AND VALIDITY CONTENT-RELATED ATTRIBUTES STUDENT AND PARTICIPANT.
Introduction Gathering Information Observation Interviewing Norm Referenced Tools Authentic Assessment Characteristics of Authentic Assessment – 7M’s Validity.
Research in Communicative Disorders1 Research Design & Measurement Considerations (chap 3) Group Research Design Single Subject Design External Validity.
MEASUREMENT: SCALE DEVELOPMENT Lu Ann Aday, Ph.D. The University of Texas School of Public Health.
Unit 5: Improving and Assessing the Quality of Behavioral Measurement
Selecting a Sample. Sampling Select participants for study Select participants for study Must represent a larger group Must represent a larger group Picked.
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Assessment and Testing
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
JS Mrunalini Lecturer RAKMHSU Data Collection Considerations: Validity, Reliability, Generalizability, and Ethics.
Psychometrics. Goals of statistics Describe what is happening now –DESCRIPTIVE STATISTICS Determine what is probably happening or what might happen in.
Review: Alternative Assessments Alternative/Authentic assessment Real-life setting Performance based Techniques: Observation Individual or Group Projects.
1 LANGUAE TEST RELIABILITY. 2 What Is Reliability? Refer to a quality of test scores, and has to do with the consistency of measures across different.
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Reliability Ability to produce similar results when repeated measurements are made under identical conditions. Consistency of the results Can you get.
Measurement Experiment - effect of IV on DV. Independent Variable (2 or more levels) MANIPULATED a) situational - features in the environment b) task.
Chapter 6 - Standardized Measurement and Assessment
Testing. Psychological Tests  Tests abilities, interests, creativity, personality, behavior  Must be standardized, reliable, and valid  Timing, instructions,
Review: Cognitive Assessments II Ambiguity (extrinsic/intrinsic) Item difficulty/discrimination relationship Questionnaires assess opinions/attitudes Open-/Close-ended.
CIV Fitness/S&C Steven Tikkanen – F129 1 Sutherland College Health & Recreation Semester Version 1.
Assessment of Sport Skills and Motor Abilities
Review: Performance-Based Assessments Performanc-based assessment Real-life setting H.O.T.S. Techniques: Observation Individual or Group Projects Portfolios.
 Characteristics or conditions that change or have different values for different individuals  Age  Gender  Score  Elapsed Time.
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 5 What is a Good Test?
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
Measurement and Scaling Concepts
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Copyright © 2009 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 47 Critiquing Assessments.
Survey Methodology Reliability and Validity
Concept of Test Validity
RELIABILITY OF QUANTITATIVE & QUALITATIVE RESEARCH TOOLS
پرسشنامه کارگاه.
RESEARCH METHODS Lecture 18
Chapter 4 Characteristics of a Good Test
Chapter 8 VALIDITY AND RELIABILITY
Presentation transcript:

Now that you know what assessment is, you know that it begins with a test. Ch 4

How can you make sure your test is trustworthy? Will it’s measurement be true? If not, what does that say about your evaluation and thus your assessment? Questions to ask yourself:

Your Test Must Have Sensitivity Reliability Objectivity Validity To Be a Good Test

Sensitivity The ability to detect a true difference.

Sensitivity A measurement device or tool….. –Used to apply a number or value - scale –Indicates and separates differences – scale –You can only be sure of your measurement when it fits the scale –Potential for error is high when sensitivity is poor. –Examples

What Happens When Sensitivity is Poor? Too High? –Detect everything –Unable to discriminate between what you’re searching for and everything else you find –Example: Signal vs. signal To Low? –Can’t detect enough –Unable to discriminate between what you’re looking for and nothing –Example: Signal vs. noise or signal vs. silence

Either case, the result is an incorrect result. Incorrect result leads to incorrect judgment. Incorrect judgment leads to incorrect intervention.

Sensitivity True False Measurement Result True State False Positive True Negative False Negative True Positive

Reliability Reliability refers to the consistency of a test –A test given on one day should yield the same results on the next day

Can One Test Trial Give Reliable Scores?

Selecting a Criterion Score Mean Score vs. Best Score –Take the average score of multiple trials –Take the best score of multiple trials

Selecting a Norm-Referenced Score Comparison against… –Individuals Best example? –Tabled Values Match the group?

Types of Reliability Stability –Use Test-Retest Method Two administrations of a test with a calculation of how similar they are

Types of Reliability Internal Consistency –Use Split Half Method One group completes odd numbered components (or questions) while the other completes evens with a calculation of how similar they are

Factors Affecting Reliability Method of scoring – objective format is more reliable than subjective format Homogeneity of the group tested – more alike = reliable Length of test – longer tests are more reliable than shorter tests (number of questions/elements) Administrative procedures – clear directions, technique, motivation of subjects, good environment

Objectivity Objectivity is also referred to as rater reliability Objectivity is the close agreement between scores assigned by two or more judges Juggler please….

Judges’ scores………

Factors Affecting Objectivity The clarity of the scoring system Neatness = 37% x difficulty factor (.87 / prime rate) Style = – (IQ x shoe size) Originality = 3.14 x (E = MC 2 ) / date of birth

Factors Affecting Objectivity The degree to which judges can assign scores accurately (fairly, no bias) –Can you say, “French Skating Judge” eh?

Validity Validity is considered the most important characteristic of a test

Types of Validity Logical (content) Validity – validity that indicates a measurement instrument (test) actually measures the capacities about which conclusions are drawn –Examples: Tennis serve for accuracy versus 40 yd. run for time/speed in football or baseball

Types of Validity Construct Validity - validity for an instrument that measures a variable or factor that is not directly observable –Examples: Attitudes, feelings, motivation Learning new skills

Factors Affecting Validity Student characteristics –Examples: sex, age, experience Administrative procedures –Examples: unclear directions, poor weather Reliability –Example: same results after repeated test

When a A Test is Not Valid….. It is a waste of everyone’s time. A true judgment (evaluation) cannot be made. So….what’s the point?

A True Professional Values his/her field. Seeks the truth. Knows that bad data are worse than no data Wastes no one’s time.

Summary A Test Must Have Sensitivity Reliability Objectivity Validity To Be a Good Test

Now to the testing process….

Tests and Their Administration Test Characteristics Administrative Concerns Participant Concerns Pre Test Procedures Administering the Test Post Test Procedures Individuals with Disabilities

Tests and Their Administration Test Characteristics Validity, Reliability, Objectivity Content Related Attributes –Important Attributes – a sample, not everything –Discrimination – best from better from good –Resemblance to the Activity - similarity –Specificity – single attribute vs. multiple attributes, ID’s limitations –Unrelated Measures – battery - different aspects – “independence”

Test Administration Administrative Concerns Mass Testability Minimal Practice Minimal Equipment Minimal Personnel Ease of Preparation Adequate Directions Norms / Criterion Useful Scores

Test Administration Participant Concerns Appropriateness to Participants Individual Scores Enjoyable Safety Confidentiality and Privacy Motivation

Test Administration Pretest Procedures -Knowing the Test -Developing Test Procedures -Developing Directions -Preparing the Participants -Planning Warm-up and Test Trials -Equipment and Facilities -Score Sheet (Recording) -Estimating Time Requirements

Test Administration Administering the Test Preparation Motivation Safety Recording Test Results

Test Administration Posttest Procedures Analyze Test Scores “Share” Information – highs vs. lows

Test Administration Individuals with Disabilities Cannot be compared to the typical student/person Few norm-referenced standards for evaluation Disabilities are varied – even within categories

Test Administration Individuals with Disabilities Disabilities Include: Mental Retardation – Mild, Moderate, Severe Serious Emotional Disturbance – Autism, Schizophrenia Orthopedic Impairment – Neurological, Musculoskeletal, Trauma Other Health Impaired – Asthma, Cardiovascular Disorders, Diabetes, Epilepsy, Obesity

Test Administration Individuals with Disabilities –Measure Physical Ability or Capacity Not learning, cognition, or language acquisition Find norm-referenced standards or validated tests

Summary Tests and Their Administration –Test Characteristics –Administrative Concerns –Participant Concerns –Pre Test Procedures –Administering the Test –Post Test Procedures –Individuals with Disabilities

Questions?

End Ch 4