Measuring Research Variables KNES 510 Research Methods in Kinesiology 1.

Slides:



Advertisements
Similar presentations
Chapter 8 Flashcards.
Advertisements

Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Topics: Quality of Measurements
Survey Methodology Reliability and Validity EPID 626 Lecture 12.
The Research Consumer Evaluates Measurement Reliability and Validity
Taking Stock Of Measurement. Basics Of Measurement Measurement: Assignment of number to objects or events according to specific rules. Conceptual variables:
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
© McGraw-Hill Higher Education. All rights reserved. Chapter 3 Reliability and Objectivity.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
VALIDITY AND RELIABILITY
Reliability & Validity.  Limits all inferences that can be drawn from later tests  If reliable and valid scale, can have confidence in findings  If.
Part II Sigma Freud & Descriptive Statistics
Measuring Research Variables
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
Reliability, the Properties of Random Errors, and Composite Scores.
Measurement. Scales of Measurement Stanley S. Stevens’ Five Criteria for Four Scales Nominal Scales –1. numbers are assigned to objects according to rules.
Reliability and Validity of Research Instruments
Part II Knowing How to Assess Chapter 5 Minimizing Error p115 Review of Appl 644 – Measurement Theory – Reliability – Validity Assessment is broader term.
RESEARCH METHODS Lecture 18
Chapter 4 Validity.
RELIABILITY & VALIDITY What is Reliability? What is Reliability?What is Reliability?What is Reliability? How Can We Measure Reliability? How Can We Measure.
VALIDITY.
Chapter 15 Conducting & Reading Research Baumgartner et al Chapter 15 Measurement Issues in Research.
Concept of Measurement
Lecture 7 Psyc 300A. Measurement Operational definitions should accurately reflect underlying variables and constructs When scores are influenced by other.
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Research Methods in MIS
Classroom Assessment A Practical Guide for Educators by Craig A
Reliability of Selection Measures. Reliability Defined The degree of dependability, consistency, or stability of scores on measures used in selection.
Measurement and Data Quality
Reliability, Validity, & Scaling
Instrumentation.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
SELECTION OF MEASUREMENT INSTRUMENTS Ê Administer a standardized instrument Ë Administer a self developed instrument Ì Record naturally available data.
Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand.
Reliability REVIEW Inferential Infer sample findings to entire population Chi Square (2 nominal variables) t-test (1 nominal variable for 2 groups, 1 continuous)
The Basics of Experimentation Ch7 – Reliability and Validity.
Assessing Learners with Special Needs: An Applied Approach, 6e © 2009 Pearson Education, Inc. All rights reserved. Chapter 4:Reliability and Validity.
6. Evaluation of measuring tools: validity Psychometrics. 2012/13. Group A (English)
Correlation & Prediction REVIEW Correlation BivariateDirect/IndirectCause/Effect Strength of relationships (is + stronger than negative?) Coefficient of.
Measurement Validity.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Measurement Issues General steps –Determine concept –Decide best way to measure –What indicators are available –Select intermediate, alternate or indirect.
Measurement MANA 4328 Dr. Jeanne Michalski
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
MEASUREMENT: PART 1. Overview  Background  Scales of Measurement  Reliability  Validity (next time)
Validity and Reliability in Instrumentation : Research I: Basics Dr. Leonard February 24, 2010.
Chapter 6 - Standardized Measurement and Assessment
TEST SCORES INTERPRETATION - is a process of assigning meaning and usefulness to the scores obtained from classroom test. - This is necessary because.
Chapter 6 Norm-Referenced Reliability and Validity.
Measuring Research Variables
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 5 What is a Good Test?
Assessing Student Performance Characteristics of Good Assessment Instruments (c) 2007 McGraw-Hill Higher Education. All rights reserved.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
Chapter 6 Norm-Referenced Measurement. Topics for Discussion Reliability Consistency Repeatability Validity Truthfulness Objectivity Inter-rater reliability.
Data Collection Methods NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN.
Measurement and Scaling Concepts
Ch. 5 Measurement Concepts.
Lecture 5 Validity and Reliability
Reliability and Validity
Questions What are the sources of error in measurement?
Concept of Test Validity
Reliability & Validity
RESEARCH METHODS Lecture 18
Chapter 11: Measuring Research Variables
Chapter 8 VALIDITY AND RELIABILITY
Presentation transcript:

Measuring Research Variables KNES 510 Research Methods in Kinesiology 1

Evidence of Validity The American Educational Research Association and American Psychological Association agree on the definition of four types of validity: Logical ValidityLogical Validity Content ValidityContent Validity Criterion ValidityCriterion Validity –Concurrent Validity –Predictive Validity Construct ValidityConstruct Validity 2

Logical or Face Validity What is logical or face validity?What is logical or face validity? Logical validity is determined subjectivelyLogical validity is determined subjectively What are some examples of logical validity?What are some examples of logical validity? 3

Content Validity What is content validity?What is content validity? A test which has content validity should adequately measure the skills and/or material which has been presented in classA test which has content validity should adequately measure the skills and/or material which has been presented in class 4

Concurrent Validity What is concurrent validity?What is concurrent validity? Determined by correlating scores on a test with scores on a criterion measureDetermined by correlating scores on a test with scores on a criterion measure The resulting correlation coefficient is called a validity coefficientThe resulting correlation coefficient is called a validity coefficient 5

Examples of Concurrent Validity VO 2 max (criterion: oxygen consumption)VO 2 max (criterion: oxygen consumption) –Distance runs (e.g., 1.0-mile, 1.5-mile, 9- minute, 12-minute, 20-minute shuttle) –Submaximal (e.g., cycle, treadmill, swimming) –Nonexercise models (e.g. self-reported physical activity) 6

Body fat (criterion: hydrostatically determined body fat)Body fat (criterion: hydrostatically determined body fat) –Skinfolds –Anthropometric measures Sport skills (criterion: game performance, expert ratings)Sport skills (criterion: game performance, expert ratings) –Sport skills tests 7

Predictive Validity What is predictive validity?What is predictive validity? When, and why, are we interested in determining predictive validity?When, and why, are we interested in determining predictive validity? 8

Examples of Predictive Validity Heart disease (criterion: heart disease developed in later life)Heart disease (criterion: heart disease developed in later life) –Present diet, exercise behaviors, blood pressure, family history Success in graduate school (criterion: grade-point average or graduation status)Success in graduate school (criterion: grade-point average or graduation status) –Graduate Record Examination scores –Undergraduate grade-point average 9

Job capabilities (criterion: successful job performance)Job capabilities (criterion: successful job performance) –Physical abilities –Cognitive abilities Predictive validity is established by correlating the scores on a test with scores on another test in the futurePredictive validity is established by correlating the scores on a test with scores on another test in the future 10

Construct Validity What is construct validity?What is construct validity? Construct validity is used with abstract rather than concrete testsConstruct validity is used with abstract rather than concrete tests An abstract test measures something that is not directly observableAn abstract test measures something that is not directly observable 11

Examples of abstract measures:Examples of abstract measures: –Attitudes –Personality characteristics –Other unobservable yet theoretically existing traits 12

Construct validity is established by finding two groups known to differ on the variable, or construct, being testedConstruct validity is established by finding two groups known to differ on the variable, or construct, being tested The test is then administered to both groups to determine if there is a significant difference between the scores for the two groupsThe test is then administered to both groups to determine if there is a significant difference between the scores for the two groups This is the known group difference methodThis is the known group difference method 13

Reliability Reliability – the degree of consistency with which a test measures what it measuresReliability – the degree of consistency with which a test measures what it measures In order to be valid, a test must also be reliableIn order to be valid, a test must also be reliable Observed score = True score + Error scoreObserved score = True score + Error score 14

Types of Reliability Stability Reliability – the degree to which an individual's scores are unchanged from day to dayStability Reliability – the degree to which an individual's scores are unchanged from day to day We use the test-retest method to obtain the stability reliability coefficientWe use the test-retest method to obtain the stability reliability coefficient 15

Each person is measured with the same test or instrument on several (usually 2) different days (Day 1, Day 2, and so on)Each person is measured with the same test or instrument on several (usually 2) different days (Day 1, Day 2, and so on) The correlation between the two sets of scores is the stability reliability coefficientThe correlation between the two sets of scores is the stability reliability coefficient The closer this coefficient is to positive one (+1.0), the more stable and reliable the scoresThe closer this coefficient is to positive one (+1.0), the more stable and reliable the scores 16

Three factors can contribute to poor score stability (a low stability reliability coefficient): 1.the people tested may perform differently 2.the measuring instrument may operate or be applied differently 3.the person administering the measurement may change 17

Internal-Consistency Reliability – the degree to which an individual's scores are unchanged within a dayInternal-Consistency Reliability – the degree to which an individual's scores are unchanged within a day We use the multiple trials within a day method to obtain the internal-consistency reliability coefficientWe use the multiple trials within a day method to obtain the internal-consistency reliability coefficient 18

To obtain an internal-consistency reliability coefficient, the evaluator must give at least 2 trials of the test within a single dayTo obtain an internal-consistency reliability coefficient, the evaluator must give at least 2 trials of the test within a single day Change in the scores of the people being tested from trial to trial indicates a lack of test reliabilityChange in the scores of the people being tested from trial to trial indicates a lack of test reliability 19

The correlation among the trial scores is the internal-consistency reliability coefficientThe correlation among the trial scores is the internal-consistency reliability coefficient What types of tests should not be evaluated for reliability using this method?What types of tests should not be evaluated for reliability using this method? 20

Stability versus Internal Consistency The internal-consistency reliability coefficient is usually higher than the stability reliability coefficientThe internal-consistency reliability coefficient is usually higher than the stability reliability coefficient With the test-retest method some learning or increase in performance will usually occur, despite the fact that it is presumed that ability will not changeWith the test-retest method some learning or increase in performance will usually occur, despite the fact that it is presumed that ability will not change 21

After completing a test for the first time, subjects will often perform better on the second administrationAfter completing a test for the first time, subjects will often perform better on the second administration This improvement can be referred to as a learning effectThis improvement can be referred to as a learning effect The learning effect is a threat to reliabilityThe learning effect is a threat to reliability How do we avoid this problem?How do we avoid this problem? 22

Methods of Calculating a Reliability Coefficient Pearson’s rPearson’s r Intraclass R from One-Way ANOVAIntraclass R from One-Way ANOVA Cronbach’s AlphaCronbach’s Alpha 23

SPSS Interclass Reliability Output 24

SPSS Output for Intraclass Reliability Analysis (Cronbach’s Alpha) Here is the SPPS output with the value for Cronbach’s alpha:Here is the SPPS output with the value for Cronbach’s alpha: 25

SPSS Output for Intraclass R from One-Way ANOVA Table This ANOVA table may be used to calculate the intraclass RThis ANOVA table may be used to calculate the intraclass R R = ( – 0.333) / = 0.989R = ( – 0.333) / =

Acceptable Reliability R = is below-averageR = is below-average R = is averageR = is average R = is above-averageR = is above-average 27

Intertester Reliability (Objectivity) Objectivity (rater reliability) – the degree to which multiple scorers agree on the magnitude of scoresObjectivity (rater reliability) – the degree to which multiple scorers agree on the magnitude of scores How can the objectivity of a test be improved?How can the objectivity of a test be improved? 28

Next Class Mock proposals 29