Quiz Do random errors accumulate? Name 2 ways to minimize the effect of random error in your data set.

Slides:



Advertisements
Similar presentations
Evaluating and Institutionalizing
Advertisements

Agenda Levels of measurement Measurement reliability Measurement validity Some examples Need for Cognition Horn-honking.
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
How good are our measurements? The last three lectures were concerned with some basics of psychological measurement: What does it mean to quantify a psychological.
Conceptualization and Measurement
The Research Consumer Evaluates Measurement Reliability and Validity
Some (Simplified) Steps for Creating a Personality Questionnaire Generate an item pool Administer the items to a sample of people Assess the uni-dimensionality.
Reliability and Validity
Reliability for Teachers Kansas State Department of Education ASSESSMENT LITERACY PROJECT1 Reliability = Consistency.
Validity In our last class, we began to discuss some of the ways in which we can assess the quality of our measurements. We discussed the concept of reliability.
Reliability, the Properties of Random Errors, and Composite Scores.
CH. 9 MEASUREMENT: SCALING, RELIABILITY, VALIDITY
Data Analysis Outline What is an experiment? What are independent vs. dependent variables in an experimental study? What are our dependent measures/variables.
Measurement  The process whereby individual instances within a defined population are scored on an attribute according to rules Usually given a numeric.
Different ways to think about validity To the extent that a measure has validity, we can say that it measures what it is supposed to measure. There are.
RESEARCH METHODS Lecture 18
Validity, Sampling & Experimental Control Psych 231: Research Methods in Psychology.
MGTO 231 Human Resources Management Personnel selection I Dr. Kin Fai Ellick WONG.
Sampling Basics 2011, 9, 13. Last Class: Measurement  Scale of measurement –Nominal scale –Ordinal scale –Interval-ratio scale  Reliability: Free of.
Psych 231: Research Methods in Psychology
Variables cont. Psych 231: Research Methods in Psychology.
Validity, Reliability, & Sampling
Reliability & Validity the Bada & Bing of YOUR tailored survey design.
Quiz Name one latent variable Name 2 manifest variables that are indicators for the latent variable.
Introduction to the design (and analysis) of experiments James M. Curran Department of Statistics, University of Auckland
Social Science Research Design and Statistics, 2/e Alfred P. Rovai, Jason D. Baker, and Michael K. Ponton Internal Consistency Reliability Analysis PowerPoint.
Reliability, Validity, & Scaling
Experimental Research
Measurement in Exercise and Sport Psychology Research EPHE 348.
Student Engagement Survey Results and Analysis June 2011.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
Goals for Today Review the basics of an experiment Learn how to create a unit-weighted composite variable and how/why it is used in psychology. Learn how.
Multiple linear indicators A better scenario, but one that is more challenging to use, is to work with multiple linear indicators. Example: Attraction.
Creating Assessments AKA how to write a test. Creating Assessments All good assessments have three key features: All good assessments have three key features:
Final Reports from the Measures of Effective Teaching Project Tom Kane Harvard University Steve Cantrell, Bill & Melinda Gates Foundation.
Teacher Engagement Survey Results and Analysis June 2011.
Chapter Thirteen Measurement Winston Jackson and Norine Verberg Methods: Doing Social Research, 4e.
Validity In our last class, we began to discuss some of the ways in which we can assess the quality of our measurements. We discussed the concept of reliability.
Multiple linear indicators A better scenario, but one that is more challenging to use, is to work with multiple linear indicators. Example: Attraction.
Reliability, the Properties of Random Errors, and Composite Scores Week 7, Psych R. Chris Fraley
Chapter 2: Behavioral Variability and Research Variability and Research 1. Behavioral science involves the study of variability in behavior how and why.
Chapter 10: Measurement Slavin Reliability & Validity Types of Measures Behavioral Observation Systems Sample Size Determination.
Psychology 3051 Psychology 305: Theories of Personality Lecture 2.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Measurement Issues General steps –Determine concept –Decide best way to measure –What indicators are available –Select intermediate, alternate or indirect.
Measurement Theory in Marketing Research. Measurement What is measurement?  Assignment of numerals to objects to represent quantities of attributes Don’t.
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Discussion Overview: Measurement I) Reliability of Measures I) Reliability of Measures II) Construct Validity II) Construct Validity III) Measurement scales.
Reliability: Introduction. Reliability Session 1.Definitions & Basic Concepts of Reliability 2.Theoretical Approaches 3.Empirical Assessments of Reliability.
Psychology 3051 Psychology 305: Theories of Personality Lecture 2.
Reliability: Introduction. Reliability Session Definitions & Basic Concepts of Reliability Theoretical Approaches Empirical Assessments of Reliability.
Measurement Experiment - effect of IV on DV. Independent Variable (2 or more levels) MANIPULATED a) situational - features in the environment b) task.
Psychology 3051 Psychology 305A: Theories of Personality Lecture 2 1.
Project VIABLE - Direct Behavior Rating: Evaluating Behaviors with Positive and Negative Definitions Rose Jaffery 1, Albee T. Ongusco 3, Amy M. Briesch.
VALIDITY What is validity? What are the types of validity? How do you assess validity? How do you improve validity?
Ch. 5 Measurement Concepts.
Reliability and Validity in Research
RELIABILITY OF QUANTITATIVE & QUALITATIVE RESEARCH TOOLS
پرسشنامه کارگاه.
5. Reliability and Validity
Questionnaire Reliability
Assessment of Personality
Scoring: Measures of Central Tendency
RESEARCH METHODS Lecture 18
Personality An individual’s characteristic pattern of thinking, feeling, and acting.
Reliability, the Properties of Random Errors, and Composite Scores
Psychological Measurement: Reliability and the Properties of Random Errors The last two lectures were concerned with some basics of psychological measurement:
Measurement Concepts and scale evaluation
Presentation transcript:

Quiz Do random errors accumulate? Name 2 ways to minimize the effect of random error in your data set.

Validity In our last class, we began to discuss some of the ways in which we can assess the quality of our measurements. We discussed the concept of reliability (i.e., the degree to which measurements are free of random error).

Why reliability alone is not enough Understanding the degree to which measurements are reliable, however, is not sufficient for evaluating their quality.

Reliability – Precision - Grouping

Validity In this example, the measurements appear reliable, but there is a problem... Validity reflects the degree to which measurements are free of both random error, E, and systematic error, S. O = T + E + S Systematic errors reflect the influence of any non-random factor beyond what we’re attempting to measure.

Validity: Does systematic error accumulate? Question: If we sum or average multiple observations (i.e., using a multiple indicators approach), how will systematic errors influence our estimates of the “true” score?

Validity: Does error accumulate? Answer: Unlike random errors, systematic errors accumulate. Systematic errors exert a constant source of influence on measurements. We will always overestimate (or underestimate) T if systematic error is present.

Note: Each measurement is 2 points higher than the true value of 10. The errors do no average out.

Note: Even when random error is present, E averages to 0 but S does not. Thus, we have reliable measures that have validity problems.

Validity: Ensuring validity What can we do to minimize the impact of systematic errors? One way to minimize their impact is to use a variety of indicators Different kinds of indicators of a latent variable may not share the same systematic errors If true, then S will behave like random error across measurements (but not within measurements)

Example As an example, let’s consider the measurement of self-esteem. –Some methods, such as self-report questionnaires, may lead people to over-estimate their self-esteem. Most people want to think highly of themselves. –Other methods, such as clinical ratings by trained observers, may lead to under-estimates of self-esteem. Clinicians, for example, may be prone to assume that people are not as well-off as they say they are.

Note: Method 1 systematically overestimates T whereas Method 2 systematically underestimates T. In combination, however, those systematic errors cancel out. Clinical ratings Self- reports

Another example One problem with the use of self-report questionnaire rating scales is that some people tend to give high (or low) answers consistently (i.e., regardless of the question being asked).

ItemTSO I think I am a worthwhile person I have high self-esteem.4+15 I am confident in my ability to meet challenges in life My friends and family value me as a person Average score:4+15 In this example, we have someone with relatively high self- esteem, but this person systematically rates questions one point higher than he or she should. 1 = strongly disagree | 5 = strongly agree

ItemTSO I think I am a worthwhile person I have high self-esteem.4+15 I am NOT confident in my ability to meet challenges in life My friends and family DO NOT value me as a person Average score:4+14 If we “reverse key” half of the items, the bias averages out. 1 = strongly disagree | 5 = strongly agree Responses to reverse keyed items are counted in the opposite direction. 1(strongly disagree) + 5(strongly agree) = 6 T: ( [6-2] + [6-2]) / 4 = 4 O: ( [6-3] + [6-3]) / 4 = 4

Validity To the extent to which a measure has validity, we say that it measures what it is supposed to measure Question: How do you assess validity? ** Very tough question to answer! ** (But, we’ll give it a shot in our next class.)