Measurement and Scaling

Slides:



Advertisements
Similar presentations
Developing a Questionnaire
Advertisements

Conceptualization and Measurement
Taking Stock Of Measurement. Basics Of Measurement Measurement: Assignment of number to objects or events according to specific rules. Conceptual variables:
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
VALIDITY AND RELIABILITY
Chapter 5 Measurement, Reliability and Validity.
4/25/2015 Marketing Research 1. 4/25/2015Marketing Research2 MEASUREMENT  An attempt to provide an objective estimate of a natural phenomenon ◦ e.g.
Part II Sigma Freud & Descriptive Statistics
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
Part II Sigma Freud & Descriptive Statistics
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
CH. 9 MEASUREMENT: SCALING, RELIABILITY, VALIDITY
5/15/2015Marketing Research1 MEASUREMENT  An attempt to provide an objective estimate of a natural phenomenon ◦ e.g. measuring height ◦ or weight.
Measurement. Scales of Measurement Stanley S. Stevens’ Five Criteria for Four Scales Nominal Scales –1. numbers are assigned to objects according to rules.
Reliability and Validity of Research Instruments
Experiment Basics: Variables Psych 231: Research Methods in Psychology.
RESEARCH METHODS Lecture 18
MEASUREMENT. Measurement “If you can’t measure it, you can’t manage it.” Bob Donath, Consultant.
Concept of Measurement
Beginning the Research Design
1 Measurement PROCESS AND PRODUCT. 2 MEASUREMENT The assignment of numerals to phenomena according to rules.
Psych 231: Research Methods in Psychology
1 Measurement Measurement Rules. 2 Measurement Components CONCEPTUALIZATION CONCEPTUALIZATION NOMINAL DEFINITION NOMINAL DEFINITION OPERATIONAL DEFINITION.
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Variables cont. Psych 231: Research Methods in Psychology.
Research Methods in MIS
Measurement and Data Quality
Reliability, Validity, & Scaling
Experimental Research
MEASUREMENT OF VARIABLES: OPERATIONAL DEFINITION AND SCALES
Measurement in Exercise and Sport Psychology Research EPHE 348.
Instrumentation.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
SELECTION OF MEASUREMENT INSTRUMENTS Ê Administer a standardized instrument Ë Administer a self developed instrument Ì Record naturally available data.
Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand.
Chapter Eight The Concept of Measurement and Attitude Scales
Chapter Nine
The Basics of Experimentation Ch7 – Reliability and Validity.
Chapter Five Measurement Concepts. Terms Reliability True Score Measurement Error.
Reliability & Validity
Counseling Research: Quantitative, Qualitative, and Mixed Methods, 1e © 2010 Pearson Education, Inc. All rights reserved. Basic Statistical Concepts Sang.
Chapter 7 Measurement and Scaling Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
Learning Objective Chapter 9 The Concept of Measurement and Attitude Scales Copyright © 2000 South-Western College Publishing Co. CHAPTER nine The Concept.
Measurement and Questionnaire Design. Operationalizing From concepts to constructs to variables to measurable variables A measurable variable has been.
Chapter 2: Behavioral Variability and Research Variability and Research 1. Behavioral science involves the study of variability in behavior how and why.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
SOCW 671: #5 Measurement Levels, Reliability, Validity, & Classic Measurement Theory.
Psychometrics. Goals of statistics Describe what is happening now –DESCRIPTIVE STATISTICS Determine what is probably happening or what might happen in.
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
Measurement Experiment - effect of IV on DV. Independent Variable (2 or more levels) MANIPULATED a) situational - features in the environment b) task.
Chapter 6 - Standardized Measurement and Assessment
Measurement Chapter 6. Measuring Variables Measurement Classifying units of analysis by categories to represent variable concepts.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
© 2009 Pearson Prentice Hall, Salkind. Chapter 5 Measurement, Reliability and Validity.
Measurement and Scaling Concepts
1 Measurement Error All systematic effects acting to bias recorded results: -- Unclear Questions -- Ambiguous Questions -- Unclear Instructions -- Socially-acceptable.
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Chapter 2 Theoretical statement:
Lecture 5 Validity and Reliability
Product Reliability Measuring
Associated with quantitative studies
CHAPTER 5 MEASUREMENT CONCEPTS © 2007 The McGraw-Hill Companies, Inc.
Reliability, validity, and scaling
RESEARCH METHODS Lecture 18
Measurement Concepts and scale evaluation
Ch 5: Measurement Concepts
The Concept of Measurement and Attitude Scales
Presentation transcript:

Measurement and Scaling Farzin Madjidi, Ed.D. Pepperdine University Graduate School of Education and Psychology

Variables Independent Dependent Extraneous Precedes, influences or predicts results Dependent Affected by or predicted by the Independent Variable Extraneous Affected by the D.V., but not controlled or measured. Causes error

Variables Confounding Intervening An extraneous variable that varies systematically (has a relationship) with the I.V. Intervening Unobservable trait that influences behavior (e.g., effect of new intervention on self-esteem may be affected by the motivation level of subjects)

Variables Control Organismic Used to eliminate the effect of extraneous variables Organismic Aka, measured, or assigned Characteristics of the subjects that cannot be manipulated

Levels of Measurements Four levels of Measurements Nominal Measures categories Ordinal Categories + rank and order Interval Equal distance between any two consecutive measures Ratio Intervals + meaningful zeros

Categories of Scales Categorical (ratings) Comparative (ranking) Score without comparison - 1 to 5 scales Comparative (ranking) Score by comparing - Smartest Preference Subjective - which do you prefer Non-preference Objective - which solution is less costly

Categories of Scales Unidimensional Multi-dimensional Involves only one aspect of the measurement Measurement by one construct Multi-dimensional Involves several aspects of a measurement Uses several dimensions to measure a single construct

Types of Scales Likert/Summated Rating Scales Semantic Differential Scales Magnitude Scaling Thruston Scales Guttman Scales

Likert Scales A very popular rating scale Measures the feelings/degree of agreement of the respondents Ideally, 4 to 7 points Examples of 5-point surveys Agreement SD D ND/NA A SA Satisfaction SD D ND/NS S SS Quality VP P Average G VG

Summative Ratings A number of items collectively measure one construct (Job Satisfaction) A number of items collectively measure a dimension of a construct and a collection of dimensions will measure the construct (Self-esteem)

Summative Likert Scales Must contain multiple items Each individual item must measure something that has an underlying, quantitative measurement continuum There can be no right/wrong answers as opposed to multiple-choice questions Items must be statements to which the respondent assigns a rating Cannot be used to measure knowledge or ability, but familiarity

Semantic Differential Scales Uses a set of scale anchored by their extreme responses using words of opposite meaning. Example: Dark ___ ___ ___ ___ ___ Light Short ___ ___ ___ ___ ___ Tall Evil ___ ___ ___ ___ ___ Good Four to seven categories are ideal

Magnitude Scaling Attempts to measure constructs along a numerical, ratio level scale Respondent is given an item with a pre-assigned numerical value attached to it to establish a “norm” The respondent is asked to rate other items with numerical values as a proportion of the “norm” Very powerful if reliability is established

Thurston Scales Thurston Scales Items are formed Panel of experts assigns values from 1 to 11 to each item Mean or median scores are calculated for each item Select statements evenly spread across the scale

Thurston Scales Example: Please check the item that best describes your level of willingness to try new tasks I seldom feel willing to take on new tasks (1.7) I will occasionally try new tasks (3.6) I look forward to new tasks (6.9) I am excited to try new tasks (9.8)

Guttman Scales Also known as Scalograms Both the respondents and items are ranked Cutting points are determined (Goodenough-Edwards technique) Coefficient of Reproducibility (CReg) - a measure of goodness of fit between the observed and predicted ideal response patterns Keep items with CReg of 0.90 or higher

Scale Construction Define Constructs Conceptual/theoretical basis from the literature Are their sub-scales (dimensions) to the scale Multiple item sub-scales Principle of Parsimony Simplest explanation among a number of equally valid explanations must be used

Item Construction Agreement items Write declarative statements Death penalty should be abolished I like to listen to classical music Frequency items (how often) I like to read Evaluation items How well did your team play How well does the police serve your community

Item Writing Mutually exclusive and collectively exhaustive items Use positively and negatively phrased questions Avoid colloquialism, expressions and jargon Avoid the use of negatives to reverse the wording of an item Don’t use: I am not satisfied with my job Use: I hate my job! Be brief, focused, and clear Use simple, unbiased questions

Sources of Error Social desirability Response sets Acquiescence Giving politically correct answers Response sets All yes, or all no responses Acquiescence Telling you what you want to hear Personal bias Wants to send a message

Sources of Error Response order Item order Recency - Respondent stops reading once s/he gets to the response s/he likes Primacy - Remember better the initial choices Fatigue Item order Answers to later items may be affected by earlier items (simple, factual items first) Respondent may not know how to answer earlier questions

Assessing Instruments Three issues to consider Validity: Does the instrument measure what its supposed to measure Reliability: Does it consistently repeat the same measurement Practicality: Is this a practical instrument

Types of Validity Face validity Content validity Does the instrument, on its face, appear to measure what it is supposed to measure Content validity Degree to which the content of the items adequately represent the universe of all relevant items under study Generally arrived at through a panel of experts

Types of Validity Criterion related Degree to which the predictor is adequate in capturing the relevant aspects of criterion Uses Correlation analysis Concurrent validity Criterion data is available at the same time as predictor score- requires high correlation between the two Predictive validity Criterion is measured after the passage of time Retrospective look at the validity of the measurement Known-groups

Types of Validity Construct Validity Measures what accounts for the variance Attempts to identify the underlying constructs Techniques used: Correlation of proposed test with other existing tests Factor analysis Multi-trait-multimethod analysis Convergent validity - Calls for high correlation between the different measures of the same construct Discriminant validity - Calls for low correlation between sub-scales within a construct

Types of Reliability Stability Test-retest: Same test is administered twice to the same subjects over a short interval (3 weeks to 6 months) Look for high correlation between the test and retest Situational factors must be minimized

Types of Reliability Equivalence Degree to which alternative forms of the same measure produce same or similar results Give parallel forms of the same test to the same group with a short delay to avoid fatigue Look for high correlation between the scores of the two forms of the test Inter-rater reliability

Types of Reliability Internal Consistency Degree to which instrument items are homogeneous and reflect the same underlying constructs Split-half testing where the test is split into two halves that contain the same types of questions Uses Cronbach’s alpha to determine internal consistency. Only one administration of the test is required Kuder-Richardson (KR20) for items with right and wrong answers

Practicality Is the survey economical Convenience Cost of producing and administering the survey Time requirement Common sense! Convenience Adequacy of instructions Easy to administer Can the measurement be interpreted by others Scoring keys Evidence of validity and reliability Established norms