Conceptualization and measurement

Slides:



Advertisements
Similar presentations
Allyn & Bacon 2003 Social Work Research Methods: Qualitative and Quantitative Approaches Topic 7: Basics of Measurement Examine Measurement.
Advertisements

Standardized Scales.
Conceptualization, Operationalization, and Measurement
Conceptualization and Measurement
The Research Consumer Evaluates Measurement Reliability and Validity
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
Professor Gary Merlo Westfield State College
Part II Sigma Freud & Descriptive Statistics
Part II Sigma Freud & Descriptive Statistics
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Measurement Reliability and Validity
Designing Research Concepts, Hypotheses, and Measurement
MEASUREMENT. Measurement “If you can’t measure it, you can’t manage it.” Bob Donath, Consultant.
Concept of Measurement
Beginning the Research Design
Research Methods in MIS
1 Measurement Measurement Rules. 2 Measurement Components CONCEPTUALIZATION CONCEPTUALIZATION NOMINAL DEFINITION NOMINAL DEFINITION OPERATIONAL DEFINITION.
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Social Research Methods: Qualitative and Quantitative Approaches, 5e
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
Classroom Assessment A Practical Guide for Educators by Craig A
Measuring Social Life Ch. 5, pp
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT OSMAN BIN SAIF Session 14.
The Practice of Social Research
Measurement and Data Quality
Slide 9-1 © 1999 South-Western Publishing McDaniel Gates Contemporary Marketing Research, 4e Understanding Measurement Carl McDaniel, Jr. Roger Gates Slides.
MEASUREMENT OF VARIABLES: OPERATIONAL DEFINITION AND SCALES
Chapter 1: Introduction to Statistics
Chapter 3 An Overview of Quantitative Research
Instrumentation.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Copyright © 2008 by Nelson, a division of Thomson Canada Limited Chapter 11 Part 3 Measurement Concepts MEASUREMENT.
Chapter Eight The Concept of Measurement and Attitude Scales
Technical Adequacy Session One Part Three.
Chapter 4 Conceptualization and Measurement
WELNS 670: Wellness Research Design Chapter 5: Planning Your Research Design.
6. Conceptualization & Measurement
Assumes that events are governed by some lawful order
URBDP 591 I Lecture 3: Research Process Objectives What are the major steps in the research process? What is an operational definition of variables? What.
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 1. The Statistical Imagination.
Learning Objective Chapter 9 The Concept of Measurement and Attitude Scales Copyright © 2000 South-Western College Publishing Co. CHAPTER nine The Concept.
Selecting a Sample. Sampling Select participants for study Select participants for study Must represent a larger group Must represent a larger group Picked.
1 The Theoretical Framework. A theoretical framework is similar to the frame of the house. Just as the foundation supports a house, a theoretical framework.
Question paper 1997.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
The Theory of Sampling and Measurement. Sampling First step in implementing any research design is to create a sample. First step in implementing any.
SOCW 671: #5 Measurement Levels, Reliability, Validity, & Classic Measurement Theory.
 Measuring Anything That Exists  Concepts as File Folders  Three Classes of Things That can be Measured (Kaplan, 1964) ▪ Direct Observables--Color of.
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
12-1 Chapter 12 Measurement Learning Objectives Understand... distinction between measuring objects, properties, and indicants of properties similarities.
Sampling Design & Measurement Scaling
Chapter7 & 8 Experimental Design & Measurement of variables Adnan Khurshid.
RESEARCH METHODS IN INDUSTRIAL PSYCHOLOGY & ORGANIZATION Pertemuan Matakuliah: D Sosiologi dan Psikologi Industri Tahun: Sep-2009.
Measurement Experiment - effect of IV on DV. Independent Variable (2 or more levels) MANIPULATED a) situational - features in the environment b) task.
Chapter 6 - Standardized Measurement and Assessment
Lesson 3 Measurement and Scaling. Case: “What is performance?” brandesign.co.za.
Measurement Chapter 6. Measuring Variables Measurement Classifying units of analysis by categories to represent variable concepts.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
Data Collection Methods NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN.
Measurement and Scaling Concepts
Chapter 3 Designing Research Concepts, Hypotheses, and Measurement.
Chapter 2 Theoretical statement:
Reliability and Validity in Research
Associated with quantitative studies
Introduction to Measurement
Research Methods: Concepts and Connections First Edition
Measuring Social Life: How Many? How Much? What Type?
Reliability and Validity of Measurement
Presentation transcript:

Conceptualization and measurement Chapter 4 Conceptualization and measurement

Every time you begin to review or design a research study, you will have to answer two questions: What do the main concepts mean in this research? (2) How are the main concepts measured? Both questions must be answered to evaluate the validity of any research.

Conceptualization A continuing challenge for social scientists rests on the fact that many of our important topics of study (social health, for instance) are not clearly defined things or objects (like trees or rocks), but are abstract concepts or ideas. A concept is an image or idea, not a simple object. It is especially important to define clearly concepts that are abstract or unfamiliar.

Conceptualization is the process of specifying what we mean by a term. In deductive research, conceptualization helps to translate portions of an abstract theory into specific variables that can be used in testable hypotheses. In inductive research, conceptualization is an important part of the process used to make sense of related observations.

Variables and constants After we define the concepts for a study, we must identify variables that correspond to those concepts. The term variable is used to refer to some specific aspect of a concept that varies, and for which we then have to select even more concrete indicators. Not every concept in a particular study is represented by a variable. Some are called a constant (it’s always the same), not a variable.

How do we know what concepts to consider and then which variables to include in a study? Examine the theories that are relevant to our research question to identify those concepts that would be expected to have some bearing on the phenomena we are investigating. Review the relevant research literature and assess the utility of variables used in prior research.

Consider the constraints and opportunities for measurement that are associated with the specific setting(s) we will study. Distinguish constants from variables in this setting. Look ahead to our analysis of the data. What role will each variable play in our analyses?

How Will We Know When We’ve Found It? Operationalization is the process of connecting concepts to observations. It is the process of specifying the operations that will indicate the value of cases on a variable. You can think of it as the empirical counterpart of the process of conceptualization. The goal is to devise operations that actually measure the concepts we intend to measure—in other words, to achieve measurement validity.

Insert exhibit 4.1

Using available data Government reports are rich, accessible sources of social science data. The data collected in many social science surveys are archived and made available for researchers who were not involved in the original survey project. We also cannot assume that available data are accurate, even when they appear to measure the concept. Another rich source of already-collected data is survey datasets archived and made available to university researchers by the Inter-University Consortium for Political and Social Research.

Constructing questions Asking people questions is the most common and probably the most versatile operation for measuring social variables. Most concepts about individuals can be defined in such a way that measurement with one or more questions becomes an option. Of course, in spite of the fact that questions are, in principle, a straightforward and efficient means to measure individual characteristics, facts about events, level of knowledge, and opinions of any sort, can easily result in misleading or inappropriate answers.

Different types of questions used in social research Measuring variables with single questions is very popular. Public opinion polls based on answers to single questions are reported frequently in newspaper articles and TV newscasts. Single questions can be designed with or without explicit response choices.

Closed-ended (fixed-choice) question: a survey question that provides preformatted response choices for the respondent to circle or check. Open-ended question: a survey question to which the respondent replies in his or her own words, either by writing or by talking.

Indexes and scales When several questions are used to measure one concept, the responses may be combined by taking the sum or average of responses. A composite measure based on this type of sum or average is termed an index. Because of the popularity of survey research, indexes already have been developed to measure many concepts. In a scale, we give different weights to the responses to different questions before summing or averaging the responses.

Insert exhibit 4.2

Making observations Observations can be used to measure characteristics of individuals, events, and places. The observations may be the primary form of measurement in a study, or they may supplement measures obtained through questioning. Direct observations can be used as indicators of some concepts. Observations may also supplement data collected in an interview study.

Content analysis Content analysis is a research method for systematically analyzing and making inferences from text. You can think of a content analysis as a “survey” of documents ranging from newspapers, books, or TV shows to persons referred to in other communications, themes expressed in government documents, or propositions made in tape-recorded debates. Computer programs for content analysis can be used to enhance reliability.

Collecting unobtrusive measures Unobtrusive measures allow us to collect data about individuals or groups without their direct knowledge or participation. Eugene Webb and his colleagues (2000) identified four types of unobtrusive measures: physical trace evidence, archives (available data), simple observation, and contrived observation (using hidden recording hardware or manipulation to elicit a response).

Unobtrusive measures can also be created from such diverse forms of media as newspaper archives or magazine articles, TV or radio talk shows, legal opinions, historical documents, personal letters, or e-mail messages.

Combining measurement operations Using available data, asking questions, making observations, and using unobtrusive indicators are interrelated measurement tools, each of which may include or be supplemented by the others. Researchers may use insights gleaned from questioning participants to make sense of the social interaction they have observed. Unobtrusive indicators can be used to evaluate the honesty of survey responses.

Triangulation Triangulation—the use of two or more different measures of the same variable—can strengthen measurement considerably. When we achieve similar results with different measures of the same variable, particularly when they are based on such different methods as survey questions and field-based observations, we can be more confident in the validity of each measure.

Levels of Measurement Level of measurement: The mathematical precision with which the values of a variable can be expressed. When we know a variable’s level of measurement, we can better understand how cases vary on that variable and so understand more fully what we have measured. The nominal level of measurement, which is qualitative, has no mathematical interpretation; the quantitative levels of measurement— ordinal, interval, and ratio—are progressively more precise mathematically.

Nominal level of measurement The nominal level of measurement identifies variables whose values have no mathematical interpretation; they vary in kind or quality but not in amount (they may also be called categorical or qualitative variables). “State” (referring to the United States) is one example. Nationality, occupation, religious affiliation, and region of the country are also measured at the nominal level.

Insert exhibit 4.3

Although the attributes of categorical variables do not have a mathematical meaning, they must be assigned to cases with great care. The attributes we use to measure, or categorize, cases must be mutually exclusive and exhaustive: A variable’s attributes or values are mutually exclusive if every case can have only one attribute. A variable’s attributes or values are exhaustive when every case can be classified into one of the categories.

Ordinal level of measurement The first of the three quantitative levels is the ordinal level of measurement. At this level, the numbers assigned to cases specify only the order of the cases, permitting “greater than” and “less than” distinctions. If the response choices to a question range from “very wrong” to “not wrong at all”, for example. The different values of a variable measured at the ordinal level must be mutually exclusive and exhaustive.

Interval level of measurement The numbers indicating the values of a variable at the interval level of measurement represent fixed measurement units but have no absolute, or fixed, zero point. The numbers can therefore be added and subtracted, but ratios are not meaningful. Again, the values must be mutually exclusive and exhaustive. This level of measurement is represented by the difference between two Fahrenheit temperatures.

Insert exhibit 4.4

Ratio level of measurement The numbers indicating the values of a variable at the ratio level of measurement represent fixed measuring units and an absolute zero point (zero means absolutely no amount of whatever the variable indicates). For most statistical analyses in social science research, the interval and ratio levels of measurement can be treated as equivalent. On a ratio scale, 10 is 2 points higher than 8 and is also 2 times greater than 5—the numbers can be compared in a ratio.

Comparison of levels of measurement Researchers choose levels of measurement in the process of operationalizing variables; the level of measurement is not inherent in the variable itself. Many variables can be measured at different levels, with different procedures. It is usually is a good idea to try to measure variables at the highest level of measurement possible. The more information available, the more ways we have to compare cases.

Insert exhibit 4.5

Did we measure what we wanted to measure? Do the operations developed to measure our variables actually do so—are they valid? If we have weighed our measurement options, carefully constructed our questions and observational procedures, and selected sensibly from the available data indicators, we should be on the right track. But we cannot have much confidence in a measure until we have empirically evaluated its validity.

Measurement validity Measurement validity refers to the extent to which measures indicate what they are intended to measure. When a measure “misses the mark”—when it is not valid—our measurement procedure has been affected by measurement error.

Face validity Researchers apply the term face validity to the confidence gained from careful inspection of a concept to see if it is appropriate “on its face.” More precisely, we can say that a measure is face valid if it obviously pertains to the meaning of the concept being measured more than to other concepts (Brewer & Hunter 1989:131). Although every measure should be inspected in this way, face validation in itself does not provide convincing evidence of measurement validity.

Content validity Content validity establishes that the measure covers the full range of the concept’s meaning. To determine that range of meaning, the researcher may solicit the opinions of experts and review literature that identifies the different aspects, or dimensions, of the concept.

Criterion validity Criterion validity is established when the scores obtained on one measure can be accurately compared to those obtained with a more direct or already validated measures of the same phenomenon (the criterion). Inconsistent findings about the validity of a measure can occur because of differences in the adequacy of a measure across settings and populations. We cannot simply assume that a measure that was validated in one study is also valid in another setting or with a different population.

The criterion that researchers select can be measured either at the same time as the variable to be validated or after that time. Concurrent validity exists when a measure yields scores that are closely related to scores on a criterion measured at the same time. Predictive validity is the ability of a measure to predict scores on a criterion measured in the future.

Construct validity Measurement validity can also be established by showing that a measure is related to a variety of other measures as specified in a theory. This validation approached, known as construct validity, is commonly used in social research when no clear criterion exists for validation purposes.

Reliability Reliability means that a measurement procedure yields consistent scores when the phenomenon being measured is not changing (or that the measured scores change in direct correspondence to actual changes in the phenomenon). If a measure is reliable, it is affected less by random error, or chance variation, than if it is unreliable. Reliability is a prerequisite for measurement validity: We cannot really measure a phenomenon if the measure we are using gives inconsistent results.

Four possible indications of unreliability test-retest reliability interim reliability alternate-forms reliability Inter-observer reliability

Test-retest reliability When researchers measure a phenomenon that does not change between two point separated by an interval of time, the degree to which the two measurements are related to each other is the test-retest reliability of the measure.

Interitem reliability When researchers use multiple items to measure a single concept, they must be concerned with interim reliability (or interim consistency). The stronger the association is among the individual items and the more items there are that are included, the higher the reliability of the index will be.

Alternate-forms reliability Researchers are testing alternate-forms reliability when they compare subjects’ answers to slightly different versions of survey questions (Litwin 1995:13–21). A related test of reliability is the split-half reliability approach. Reliability is achieved when responses to the same questions by two randomly selected halves of a sample are about the same.

Interobserver reliability When researchers use more than one observer to rate the same people, events, or places, interobserver reliability is their goal. Assessing interobserver reliability is most important when the rating task is complex.

Ways to improve reliability and validity Whatever the concept measured or the validation method used, no measure is without some error, nor can we expect it to be valid for all times and places. Remember that a reliable measure is not necessarily a valid measure. Unfortunately, many measures are judged to be worthwhile on the basis of only a reliability test.

Remember always that measurement validity is a necessary foundation for social research. Gathering data without careful conceptualization or conscientious efforts to operationalize key concepts often is a wasted effort. The difficulties of achieving valid measurement vary with the concept being operationalized and the circumstances of the particular study.

Planning ahead is the key to achieving valid measurement in your own research; careful evaluation is the key to sound decisions about the validity of measures in others’ research. Statistical tests can help to determine whether a given measure is valid after data have been collected, but if it appears after the fact that a measure is invalid, little can be done to correct the situation. If you cannot tell how key concepts were operationalized when you read a research report, don’t trust the findings. And if a researcher does not indicate the results of tests used to establish the reliability and validity of key measures, remain skeptical.