Larry D. Gruppen, Ph.D. University of Michigan From Concepts to Data: Conceptualization, Operationalization, and in Educational Research Measurement.

Slides:



Advertisements
Similar presentations
Questionnaire Development
Advertisements

Agenda Levels of measurement Measurement reliability Measurement validity Some examples Need for Cognition Horn-honking.
Chapter 8 Flashcards.
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Engineering Education Larry D. Gruppen, Ph.D. University of Michigan Rigorous Research.
Conceptualization and Measurement
Cal State Northridge Psy 427 Andrew Ainsworth PhD
The Research Consumer Evaluates Measurement Reliability and Validity
Taking Stock Of Measurement. Basics Of Measurement Measurement: Assignment of number to objects or events according to specific rules. Conceptual variables:
Research Methodology Lecture No : 11 (Goodness Of Measures)
Part II Sigma Freud & Descriptive Statistics
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Measurement the process by which we test hypotheses and theories. assesses traits and abilities by means other than testing obtains information by comparing.
Chapter 4A Validity and Test Development. Basic Concepts of Validity Validity must be built into the test from the outset rather than being limited to.
Measurement. Scales of Measurement Stanley S. Stevens’ Five Criteria for Four Scales Nominal Scales –1. numbers are assigned to objects according to rules.
Chapter 4 Validity.
MEASUREMENT. Measurement “If you can’t measure it, you can’t manage it.” Bob Donath, Consultant.
Concept of Measurement
The Methods of Social Psychology
Comparing Assessment to Educational Research. In this session, we will… Construct and discuss the (related but different) goals and attributes of assessment.
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Refining Your Research Question. In this session, we will… Discuss guidelines for creating a ‘good’ research question Provide time to revisit and revise.
Classroom Assessment A Practical Guide for Educators by Craig A
Reliability and Validity. Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? How do we judge.
Psychometrics Timothy A. Steenbergh and Christopher J. Devers Indiana Wesleyan University.
Ch 6 Validity of Instrument
Research Methods Key Points What is empirical research? What is the scientific method? How do psychologists conduct research? What are some important.
Reliability and Validity what is measured and how well.
Instrumentation.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Evaluating a Research Report
Measurement Validity.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
MEASUREMENT. MeasurementThe assignment of numbers to observed phenomena according to certain rules. Rules of CorrespondenceDefines measurement in a given.
MOI UNIVERSITY SCHOOL OF BUSINESS AND ECONOMICS CONCEPT MEASUREMENT, SCALING, VALIDITY AND RELIABILITY BY MUGAMBI G.K. M’NCHEBERE EMBA NAIROBI RESEARCH.
Measurement and Scaling
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
SOCW 671: #5 Measurement Levels, Reliability, Validity, & Classic Measurement Theory.
The Development and Validation of the Evaluation Involvement Scale for Use in Multi-site Evaluations Stacie A. ToalUniversity of Minnesota Why Validate.
Validity and Reliability in Instrumentation : Research I: Basics Dr. Leonard February 24, 2010.
SECOND EDITION Chapter 5 Standardized Measurement and Assessment
Chapter 6 - Standardized Measurement and Assessment
Reliability a measure is reliable if it gives the same information every time it is used. reliability is assessed by a number – typically a correlation.
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
MGMT 588 Research Methods for Business Studies
Reliability and Validity
Writing a sound proposal
Ch. 5 Measurement Concepts.
Lecture 5 Validity and Reliability
Leacock, Warrican and Rose (2009)
QUESTIONNAIRE DESIGN AND VALIDATION
Reliability and Validity
Concept of Test Validity
Test Validity.
CHAPTER 5 MEASUREMENT CONCEPTS © 2007 The McGraw-Hill Companies, Inc.
Tests and Measurements: Reliability
Journalism 614: Reliability and Validity
Reliability & Validity
Introduction to Measurement
Human Resource Management By Dr. Debashish Sengupta
پرسشنامه کارگاه.
5. Reliability and Validity
Reliability and Validity of Measurement
Unit IX: Validity and Reliability in nursing research
RESEARCH METHODS Lecture 18
Measurement Concepts and scale evaluation
Conducting Quality Service-Learning Research Andrew Furco Associate Vice President for Public Engagement March 22, 2019 Hong Kong Polytechnic University.
Presentation transcript:

Larry D. Gruppen, Ph.D. University of Michigan From Concepts to Data: Conceptualization, Operationalization, and in Educational Research Measurement

Objectives Identify key research design issues Wrestle with the complexities of educational measurement Explain the concepts of reliability and validity in educational measurement Apply criteria for measurement quality when conducting educational research

Agenda A brief nod to design From theory to measurement Criteria for measurement quality –Reliability –Validity Application: analyze an article

Guiding Principles for Scientific Research in Education 1.Question: pose significant question that can be investigated empirically 2.Theory: link research to relevant theory 3.Methods: use methods that permit direct investigation of the question 4.Reasoning: provide coherent, explicit chain of reasoning 5.Replicate and generalize across studies 6.Disclose research to encourage professional scrutiny and critique

Study design Study design consists of: –Your measurement method(s) –The participants and how they are assigned –The intervention –The sequence and timing of measurements and interventions

Randomization Purpose: attempt to balance effects that are unknown or uncontrollable Randomization in selecting participants (sampling) Randomization in assigning participants to study conditions

Comparison Group Pre-post design - compare intervention group to itself Non-equivalent control group design - compare intervention group to an existing group Randomized control group design - compare to equivalent controls

Overview of Study Designs Symbols –Each line represents a group. – x = Intervention (e.g. treatment) –O 1, O 2, O 3 …= Observation (measurement) at Time 1, Time 2, Time 3, etc. –R = Random assignment

Non-Experimental Designs

x O 1 One-Group Posttest x O 1

Quasi-Experimental Designs

x O 1 O 1 Posttest-Only Control Group

O 1 x O 2 One-Group Pretest-Posttest

O 1 x O 2 O 1 O 2 Control Group Pretest-Posttest

Experimental Designs

Posttest Only Randomized Control Group R x O 1 R O 1

R O 1 x O 2 R O 1 O 2 Randomized Control Group Pretest- Posttest

Theory Constructs Operational Definition Measurement From Theory to Measurement

An Example: Self-assessment Theory –Self-assessment guides self-directed learning –Individuals may differ in how self-assessment guides self-directed learning Research question: To what extent do students spend more time learning in areas in which they perceive themselves to be weakest Constructs –Self-assessment of knowledge and skill –Self-directed learning

An Example: Self-assessment Operational definition –Self-assessment: Relative ranking of strengths and weaknesses in knowledge and skills - not in reference to other people –Self-directed learning: Amount of time they spent learning about 14 categories of patient presentation (e.g., cough, abdominal pain)

An Example: Self-assessment Measurement –Self-assessment: Rate on 10-point scale (1=least confident, 10=most confident) relative confidence in knowledge and skill on 14 presenting complaints –Self-directed learning: Self-report of time spent learning each of the 14 complaints using 11-point scale (0=no time at all, 10=a great deal)

Measurement Measurement: assignment of numbers to objects or events according to rules Quality: reliability and validity

The Challenge of Educational Measurement Almost all of the constructs we are interested in are buried inside the individual Measurement depends on transforming these internal states, events, capabilities, etc. into something observable Making them observable may alter the thing we are measuring

Examples of Measurement Methods Tests (knowledge, performance): defined response, constructed response, simulations Questionnaires (attitudes, beliefs, preferences): rating scales, checklists, open-ended responses Observations (performance, skills): tasks (varying degrees of authenticity), problems, real- world behaviors, records (documents)

Reliability Dependability (consistency or stability) of measurement A necessary condition for validity

Types of Reliability Stability (produces the same results with repeated measurements over time): –Test-retest –Correlation between scores at 2 times Equivalence/Internal Consistency (produces same results with parallel items on alternate forms): –Alternate forms; split-half; Kuder-Richardson; Chronbach’s alpha –Correlation between scores on different forms; Calculate coefficient alpha (a) Consistency (produces the same results with different observers or raters): –Inter-rater agreement –Correlation between scores from different raters; kappa coefficient

Validity Refers to the accuracy of inferences based on data obtained from measurement Technically, measures aren’t valid, inferences are No such thing as validity in the abstract: the key issue is ‘valid’ for what inference Want to reduce systematic, non-random error Unreliability lowers correlations, reducing validity claims

Conventional View of Validity Face validity: logical link between items and purpose— makes sense on the surface Content validity: items cover the range of meaning included in the construct or domain. Expert judgment Criterion validity: relationship between performance on one measurement and performance on another (or actual behavior) Concurrent and Predictive Correlation coefficients Construct validity: directly connect measurement with theory. Allows interpretation of empirical evidence in terms of theoretical relationships. Based on weight of evidence. Convergent and discriminant evidence. Multitrait-MultiMethod Analysis (MTMM)

Unified View of Construct Validity (Messick S, Amer Psych, 1995) Validity is not a property of an instrument but rather of the meaning of the scores. Must be considered holistically. 6 Aspects of Construct Validity Evidence –Content—content relevance & representativeness –Substantive—theoretical rationale for observed consistencies in test responses –Structural—fidelity of scoring structure to structure of construct domain –Generalizability—generalization to the population and across populations –External—convergent and discriminant evidence –Consequential—intended and unintended consequences of score interpretation; social consequence of assessment (fairness, justice)

Finding Measurement Instruments Scan the engineering education literature (obviously) engineering ed researchers (use the network) Examine literature for instruments used in prior studies General education/social science instrument databases –Buros Institute of Mental Measurements (Mental Measurement Yearbook, Tests in Print) –ERIC databases –Educational Testing Service Test Collection Construct your own (last resort!) –Get some expert consultation (test writing, survey design, questionnaire construction, etc.)

Example In your groups, analyze the Steif & Dantzler statics concept inventory article. Look for: –Theoretical framework –Constructs used in the study –How constructs were operationalized –Measurement process Attention to reliability and validity

References Campbell DT, Stanley JC. Experimental and quasi- experimental designs for research. Chicago: Rand McNally; Cook, T.D. and Campbell, D.T. (1979). Quasi- Experimentation: Design and Analysis for Field Settings. Rand McNally, Chicago, Illinois. Messick S. Validity of psychological assessment: validation of inferences from persons' responses and performances as scientific inquiry into score meaning. American Psychologist. 1995;50: Messick S. Validity. In: Linn RL, ed. Educational measurement. 3rd ed. New York: American Council on Education & Macmillan; 1989: