Presentation on theme: "1. Observational 2. Physiological and Neuroscientific 3. Self-report --majority of social & behavioral science research."— Presentation transcript:
1. Observational 2. Physiological and Neuroscientific 3. Self-report --majority of social & behavioral science research
Self-report measures People’s replies to written questionnaires or interviews Can measure: ▪ thoughts (cognitive self-reports) ▪ feelings (affective self-reports) ▪ actions (behavioral self-reports)
Self-reported momentary emotions: Positive and Negative Affect Schedule (PANAS) (Watson, Clark & Tellegen,1988) Indicate the extent you feel this way right now: enthusiastic Not at all enthusiastic Very enthusiastic Indicate the extent you feel this way right now: upset Not at all upset Very upset
Nominal Hot = 1 Warm = 3 Cold = 2 Ordinal 1 st Place Sample 2 nd Place Sample 3 rd Place Sample 4 th Place Sample 5 th Place Sample Thing being measured Interval Interval Ratio
Distinction between scales is due to the meaning of numbers 1. Nominal Scale—numbers assigned are only labels. 2. Ordinal Scale—a rank ordering. 3. Interval Scale—each number equidistant from the next, but no zero point (majority of measures). 4. Ratio Scale—each number is equidistant and there is a true zero point.
Type of Scale Determines Statistics and Power StatisticsPower NominalChi-squareLow OrdinalRank-order testsModerate IntervalParametric tests (F-tests, t-tests) High RatioParametric tests and math operations High
Valid: measure assesses the construct it is intended to and is not influenced by other factors Reliable: the consistency of a measure, does it provide the same result repeatedly.
Reliable but not Valid Dependable measure, but doesn’t measure what it should Example: Arm length to measure self-esteem. Valid but not Reliable Measures what it should, but not dependably Example: Stone as a measure of weight in Great Britain.
Central dot = construct we are seeking to measure
Test-Retest Reliability Measure administered at two points in time to assess consistency. Works best for things that do not change over time (e.g., intelligence). Internal Consistency Reliability Judgments of consistency of results across items in the same test administration session. 1. Intercorrelation: Chronbach’s α (>.65 is preferred) 2. Split halves reliability
Content Validity Does the measure represent the range of possible items the it should cover based on the meaning of the measure. Predictive Validity measure predicts criterion measures that are assessed at a later time. Ex: Does aptitude assessment predict later success? Construct Validity Does the measure actually tap into intended construct?
Guided spontaneous response from individuals in sample population (thought listings, essay questions…) Face valid items: develop items that appear to measure your construct. Pilot test a larger set of items and choose those that are more reliable & valid. Reversed coded items indicate whether participants are paying attention.
Likert Scale: To what extent do you agree with the following statement… (0 to 9, strongly disagree-strongly agree) Semantic Differential: What is your response to (insert person, object, place, issue)? (-5 to +5, good-bad, like-dislike, warm-cold)
The measure exists already in the literature Restriction of range: responses either at high or low end of scale (skew). Can you trust responses? Social desirability, demand characteristics & satisficing.
1. Develop subjective and objective versions of a new scale Example: Contact with Blacks scale: Objective: % of your neighborhood growing up Subjective: No Blacks—a lot of Blacks 2. Using 5+ items worded similarly provides greatly increased reliability and likelihood of success. 3. Human targets are rarely evaluated below the midpoint of the scale, so use more scale points (9 instead of 5 points).
**Most Important** If you have a larger study ready and a great idea for a new scale comes up, build something and give it a shot!
Response time measures Physiological measures Neuroscience: fMRI and other brain imaging Indirect measures: projective tests, etc. Facial and other behavior coding schemes (verbal/nonverbal) Cognitive measures: (memory, perception…) Task performance: academic, physical… Game theory: prisoner’s dilemma…
Chronbach’s α: Analyze Scale Reliability Analysis Pull over all scale items Click Statistics, select inter-item correlations OK Try Van Camp, Barden & Sloan (2010) data file. Centrality1- Centrality8. Compare to manuscript. Many other reliability analyses involve correlations (test-retest, split halves) or probabilities (inter-rater reliability).
Case Processing Summary N% CasesValid Excluded a Total a. Listwise deletion based on all variables in the procedure. Reliability Statistics Cronbach's Alpha Cronbach's Alpha Based on Standardized ItemsN of Items Inter-Item Correlation Matrix centrality1revcentrality2centrality3centrality4revcentrality5centrality6centrality7centrality8rev centrality1rev centrality centrality centrality4rev centrality centrality centrality centrality8rev
Factor Analysis: determines factor structure of measures (does your measure assess one construct or multiple constructs? Is your proposed construct coherent?) Multi-trait Multi-method Matrix: using combination of existing measures and manipulations to establish convergent/ divergent validity with measure.
Inter-rater Reliability Independent judges score participant responses and the % of agreement is assessed to indicate reliability. Used particularly for measures requiring coding (video coding, spontaneous responses…).