Correlation, Reliability and Regression Chapter 7.

Slides:



Advertisements
Similar presentations
Quantitative Methods in HPELS 440:210
Advertisements

Chapter 16: Correlation.
Wednesday AM  Presentation of yesterday’s results  Associations  Correlation  Linear regression  Applications: reliability.
Action Research Correlation and Regression
RELIABILITY Reliability refers to the consistency of a test or measurement. Reliability studies Test-retest reliability Equipment and/or procedures Intra-
Correlation Chapter 6. Assumptions for Pearson r X and Y should be interval or ratio. X and Y should be normally distributed. Each X should be independent.
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
Statistics Measures of Regression and Prediction Intervals.
Describing Relationships Using Correlation and Regression
Correlation CJ 526 Statistical Analysis in Criminal Justice.
Chapter 15 (Ch. 13 in 2nd Can.) Association Between Variables Measured at the Interval-Ratio Level: Bivariate Correlation and Regression.
CJ 526 Statistical Analysis in Criminal Justice
Chapter 13 Conducting & Reading Research Baumgartner et al Data Analysis.
Lecture 11 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
A quick introduction to the analysis of questionnaire data John Richardson.
Correlation-Regression The correlation coefficient measures how well one can predict X from Y or Y from X.
Chapter Seven The Correlation Coefficient. Copyright © Houghton Mifflin Company. All rights reserved.Chapter More Statistical Notation Correlational.
Chap 3-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 3 Describing Data: Numerical Statistics for Business and Economics.
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Linear Regression/Correlation
Aim: How do we calculate and interpret correlation coefficients with SPSS? SPSS Assignment Due Friday 2/12/10.
Chapter 9 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 What is a Perfect Positive Linear Correlation? –It occurs when everyone has the.
Relationships Among Variables
Correlation Coefficient Correlation coefficient refers to the type of relationship between variables that allows one to make predications from one variable.
Chapter 21 Correlation. Correlation A measure of the strength of a linear relationship Although there are at least 6 methods for measuring correlation,
© 2011 Cengage Learning. All Rights Reserved. May not be copied, scanned, or duplicated, in whole or in part, except for use as permitted in a license.
Chapter 12 Correlation and Regression Part III: Additional Hypothesis Tests Renee R. Ha, Ph.D. James C. Ha, Ph.D Integrative Statistics for the Social.
Correlation and regression 1: Correlation Coefficient
Chapter 14 – Correlation and Simple Regression Math 22 Introductory Statistics.
Chapter 15 Correlation and Regression
Statistics 11 Correlations Definitions: A correlation is measure of association between two quantitative variables with respect to a single individual.
Correlation is a statistical technique that describes the degree of relationship between two variables when you have bivariate data. A bivariate distribution.
Experimental Research Methods in Language Learning Chapter 11 Correlational Analysis.
Hypothesis of Association: Correlation
METHODS IN BEHAVIORAL RESEARCH NINTH EDITION PAUL C. COZBY Copyright © 2007 The McGraw-Hill Companies, Inc.
Basic Statistics Correlation Var Relationships Associations.
Figure 15-3 (p. 512) Examples of positive and negative relationships. (a) Beer sales are positively related to temperature. (b) Coffee sales are negatively.
Investigating the Relationship between Scores
© 2011 Cengage Learning. All Rights Reserved. May not be copied, scanned, or duplicated, in whole or in part, except for use as permitted in a license.
Correlation and Regression Chapter 9. § 9.3 Measures of Regression and Prediction Intervals.
Elementary Statistics Correlation and Regression.
Examining Relationships in Quantitative Research
Introduction to Behavioral Statistics Correlation & Regression.
By: Amani Albraikan.  Pearson r  Spearman rho  Linearity  Range restrictions  Outliers  Beware of spurious correlations….take care in interpretation.
Correlation & Regression Chapter 15. Correlation It is a statistical technique that is used to measure and describe a relationship between two variables.
ITEC6310 Research Methods in Information Technology Instructor: Prof. Z. Yang Course Website: c6310.htm Office:
Describing Relationships Using Correlations. 2 More Statistical Notation Correlational analysis requires scores from two variables. X stands for the scores.
11/23/2015Slide 1 Using a combination of tables and plots from SPSS plus spreadsheets from Excel, we will show the linkage between correlation and linear.
Correlations. Distinguishing Characteristics of Correlation Correlational procedures involve one sample containing all pairs of X and Y scores Correlational.
Chapter 14 Correlation and Regression
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Chapter Thirteen Bivariate Correlation and Regression Chapter Thirteen.
Chapter Eleven Performing the One-Sample t-Test and Testing Correlation.
Chapter 13 Understanding research results: statistical inference.
© 2011 Cengage Learning. All Rights Reserved. May not be copied, scanned, or duplicated, in whole or in part, except for use as permitted in a license.
Correlation and Regression Elementary Statistics Larson Farber Chapter 9 Hours of Training Accidents.
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 3 Investigating the Relationship of Scores.
Chapter 11 Linear Regression and Correlation. Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and.
Chapter 12 Understanding Research Results: Description and Correlation
12 Inferential Analysis.
Understanding Research Results: Description and Correlation
Introduction to Behavioral Statistics
Linear Regression/Correlation
The Pearson Correlation
STATISTICS Topic 1 IB Biology Miss Werba.
12 Inferential Analysis.
4/4/2019 Correlations.
Correlations: Correlation Coefficient:
Introduction to Regression
Presentation transcript:

Correlation, Reliability and Regression Chapter 7

Correlation  Statistic that describes the relationship between scores (Pearson r).  Number is the correlation coefficient.  Ranges between and –1.00.  Positive is direct relationship.  Negative is inverse relationship. .00 is no relationship.  Does not mean cause and effect.  Measured by a Z score.  Generally looking for scores greater than.5

Reliability  Statistic used to determine repeatability  Number ranges between 0 and 1  Always positive  Closer to one is greater reliability  Closer to 0 is less reliability  Generally looking for values greater than.8

Scattergram or Scatterplot  Designate one variable X and one Y.  Draw and label axes.  Lowest scores are bottom left.  Plot each pair of scores.  Positive means high on both scores.  Negative means high on one and low on the other.  IQ and GPA?  ~ 0.68

Example (high correlation with systematic bias) Trial 1 Trial S1 S2 S3 S4 S5 S6

Positive Data

Positive Plot r=.93

Negative Data

Negative Plot r=-.92

Null Data

Null Plot (orthogonal) r=.34r=-.00

Pearson (Interclass Correlation) Ignores the systematic bias Has agreement (rank) but not correspondence (raw score) The order and the SD of the scores remain the same The mean may be different between the two tests The r can still be high (i.e. close to 1.0)

Calculation of r

ICC (Intraclass Correlation)  Addresses correspondence and agreement  R =  1.0 is perfect reliability  0.0 is no reliability

ICC  Advantages  More then two variables (ratings, raters etc.)  Will find the systematic bias  Interval, ratio or ordinal data

Example Trial 1 Trial S1 S2 S3 S4 S5 S6

ICC ICC = BMS - EMS BMS + (k-1) EMS

Trial 1 & 2 BMS - EMS BMS +(k-1)EMS + k [(TMS-EMS)/n] ICC = – (2-1) [( )/15] ICC = = 0.84 Pearson r = 0.91

Example Trial 1 Trial S1 S2 S3 S4 S5 S6 BMS TMS EMS

What is a Mean Square?  Sum of squared deviations divided by the degrees of freedom (df=values free to vary when sum is set)  SSx = sum of squared deviations about the mean which is a variance estimator

Running ICC on SPSS  Analyze, scale, reliability analysis  Choose two or more variables  Click statistics, check ICC at bottom  Two-way mixed, consistency  Use single measures on output

Pearson vs. ICC Trial 1Trial

Interpretation (positive or negative)  <.20 very low  low  moderate  high  very high  Must also consider the p-value

Correlation Conclusion Statement 1. Always past tense 2. Include interpretation 3. Include ‘significant’ 4. Include p or alpha value 5. Include direction 6. Include r value 7. Use variable names  There was a high significant (p<0.05) positive correlation (r=.78) between X and Y.

Pearson vs. ICC mean

Curvilinear  Scores curve around line of best fit.  Also called a trend analysis.  More complex statistics.

Coefficient of Determination  Represents the common variance between scores.  Square of the r value.  % explained.  How much one variable affects the other

R 2 the proportion of variance that two measures have in common-overlap determines relationship-explained variance

Partial Correlation the degree of relationship between two variables while ruling out that degree of correlation attributable to other variables

Simple Linear Regression  Predict one variable from one another  If measurement on one variable is difficult or missing  Prediction is not perfect but contains error  High reliability if error is low and R is high

Residual is vertical distance of any point from the line of best fit (predicted) r=.93 Positive and negative distances are equal

Prediction  Y=(bx)+c  Y is the predicted value  B is the slope of the line  X is the raw value of the predictor  C is the Y intercept (Y when x = zero)  Y vertical/X horizontal

SPSS

SPSS Printout Z-scores

Prediction  Y p = (bx)+c  HT p = (.85x80)+(108.58)  HT p = (68)+(108.58)  HT p =  Residual (error) = diff between predicted and actual  Subject must come from that population! HTWT

Standard Error of the Estimate (SEE) is the standard deviation of the distribution of residual scores  Error associated with the predicted value  Read as a SD or SEM value  68%, 95%, 99%  SEE x 2 then add and subtract it from the predicted score to determine 95% CI of the predicted score. SEE SEE = the square root of The squared residuals Divided by the number of pairs

Prediction  Y p = (bx)+c  HT p = (.85x80)+(108.58)  HT p = (68)+(108.58)  HT p =  SEE x 2 = 19.5  95% CI = – HTWT

Multiple Regression  Uses multiple X variables to predict Y  Results in beta weights for each X variable  Y=(b 1 x 1 ) + (b 2 x 2 ) + (b 3 x 3 ) … + c

SPSS - R Prediction Y=(WT x.40) - (skinfold x 1.04)

Equation  Y=(WTx.40)- (skinfoldx1.04)  Y=(80x.40)- (11x1.04)  Y=(32)-(11.44)  Y=  SEE=8.67 (x2=17.34)  CI= HTWTSkin

Next Class  Chapter 8 & 13 t-tests and chi-square.

Homework 1. Make a scatterplot with trendline and r and r 2 of two ratio variables. 2. Run Pearson r on four different variables and hand draw a scatterplot for two. 3. Run ICC between VJ1 and VJ2. 4. Run linear regression on standing long jump and predict stair up time. Work out the equation and CI for subject #2. 5. Run multiple regression on subject #2 and add vjump running, circumference and weight to predictors. Also out the equation and CI.