BPK 304W Correlation.

Slides:



Advertisements
Similar presentations
Managerial Economics in a Global Economy
Advertisements

Kin 304 Regression Linear Regression Least Sum of Squares
Forecasting Using the Simple Linear Regression Model and Correlation
Inference for Regression
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Correlation and Regression
Chapter 10 Regression. Defining Regression Simple linear regression features one independent variable and one dependent variable, as in correlation the.
Simple Linear Regression
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
Chapter 10 Simple Regression.
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
Chapter Eighteen MEASURES OF ASSOCIATION
Chapter Topics Types of Regression Models
Chapter 11 Multiple Regression.
Probability & Statistics for Engineers & Scientists, by Walpole, Myers, Myers & Ye ~ Chapter 11 Notes Class notes for ISE 201 San Jose State University.
Simple Linear Regression Analysis
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
Chapter 7 Forecasting with Simple Regression
Simple Linear Regression Analysis
Relationships Among Variables
Lecture 5 Correlation and Regression
Correlation & Regression
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Introduction to Linear Regression and Correlation Analysis
Inference for regression - Simple linear regression
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Correlation and Regression
BPS - 3rd Ed. Chapter 211 Inference for Regression.
Managerial Economics Demand Estimation. Scatter Diagram Regression Analysis.
Correlation and Regression Used when we are interested in the relationship between two variables. NOT the differences between means or medians of different.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
Production Planning and Control. A correlation is a relationship between two variables. The data can be represented by the ordered pairs (x, y) where.
Correlation and Regression Chapter 9. § 9.3 Measures of Regression and Prediction Intervals.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 14 Inference for Regression AP Statistics 14.1 – Inference about the Model 14.2 – Predictions and Conditions.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Kin 304 Correlation Correlation Coefficient r Limitations of r
Lesson 14 - R Chapter 14 Review. Objectives Summarize the chapter Define the vocabulary used Complete all objectives Successfully answer any of the review.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
BPS - 5th Ed. Chapter 231 Inference for Regression.
Stats Methods at IC Lecture 3: Regression.
Chapter 4: Basic Estimation Techniques
BPK 304W Correlation Correlation Coefficient r Limitations of r
Regression and Correlation of Data Summary
Regression Analysis AGEC 784.
Correlation, Bivariate Regression, and Multiple Regression
AP Statistics Chapter 14 Section 1.
Basic Estimation Techniques
Kin 304 Regression Linear Regression Least Sum of Squares
BPK 304W Regression Linear Regression Least Sum of Squares
Quantitative Methods Simple Regression.
Basic Estimation Techniques
Kin 304 Correlation Correlation Coefficient r Limitations of r
Simple Linear Regression
Basic Practice of Statistics - 3rd Edition Inference for Regression
Correlation and Regression
Simple Linear Regression
SIMPLE LINEAR REGRESSION
Chapter 14 Inference for Regression
Introduction to Regression
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

BPK 304W Correlation

Correlation Coefficient (r) Correlation Coefficient (r) is a measure of association between two variables Varies from -1 to +1 r is a ratio of variability in X to that of Y. 0 = no relationship; 1 = perfect relationship Correlation

Correlation

Linear Fit High Correlation Coefficient does not mean a linear fit

Correlation does not mean causation Spurious Correlations – coincidental correlation between two unrelated variables A study of boys aged 6 to 18 years produced correlations of standing broad jump with other measures. Which had the highest correlation? Correlation

Range of the Data affects the Correlation Coefficient

Correlation coefficient depends upon the orientation of the two groups

Significance of the Correlation Coefficient The critical value of the correlation coefficient is determined by the sample size Bigger sample size = lower critical value of r Statistical significance of r does not infer “practical significance” Correlation

Degrees Probability   of Freedom 0.05 0.01 1 .997 1.000 24 .388 .496 2 .950 .990 25 .381 .487 3 .878 .959 26 .374 .478 4 .811 .917 27 .367 .470 5 .754 .874 28 .361 .463 6 .707 .834 29 .355 .456 7 .666 .798 30 .349 .449 8 .632 .765 35 .325 .418 9 .602 .735 40 .304 .393 10 .576 .708 45 .288 .372 11 .553 .684 50 .273 .354 12 .532 .661 60 .250 13 .514 .641 70 .232 .302 14 .497 .623 80 .217 .283 15 .482 .606 90 .205 .267 16 .468 .590 100 .195 .254 17 .575 125 .174 .228 18 .444 .561 150 .159 .208 19 .433 .549 200 .138 .181 20 .423 .537 300 .113 .148 21 .413 .526 400 .098 .128 22 .404 .515 500 .088 .115 23 .396 .505 1,000 .062 .081 Table 2-4.2: Critical Values of the Correlation Coefficient

Coefficient of Determination R squared (r2) The circle represents the total variance in the measure Weight 75% unexplained Correlation of Weight with Arm Girth r = 0.5, r2 = 0.25 Therefore 25% of the variance in weight is explained by arm girth 25% Arm Girth

Correlations between all variables Correlation Matrix Correlations between all variables Weight vs Arm Girth r = 0.5, r2 = 0.25 Weight vs Calf Girth r = 0.6 , r2 = 0 .36 Arm Girth vs Calf Girth r = 0.4 , r2 = 0 .16 Weight Arm Girth Calf Girth

BPK 304W Regression

Prediction Can we predict one variable from another? Linear Regression Analysis Y = mX + c m = slope; c = intercept Regression

Linear Regression Correlation Coefficient (r) how well the line fits Standard Error of Estimate (S.E.E.) how well the line predicts Regression

Least Sum of Squares Curve Fitting (Residual) Regression

Assumptions about the relationship between Y and X For each value of X there is a normal distribution of Y from which the sample value of Y is drawn The population of values of Y corresponding to a selected X has a mean that lies on the straight line In each population the standard deviation of Y about its mean has the same value

Standard Error of Estimate measure of how well the equation predicts Y has units of Y true score 68.26% of time is within plus or minus 1 SEE of predicted score Standard deviation of the normal distribution of residuals Regression

Right Hand L. = 0.99Left Hand L. + 0.254 r = 0.94 S.E.E. = 0.38cm Regression

How good is my equation? Regression equations are sample specific Cross-validation Studies Test your equation on a different sample Split sample studies Take a 50% random sample and develop your equation then test it on the other 50% of the sample Regression

Multiple Regression More than one independent variable Y = m1X1 + m2X2 + m3X3 …… + c Same meaning for r, and S.E.E., just more measures used to predict Y Stepwise regression variables are entered into the equation based upon their relative importance Regression

Building a multiple regression equation X1 has the highest correlation with Y, therefore it would be the first variable included in the equation. X3 has a higher correlation with Y than X2. However, X2 would be a better choice than X3. to include in an equation with X1, to predict Y. X2 has a low correlation with X1 and explains some of the variance that X1 does not. X3 Y X1 X2

Standardized Regression The numerical value is of mn is dependent upon the size of the independent variable Y = m1X1 + m2X2 + m3X3 …… + c Variables are transformed into standard scores before regression analysis, therefore mean and standard deviation of all independent variables are 0 and 1 respectively. The numerical value of zmn now represents the relative importance of that independent variable to the prediction Y = zm1X1 + zm2X2 + zm3X3 …… + c Regression