Correlation and Simple Linear Regression

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Chapter 12 Simple Linear Regression
Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Forecasting Using the Simple Linear Regression Model and Correlation
Inference for Regression
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Simple Linear Regression and Correlation
Chapter 12 Simple Linear Regression
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Chapter 10 Simple Regression.
The Simple Regression Model
SIMPLE LINEAR REGRESSION
Chapter Topics Types of Regression Models
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Correlation and Regression Analysis
Simple Linear Regression Analysis
Relationships Among Variables
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Lecture 5 Correlation and Regression
Correlation & Regression
SIMPLE LINEAR REGRESSION
Introduction to Linear Regression and Correlation Analysis
Correlation and Regression
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Introduction to Linear Regression
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Regression Analysis © 2007 Prentice Hall17-1. © 2007 Prentice Hall17-2 Chapter Outline 1) Correlations 2) Bivariate Regression 3) Statistics Associated.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Agresti/Franklin Statistics, 1 of 88 Chapter 11 Analyzing Association Between Quantitative Variables: Regression Analysis Learn…. To use regression analysis.
Correlation & Regression Analysis
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Chapter 13 Linear Regression and Correlation. Our Objectives  Draw a scatter diagram.  Understand and interpret the terms dependent and independent.
Correlation and Regression
Chapter 4: Basic Estimation Techniques
Lecture 11: Simple Linear Regression
Chapter 20 Linear and Multiple Regression
Regression and Correlation
Statistics for Business and Economics
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Statistics for Managers using Microsoft Excel 3rd Edition
Correlation and Simple Linear Regression
Basic Estimation Techniques
Linear Regression and Correlation Analysis
Chapter 11: Simple Linear Regression
Simple Linear Regression
Chapter 11 Simple Regression
Statistics for Business and Economics (13e)
Slides by JOHN LOUCKS St. Edward’s University.
Correlation and Simple Linear Regression
Basic Estimation Techniques
PENGOLAHAN DAN PENYAJIAN
Correlation and Simple Linear Regression
SIMPLE LINEAR REGRESSION
Simple Linear Regression and Correlation
Product moment correlation
SIMPLE LINEAR REGRESSION
Warsaw Summer School 2017, OSU Study Abroad Program
Introduction to Regression
St. Edward’s University
Correlation and Simple Linear Regression
Presentation transcript:

Correlation and Simple Linear Regression

Correlation Analysis Correlation analysis is used to describe the degree to which one variable is linearly related to another. There are two measures for describing correlation: The Coefficient of Correlation or Coefficient of linear correlation or Pearson’s coefficient of linear correlation (ρ / r / R) The Coefficient of Determination (r2 / R2 )

Correlation The correlation between two random variables, X and Y, is a measure of the degree of linear association between the two variables. The population correlation, denoted by, can take on any value from -1 to 1.    indicates a perfect negative linear relationship -1 <  < 0 indicates a negative linear relationship    indicates no linear relationship 0 <  < 1 indicates a positive linear relationship    indicates a perfect positive linear relationship The absolute value of  indicates the strength or exactness of the relationship.

Illustrations of Correlation Y X  = 1 Y X  = -1 Y X  = 0 Y X  = -.8 Y X  = 0 Y X  = .8

The coefficient of correlation: Sample Coefficient of Determination

The Coefficient of Correlation or Karl Pearson’s Coefficient of Correlation The coefficient of correlation is the square root of the coefficient of determination. The sign of r indicates the direction of the relationship between the two variables X and Y.

Simple Linear Regression Regression refers to the statistical technique of modeling the relationship between variables. In simple linear regression, we model the relationship between two variables. One of the variables, denoted by Y, is called the dependent variable and the other, denoted by X, is called the independent variable. The model we will use to depict the relationship between X and Y will be a straight-line relationship. A graphical sketch of the pairs (X, Y) is called a scatter plot.

Using Statistics This scatterplot locates pairs of observations of advertising expenditures on the x-axis and sales on the y-axis. We notice that: Larger (smaller) values of sales tend to be associated with larger (smaller) values of advertising. S c a t e r p l o f A d v i s n g E x u ( X ) Y 5 4 3 2 1 8 6 The scatter of points tends to be distributed around a positively sloped straight line. The pairs of values of advertising expenditures and sales are not located exactly on a straight line. The scatter plot reveals a more or less strong tendency rather than a precise linear relationship. The line represents the nature of the relationship on average.

Examples of Other Scatterplots Y X Y X Y X X Y X Y X Y

Simple Linear Regression Model The equation that describes how y is related to x and an error term is called the regression model. The simple linear regression model is: y = a+ bx +e where: a and b are called parameters of the model, a is the intercept and b is the slope. e is a random variable called the error term.

Assumptions of the Simple Linear Regression Model The relationship between X and Y is a straight-line relationship. The errors i are normally distributed with mean 0 and variance 2. The errors are uncorrelated (not related) in successive observations. That is: ~ N(0,2) X Y E[Y]=0 + 1 X Assumptions of the Simple Linear Regression Model Identical normal distributions of errors, all centered on the regression line.

Errors in Regression Y . { X Xi

Estimating Using the Regression Line SIMPLE REGRESSION AND CORRELATION Estimating Using the Regression Line First, lets look at the equation of a straight line is: Independent variable Dependent variable Slope of the line Y-intercept

The Method of Least Squares SIMPLE REGRESSION AND CORRELATION The Method of Least Squares To estimate the straight line we have to use the least squares method. This method minimizes the sum of squares of error between the estimated points on the line and the actual observed points. The sign of r will be the same as the sign of the coefficient “b” in the regression equation Y = a + b X

Slope of the best-fitting Regression Line SIMPLE REGRESSION AND CORRELATION The estimating line Slope of the best-fitting Regression Line Y-intercept of the Best-fitting Regression Line

SIMPLE REGRESSION – EXAMPLE (Appliance store) Suppose an appliance store conducts a five-month experiment to determine the effect of advertising on sales revenue. The results are shown below. (File: PPT_Regr_example) Advertising Exp.($100s) Sales Rev.($1000S) 1 1 2 1 3 2 4 2 5 4

If the slope of the estimating line is positive SIMPLE REGRESSION AND CORRELATION If the slope of the estimating line is positive line is negative :- r is the positive square root :- r is the negative The relationship between the two variables is direct

Steps in Hypothesis Testing using SPSS State the null and alternative hypotheses Define the level of significance (α) Calculate the actual significance : p-value Make decision : Reject null hypothesis, if p≤ α, for 2-tail test Conclusion

Std. Error of the Estimate Summary of SPSS Regression Analysis Output Model Summary Model R R Square Adjusted R Square Std. Error of the Estimate 1 .904a .817 .756 .606 a. Predictors: (Constant), Advertising expenses ($00) ANOVAb Model Sum of Squares df Mean Square F Sig. 1 Regression 4.900 13.364 .035a Residual 1.100 3 .367 Total 6.000 4 a. Predictors: (Constant), Advertising expenses ($00) b. Dependent Variable: Sales revenue ($000)

Advertising expenses ($00) Excel and SPSS Correlation Outputs   Advt. Sales 1 0.903696 Correlations Advertising expenses ($00) Sales revenue ($000) Pearson Correlation 1 .904* Sig. (2-tailed) .035 N 5 *. Correlation is significant at the 0.05 level (2-tailed).

Unstandardized Coefficients Standardized Coefficients Coefficientsa Model Unstandardized Coefficients Standardized Coefficients t Sig. B Std. Error Beta 1 (Constant) -.100 .635 -.157 .885 Advertising expenses ($00) .700 .191 .904 3.656 .035 a. Dependent Variable: Sales revenue ($000)

Hypothesis Tests for the Correlation Coefficient H0:  = 0 (No significant linear relationship) H1:   0 (Linear relationship is significant) Use p-value for decision making.

Analysis-of-Variance Table and an F Test of the Regression Model H0 : The regression model is not significant H1 : The regression model is significant

The p-value is 0.035 Conclusion: There is sufficient evidence to reject the null hypothesis in favor of the alternative hypothesis. b is not equal to zero. Thus, the independent variable is linearly related to y. This linear regression model is valid

Testing for the existence of linear relationship We test the hypothesis: H0: b = 0 (the independent variable is not a significant predictor of the dependent variable) H1: b is not equal to zero (the independent variable is a significant predictor of the dependent variable). If b is not equal to zero (if the null hypothesis is rejected), we can conclude that the Independent variable contributes significantly in predicting the Dependent variable.

Conclusion: Alternately, the actual significance is 0.035. Therefore we will reject the null hypothesis. The advertising expenses is a significant explanatory variable.