©2006 Thomson/South-Western 1 Chapter 14 – Multiple Linear Regression Slides prepared by Jeff Heyl Lincoln University ©2006 Thomson/South-Western Concise.

Slides:



Advertisements
Similar presentations
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 14-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter.
Advertisements

12-1 Multiple Linear Regression Models Introduction Many applications of regression analysis involve situations in which there are more than.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Chapter 10 Simple Regression.
Correlation and Simple Regression Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
© 2000 Prentice-Hall, Inc. Chap Multiple Regression Models.
Multiple Regression Models. The Multiple Regression Model The relationship between one dependent & two or more independent variables is a linear function.
Multiple Linear Regression Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 13-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Chapter Topics Types of Regression Models
Irwin/McGraw-Hill © The McGraw-Hill Companies, Inc., 2000 LIND MASON MARCHAL 1-1 Chapter Twelve Multiple Regression and Correlation Analysis GOALS When.
Predictive Analysis in Marketing Research
Statistics for Business and Economics Chapter 11 Multiple Regression and Model Building.
Multiple Linear Regression
Multiple Regression and Correlation Analysis
1 1 Slide © 2015 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
1 1 Slide © 2003 South-Western/Thomson Learning™ Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Regression Model Building Setting: Possibly a large set of predictor variables (including interactions). Goal: Fit a parsimonious model that explains variation.
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression Analysis
Forecasting Revenue: An Example of Regression Model Building Setting: Possibly a large set of predictor variables used to predict future quarterly revenues.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Correlation & Regression
Quantitative Business Analysis for Decision Making Multiple Linear RegressionAnalysis.
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Introduction to Linear Regression and Correlation Analysis
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 12-1 Chapter 12 Simple Linear Regression Statistics for Managers Using.
Chapter 12 Multiple Regression and Model Building.
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
1 1 Slide © 2003 Thomson/South-Western Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple Coefficient of Determination.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved OPIM 303-Lecture #9 Jose M. Cruz Assistant Professor.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Business Statistics, 4e, by Ken Black. © 2003 John Wiley & Sons Business Statistics, 4e by Ken Black Chapter 15 Building Multiple Regression Models.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
©2006 Thomson/South-Western 1 Chapter 13 – Correlation and Simple Regression Slides prepared by Jeff Heyl Lincoln University ©2006 Thomson/South-Western.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
14- 1 Chapter Fourteen McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved.
Chapter 13 Multiple Regression
Simple Linear Regression (SLR)
Simple Linear Regression (OLS). Types of Correlation Positive correlationNegative correlationNo correlation.
Lecture 10: Correlation and Regression Model.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Model Building and Model Diagnostics Chapter 15.
Copyright © 2012 Pearson Education, Inc. All rights reserved Chapter 12 Multiple Regression and Model Building.
Multiple Regression  Similar to simple regression, but with more than one independent variable R 2 has same interpretation R 2 has same interpretation.
1 1 Slide © 2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Copyright © 2004 by The McGraw-Hill Companies, Inc. All rights reserved.
Chap 13-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 13 Multiple Regression and.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Multiple Regression Chapter 14.
1 Multiple Regression. 2 Model There are many explanatory variables or independent variables x 1, x 2,…,x p that are linear related to the response variable.
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
1 BUSI 6220 By Dr. Nick Evangelopoulos, © 2012 Brief overview of Linear Regression Models (Pre-MBA level)
Yandell – Econ 216 Chap 15-1 Chapter 15 Multiple Regression Model Building.
Chapter 15 Multiple Regression and Model Building
Essentials of Modern Business Statistics (7e)
Chapter 13 Created by Bethany Stubbe and Stephan Kogitz.
Multiple Regression.
Business Statistics, 4e by Ken Black
Business Statistics Multiple Regression This lecture flows well with
Chapter 15 – Multiple Linear Regression
Prepared by Lee Revere and John Large
Chapter 14 – Correlation and Simple Regression
Multiple Regression Chapter 14.
Chapter Fourteen McGraw-Hill/Irwin
Business Statistics, 4e by Ken Black
Presentation transcript:

©2006 Thomson/South-Western 1 Chapter 14 – Multiple Linear Regression Slides prepared by Jeff Heyl Lincoln University ©2006 Thomson/South-Western Concise Managerial Statistics Concise Managerial Statistics KVANLI PAVUR KEELING KVANLI PAVUR KEELING

©2006 Thomson/South-Western 2 Multiple Regression Model Y =  0 +  1 X 1 +  2 X 2 +  k X k + e Deterministic component  0 +  1 X 1 +  2 X 2 +  k X k Least Squares Estimate SSE = ∑(Y - Y) 2 ^

©2006 Thomson/South-Western 3 Multiple Regression Model Figure 14.1 Y X1X1X1X1 X2X2X2X2 Y =  0 +  1 X 1 +  2 X 2 e (positive) e (negative)

©2006 Thomson/South-Western 4 Housing Example Y = Home square footage (100s) X1 = Annual Income ($1,000s) X2 = Family Size X3 = Combined years of education beyond high school for all household members

©2006 Thomson/South-Western 5 Multiple Regression Model Figure 14.2

©2006 Thomson/South-Western 6 Multiple Regression Model Figure 14.3

©2006 Thomson/South-Western 7 Multiple Regression Model Figure 14.4

©2006 Thomson/South-Western 8 Assumptions of the Multiple Regression Model The errors follow a normal distribution, centered at zero, with common variance The errors follow a normal distribution, centered at zero, with common variance The errors are independent The errors are independent

©2006 Thomson/South-Western 9 Errors in Multiple Linear Regression Figure 14.5 Y X1X1X1X1 X2X2X2X2 Y =  0 +  1 X 1 +  2 X 2 X 1 = 30, X 2 = 8 X 1 = 50, X 2 = 2 e e

©2006 Thomson/South-Western 10 Multiple Regression Model An estimate of  e 2 s 2 =  e 2 = = SSE n - (k + 1) SSE n - k - 1 ^

©2006 Thomson/South-Western 11 Hypothesis Test for the Significance of the Model H o :  1 =  2 = … =  k H a : at least one of the  ’s ≠ 0 Reject H o if F > F ,k,n-k-1 F =F =F =F =MSRMSE

©2006 Thomson/South-Western 12 Associated F Curve reject H 0 F,v,vF,v,vF,v,vF,v,v 12 Area =  Figure 14.6

©2006 Thomson/South-Western 13 Test for H o :  i = 0 reject H o if |t| > t ./2,n-k-1 t = b1 b1sbsb b1 b1sbsb1 H o :  1 = 0 (X 1 does not contribute) H a :  1 ≠ 0 (X 1 does contribute) H o :  2 = 0 (X 2 does not contribute) H a :  2 ≠ 0 (X 2 does contribute) H o :  3 = 0 (X 3 does not contribute) H a :  3 ≠ 0 (X 3 does contribute) b i - t  /2,n-k-1 s b to b i + t  /2,n-k-1 s b ii (1-  ) 100% Confidence Interval

©2006 Thomson/South-Western 14 Housing Example Figure 14.7

©2006 Thomson/South-Western 15 BB Investments Example BB Investments wants to develop a model to predict the amount of money invested by various clients in their portfolio of high-risk securities Y = Investment Amount ($) X1 = Annual Income ($1,000s) X2 = Economic Index, showing expected increase in interest levels, manufacturing costs, and price inflation ( scale)

©2006 Thomson/South-Western 16 BB Investments Example Figure 14.8

©2006 Thomson/South-Western 17 BB Investments Example Figure 14.9

©2006 Thomson/South-Western 18 BB Investments Example Figure 14.10

©2006 Thomson/South-Western 19 Coefficient of Determination SST= total sum of squares = SS Y = ∑(Y - Y) 2 = ∑Y 2 - (∑Y) 2 n R 2 = 1 - SSESST F =F =F =F = R 2 / k (1 - R 2 ) / (n - k - 1)

©2006 Thomson/South-Western 20 Partial F Test R c 2 = the value of R 2 for the complete model R r 2 = the value of R 2 for the reduced model Test statistic F =F =F =F = (R c 2 - R r 2 ) / v 1 (1 - R c 2 ) / v 2

©2006 Thomson/South-Western 21 Motormax Example Motormax produces electric motors in home furnaces. They want to study the relationship between the dollars spent per week in inspecting finished products (X) and the number of motors produced during that week that were returned to the factory by the customer (Y)

©2006 Thomson/South-Western 22 Motormax Example Figure 14.11

©2006 Thomson/South-Western 23 Quadratic Curves |11|111 |22|222|3|4| |11|111 |22|222|3|4|5 Figure YX (a) XXYY(b)(b)

©2006 Thomson/South-Western 24 Motormax Example Figure 14.13

©2006 Thomson/South-Western 25 Error From Extrapolation Figure Predicted Actual YX |1|2|3|4|5

©2006 Thomson/South-Western 26 Multicollinearity Occurs when independent variables are highly correlated with each other Often detectable through pairwise correlations readily available in statistical packages Often detectable through pairwise correlations readily available in statistical packages The variance inflation factor can also be used The variance inflation factor can also be used VIF j = R j 2 Conclude severe multicollinearity exists when the maximum VIF j > 10

©2006 Thomson/South-Western 27 Multicollinearity Example Figure 14.15

©2006 Thomson/South-Western 28 Multicollinearity Example Figure 14.16

©2006 Thomson/South-Western 29 Multicollinearity Example Figure 14.17

©2006 Thomson/South-Western 30 Multicollinearity The stepwise selection process can help eliminate correlated predictor variables The stepwise selection process can help eliminate correlated predictor variables Other advanced procedures such as ridge regression can also be applied Other advanced procedures such as ridge regression can also be applied Care should be taken during the model selection phase as multicollinearity can be difficult to detect and eliminate Care should be taken during the model selection phase as multicollinearity can be difficult to detect and eliminate

©2006 Thomson/South-Western 31 Dummy Variables Dummy, or indicator, variables allow for the inclusion of qualitative variables in the model For example: X1 =X1 =X1 =X1 = 1if female 0if male

©2006 Thomson/South-Western 32 Dummy Variable Example Figure 14.18

©2006 Thomson/South-Western 33 Stepwise Procedures Procedures either choose or eliminate variables, one at a time, in an effort to avoid including variables with either no predictive ability or are highly correlated with other predictor variables Forward regression Add one variable at a time until contribution is insignificant Forward regression Add one variable at a time until contribution is insignificant Backward regression Remove one variable at a time starting with the “worst” until R 2 drops significantly Backward regression Remove one variable at a time starting with the “worst” until R 2 drops significantly Stepwise regression Forward regression with the ability to remove variables that become insignificant Stepwise regression Forward regression with the ability to remove variables that become insignificant

©2006 Thomson/South-Western 34 Stepwise Regression Figure Include X 3 Include X 6 Include X 2 Include X 5 Remove X 2 (When X 5 was inserted into the model X 2 became unnecessary) Include X 7 Remove X 7 - it is insignificant Stop Final model includes X 3, X 5 and X 6

©2006 Thomson/South-Western 35 Checking Model Assumptions Checking Assumption 1 - Normal distribution Construct a histogram Checking Assumption 3 - Errors are independent Durbin-Watson statistic Checking Assumption 2 - Constant variance Plot residuals versus predicted Y values ^

©2006 Thomson/South-Western 36 Detecting Sample Outliers Sample leverages Sample leverages Standardized residuals Standardized residuals Cook’s distance measure Cook’s distance measure Standardized residual = Y i – Y i s 1 - h i ^

©2006 Thomson/South-Western 37 Cook’s Distance Measure k1 or 23 or 4≥ 5 DMAX Table 14.1 D i = (standardized residual) 2 1 k + 1 h i 1 - h i = (Y i - Y i ) 2 (k + 1)s 2 h i (1 – h i ) 2 ^

©2006 Thomson/South-Western 38 Residual Analysis Figure 14.20

©2006 Thomson/South-Western 39 Residual Analysis Figure 14.21

©2006 Thomson/South-Western 40 Residual Analysis Figure 14.22

©2006 Thomson/South-Western 41 Prediction Using Multiple Regression Figure 14.23

©2006 Thomson/South-Western 42 Prediction Using Multiple Regression Figure 14.24

©2006 Thomson/South-Western 43 Prediction Using Multiple Regression Figure 14.25

©2006 Thomson/South-Western 44 Prediction Using Multiple Regression Confidence and Prediction Intervals Y - t  /2,n-k-1 s Y to Y + t  /2,n-k-1 s Y ^^ ^^ (1-  ) 100% Confidence Interval for µ Y | X 0 (1-  ) 100% Confidence Interval for Y x 0 Y - t  /2,n-k-1 s 2 + s Y 2 to Y + t  /2,n-k-1 s 2 + s Y 2 ^^ ^^

©2006 Thomson/South-Western 45 Interaction Effects Implies how variables occur together has an impact on prediction of the dependent variable Y =  0 +  1 X 1 +  2 X 2 +  3 X 1 X 2 + e µ Y =  0 +  1 X 1 +  2 X 2 +  3 X 1 X 2

©2006 Thomson/South-Western 46 Interaction Effects Figure µYµYµYµY X1X1X1X1 (a) |1|2 µ Y = X 1 µ Y = X 1 X 2 = 2 X 2 = 5 µ Y = X 1 µ Y = X 1 X 2 = 2 X 2 = 5 60 – 50 – 40 – 30 – 20 – 10 – 60 – 50 – 40 – 30 – 20 – 10 – X1X1X1X1 X1X1X1X1 (b)(b) |1|1|2|2 µYµYµYµY µYµYµYµY

©2006 Thomson/South-Western 47 Quadratic and Second-Order Models Y =  0 +  1 X 1 +  2 X e Quadratic Effects Y =  0 +  1 X 1 +  2 X 2 +  3 X 1 X 2 +  4 X  5 X e Complete Second-Order Models Y =  0 +  1 X 1 +  2 X 2 +  3 X 3 +  4 X 1 X 2 +  5 X 2 X 3 +  6 X 2 X 3 +  7 X  8 X  9 X e

©2006 Thomson/South-Western 48 Financial Example Figure 14.27

©2006 Thomson/South-Western 49 Financial Example Figure 14.28

©2006 Thomson/South-Western 50 Financial Example Figure 14.29

©2006 Thomson/South-Western 51 Financial Example Figure 14.30