Download presentation

1
**12 Multiple Linear Regression CHAPTER OUTLINE**

Multiple Linear Regression Model Confidence Intervals in Multiple Linear Regression Introduction Least squares estimation of the parameters Use of t-tests Confidence interval on the mean response Matrix approach to multiple linear regression Prediction of New Observations Model Adequacy Checking Properties of the least squares estimators Residual analysis Hypothesis Tests in Multiple Linear Regression Influential observations Aspects of Multiple Regression Modeling Test for significance of regression Tests on individual regression coefficients & subsets of coefficients Polynomial regression models Categorical regressors & indicator variables Selection of variables & model building Multicollinearity Dear Instructor: This file is an adaptation of the 4th edition slides for the 5th edition. It will be replaced as slides are developed following the style of the Chapters 1-7 slides.

2
**Learning Objectives for Chapter 12**

After careful study of this chapter, you should be able to do the following: Use multiple regression techniques to build empirical models to engineering and scientific data. Understand how the method of least squares extends to fitting multiple regression models. Assess regression model adequacy. Test hypotheses and construct confidence intervals on the regression coefficients. Use the regression model to estimate the mean response, and to make predictions and to construct confidence intervals and prediction intervals. Build regression models with polynomial terms. Use indicator variables to model categorical regressors. Use stepwise regression and other model building techniques to select the appropriate set of variables for a regression model.

3
**12-1: Multiple Linear Regression Models**

Introduction Many applications of regression analysis involve situations in which there are more than one regressor variable. A regression model that contains more than one regressor variable is called a multiple regression model.

4
**12-1: Multiple Linear Regression Models**

Introduction For example, suppose that the effective life of a cutting tool depends on the cutting speed and the tool angle. A possible multiple regression model could be where Y – tool life x1 – cutting speed x2 – tool angle

5
**12-1: Multiple Linear Regression Models**

Introduction Figure 12-1 (a) The regression plane for the model E(Y) = x1 + 7x2. (b) The contour plot

6
**12-1: Multiple Linear Regression Models**

Introduction

7
**12-1: Multiple Linear Regression Models**

Introduction Figure 12-2 (a) Three-dimensional plot of the regression model E(Y) = x1 + 7x2 + 5x1x2. (b) The contour plot

8
**12-1: Multiple Linear Regression Models**

Introduction Figure 12-3 (a) Three-dimensional plot of the regression model E(Y) = x1 + 7x2 – 8.5x12 – 5x22 + 4x1x2. (b) The contour plot

9
**12-1: Multiple Linear Regression Models**

Least Squares Estimation of the Parameters

10
**12-1: Multiple Linear Regression Models**

Least Squares Estimation of the Parameters The least squares function is given by The least squares estimates must satisfy

11
**12-1: Multiple Linear Regression Models**

Least Squares Estimation of the Parameters The least squares normal Equations are The solution to the normal Equations are the least squares estimators of the regression coefficients.

12
**12-1: Multiple Linear Regression Models**

Example 12-1

13
**12-1: Multiple Linear Regression Models**

Example 12-1

14
**12-1: Multiple Linear Regression Models**

Figure 12-4 Matrix of scatter plots (from Minitab) for the wire bond pull strength data in Table 12-2.

15
**12-1: Multiple Linear Regression Models**

Example 12-1

16
**12-1: Multiple Linear Regression Models**

Example 12-1

17
**12-1: Multiple Linear Regression Models**

Example 12-1

18
**12-1: Multiple Linear Regression Models**

Matrix Approach to Multiple Linear Regression Suppose the model relating the regressors to the response is In matrix notation this model can be written as

19
**12-1: Multiple Linear Regression Models**

Matrix Approach to Multiple Linear Regression where

20
**12-1: Multiple Linear Regression Models**

Matrix Approach to Multiple Linear Regression We wish to find the vector of least squares estimators that minimizes: The resulting least squares estimate is

21
**12-1: Multiple Linear Regression Models**

Matrix Approach to Multiple Linear Regression

22
**12-1: Multiple Linear Regression Models**

Example 12-2

23
Example 12-2

24
**12-1: Multiple Linear Regression Models**

Example 12-2

25
**12-1: Multiple Linear Regression Models**

Example 12-2

26
**12-1: Multiple Linear Regression Models**

Example 12-2

27
**12-1: Multiple Linear Regression Models**

Example 12-2

29
**12-1: Multiple Linear Regression Models**

Estimating 2 An unbiased estimator of 2 is

30
**12-1: Multiple Linear Regression Models**

Properties of the Least Squares Estimators Unbiased estimators: Covariance Matrix:

31
**12-1: Multiple Linear Regression Models**

Properties of the Least Squares Estimators Individual variances and covariances: In general,

32
**12-2: Hypothesis Tests in Multiple Linear Regression**

Test for Significance of Regression The appropriate hypotheses are The test statistic is

33
**12-2: Hypothesis Tests in Multiple Linear Regression**

Test for Significance of Regression

34
**12-2: Hypothesis Tests in Multiple Linear Regression**

Example 12-3

35
**12-2: Hypothesis Tests in Multiple Linear Regression**

Example 12-3

36
**12-2: Hypothesis Tests in Multiple Linear Regression**

Example 12-3

37
**12-2: Hypothesis Tests in Multiple Linear Regression**

Example 12-3

38
**12-2: Hypothesis Tests in Multiple Linear Regression**

R2 and Adjusted R2 The coefficient of multiple determination For the wire bond pull strength data, we find that R2 = SSR/SST = / = Thus, the model accounts for about 98% of the variability in the pull strength response.

39
**12-2: Hypothesis Tests in Multiple Linear Regression**

R2 and Adjusted R2 The adjusted R2 is The adjusted R2 statistic penalizes the analyst for adding terms to the model. It can help guard against overfitting (including regressors that are not really useful)

40
**12-2: Hypothesis Tests in Multiple Linear Regression**

Tests on Individual Regression Coefficients and Subsets of Coefficients The hypotheses for testing the significance of any individual regression coefficient:

41
**12-2: Hypothesis Tests in Multiple Linear Regression**

Tests on Individual Regression Coefficients and Subsets of Coefficients The test statistic is Reject H0 if |t0| > t/2,n-p. This is called a partial or marginal test

42
**12-2: Hypothesis Tests in Multiple Linear Regression**

Example 12-4

43
**12-2: Hypothesis Tests in Multiple Linear Regression**

Example 12-4

44
**12-2: Hypothesis Tests in Multiple Linear Regression**

The general regression significance test or the extra sum of squares method: We wish to test the hypotheses:

45
**12-2: Hypothesis Tests in Multiple Linear Regression**

A general form of the model can be written: where X1 represents the columns of X associated with 1 and X2 represents the columns of X associated with 2

46
**12-2: Hypothesis Tests in Multiple Linear Regression**

For the full model: If H0 is true, the reduced model is

47
**12-2: Hypothesis Tests in Multiple Linear Regression**

The test statistic is: Reject H0 if f0 > f,r,n-p The test in Equation (12-32) is often referred to as a partial F-test

48
**12-2: Hypothesis Tests in Multiple Linear Regression**

Example 12-6

49
**12-2: Hypothesis Tests in Multiple Linear Regression**

Example 12-6

50
**12-2: Hypothesis Tests in Multiple Linear Regression**

Example 12-6

51
**12-3: Confidence Intervals in Multiple Linear Regression**

Confidence Intervals on Individual Regression Coefficients Definition

52
**12-3: Confidence Intervals in Multiple Linear Regression**

Example 12-7

53
**12-3: Confidence Intervals in Multiple Linear Regression**

Confidence Interval on the Mean Response The mean response at a point x0 is estimated by The variance of the estimated mean response is

54
**12-3: Confidence Intervals in Multiple Linear Regression**

Confidence Interval on the Mean Response Definition

55
**12-3: Confidence Intervals in Multiple Linear Regression**

Example 12-8

56
**12-3: Confidence Intervals in Multiple Linear Regression**

Example 12-8

57
**12-4: Prediction of New Observations**

A point estimate of the future observation Y0 is A 100(1-)% prediction interval for this future observation is

58
**12-4: Prediction of New Observations**

Figure 12-5 An example of extrapolation in multiple regression

59
**12-4: Prediction of New Observations**

Example 12-9

60
**12-5: Model Adequacy Checking**

Residual Analysis Example 12-10 Figure 12-6 Normal probability plot of residuals

61
**12-5: Model Adequacy Checking**

Residual Analysis Example 12-10

62
**12-5: Model Adequacy Checking**

Residual Analysis Example 12-10 Figure 12-7 Plot of residuals

63
**12-5: Model Adequacy Checking**

Residual Analysis Example 12-10 Figure 12-8 Plot of residuals against x1.

64
**12-5: Model Adequacy Checking**

Residual Analysis Example 12-10 Figure 12-9 Plot of residuals against x2.

65
**12-5: Model Adequacy Checking**

Residual Analysis

66
**12-5: Model Adequacy Checking**

Residual Analysis The variance of the ith residual is

67
**12-5: Model Adequacy Checking**

Residual Analysis

68
**12-5: Model Adequacy Checking**

Influential Observations Figure A point that is remote in x-space.

69
**12-5: Model Adequacy Checking**

Influential Observations Cook’s distance measure

70
**12-5: Model Adequacy Checking**

Example 12-11

71
**12-5: Model Adequacy Checking**

Example 12-11

72
**12-6: Aspects of Multiple Regression Modeling**

Polynomial Regression Models

73
**12-6: Aspects of Multiple Regression Modeling**

Example 12-12

74
**12-6: Aspects of Multiple Regression Modeling**

Example 12-11 Figure Data for Example

75
Example 12-12

76
**12-6: Aspects of Multiple Regression Modeling**

Example 12-12

77
**12-6: Aspects of Multiple Regression Modeling**

Categorical Regressors and Indicator Variables Many problems may involve qualitative or categorical variables. The usual method for the different levels of a qualitative variable is to use indicator variables. For example, to introduce the effect of two different operators into a regression model, we could define an indicator variable as follows:

78
**12-6: Aspects of Multiple Regression Modeling**

Example 12-13

79
**12-6: Aspects of Multiple Regression Modeling**

Example 12-13

80
**12-6: Aspects of Multiple Regression Modeling**

Example 12-13

81
Example 12-12

82
**12-6: Aspects of Multiple Regression Modeling**

Example 12-13

83
**12-6: Aspects of Multiple Regression Modeling**

Example 12-13

84
**12-6: Aspects of Multiple Regression Modeling**

Selection of Variables and Model Building

85
**12-6: Aspects of Multiple Regression Modeling**

Selection of Variables and Model Building All Possible Regressions – Example 12-14

86
**12-6: Aspects of Multiple Regression Modeling**

Selection of Variables and Model Building All Possible Regressions – Example 12-14

87
**12-6: Aspects of Multiple Regression Modeling**

Selection of Variables and Model Building All Possible Regressions – Example 12-14 Figure A matrix of Scatter plots from Minitab for the Wine Quality Data.

89
**12-6.3: Selection of Variables and Model Building - Stepwise Regression**

Example 12-14

90
**12-6.3: Selection of Variables and Model Building - Backward Regression**

Example 12-14

91
**12-6: Aspects of Multiple Regression Modeling**

Multicollinearity Variance Inflation Factor (VIF)

92
**12-6: Aspects of Multiple Regression Modeling**

Multicollinearity The presence of multicollinearity can be detected in several ways. Two of the more easily understood of these are:

93
**Important Terms & Concepts of Chapter 12**

All possible regressions Analysis of variance test in multiple regression Categorical variables Confidence intervals on the mean response Cp statistic Extra sum of squares method Hidden extrapolation Indicator variables Inference (test & intervals) on individual model parameters Influential observations Model parameters & their interpretation in multiple regression Multicollinearity Multiple regression Outliers Polynomial regression model Prediction interval on a future observation PRESS statistic Residual analysis & model adequacy checking Significance of regression Stepwise regression & related methods Variance Inflation Factor (VIF)

Similar presentations

© 2019 SlidePlayer.com Inc.

All rights reserved.

To make this website work, we log user data and share it with processors. To use this website, you must agree to our Privacy Policy, including cookie policy.

Ads by Google