Stat 112 Notes 6 Today: –Chapters 4.2 (Inferences from a Multiple Regression Analysis)

Slides:



Advertisements
Similar presentations
Stat 112: Lecture 7 Notes Homework 2: Due next Thursday The Multiple Linear Regression model (Chapter 4.1) Inferences from multiple regression analysis.
Advertisements

Inference for Regression
Simple Linear Regression. Start by exploring the data Construct a scatterplot  Does a linear relationship between variables exist?  Is the relationship.
Objectives (BPS chapter 24)
Class 16: Thursday, Nov. 4 Note: I will you some info on the final project this weekend and will discuss in class on Tuesday.
Chapter 13 Multiple Regression
Chapter 10 Simple Regression.
Lecture 23: Tues., Dec. 2 Today: Thursday:
Class 15: Tuesday, Nov. 2 Multiple Regression (Chapter 11, Moore and McCabe).
Chapter 12 Multiple Regression
Lecture 23: Tues., April 6 Interpretation of regression coefficients (handout) Inference for multiple regression.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
1 Pertemuan 13 Uji Koefisien Korelasi dan Regresi Matakuliah: A0392 – Statistik Ekonomi Tahun: 2006.
Lecture 6 Notes Note: I will homework 2 tonight. It will be due next Thursday. The Multiple Linear Regression model (Chapter 4.1) Inferences from.
Stat 112: Lecture 8 Notes Homework 2: Due on Thursday Assessing Quality of Prediction (Chapter 3.5.3) Comparing Two Regression Models (Chapter 4.4) Prediction.
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Pengujian Parameter Koefisien Korelasi Pertemuan 04 Matakuliah: I0174 – Analisis Regresi Tahun: Ganjil 2007/2008.
Chapter Topics Types of Regression Models
Lecture 24: Thurs., April 8th
Chapter 11 Multiple Regression.
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
Simple Linear Regression Analysis
Stat 112: Lecture 13 Notes Finish Chapter 5: –Review Predictions in Log-Log Transformation. –Polynomials and Transformations in Multiple Regression Start.
Multiple Regression and Correlation Analysis
Ch. 14: The Multiple Regression Model building
Stat Notes 5 p-values for one-sided tests Caution about forecasting outside the range of the explanatory variable (Chapter 3.7.2) Fitting a linear.
Pertemua 19 Regresi Linier
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Statistics 350 Lecture 17. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
Stat 112: Lecture 9 Notes Homework 3: Due next Thursday
Chapter 7 Forecasting with Simple Regression
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
Correlation & Regression
Chapter 8 Forecasting with Multiple Regression
Copyright © 2011 Pearson Education, Inc. Multiple Regression Chapter 23.
Regression and Correlation Methods Judy Zhong Ph.D.
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Inference for regression - Simple linear regression
Chapter 13: Inference in Regression
Lecture 14 Multiple Regression Model
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Chapter 14 Introduction to Multiple Regression
Copyright © 2014, 2011 Pearson Education, Inc. 1 Chapter 23 Multiple Regression.
© 2003 Prentice-Hall, Inc.Chap 13-1 Basic Business Statistics (9 th Edition) Chapter 13 Simple Linear Regression.
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Chap 14-1 Copyright ©2012 Pearson Education, Inc. publishing as Prentice Hall Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics.
Chapter 5: Regression Analysis Part 1: Simple Linear Regression.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Lecture 10: Correlation and Regression Model.
Stat 112 Notes 5 Today: –Chapter 3.7 (Cautions in interpreting regression results) –Normal Quantile Plots –Chapter 3.6 (Fitting a linear time trend to.
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 14-1 Chapter 14 Introduction to Multiple Regression Statistics for Managers using Microsoft.
Stat 112 Notes 6 Today: –Chapter 4.1 (Introduction to Multiple Regression)
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Lesson 14 - R Chapter 14 Review. Objectives Summarize the chapter Define the vocabulary used Complete all objectives Successfully answer any of the review.
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
Stat 112 Notes 14 Assessing the assumptions of the multiple regression model and remedies when assumptions are not met (Chapter 6).
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Multiple Regression Chapter 14.
Lesson Testing the Significance of the Least Squares Regression Model.
Stat 112 Notes 8 Today: –Chapters 4.3 (Assessing the Fit of a Regression Model) –Chapter 4.4 (Comparing Two Regression Models) –Chapter 4.5 (Prediction.
Chapter 14 Introduction to Multiple Regression
Inference for Least Squares Lines
Unit 3 – Linear regression
Simple Linear Regression
Presentation transcript:

Stat 112 Notes 6 Today: –Chapters 4.2 (Inferences from a Multiple Regression Analysis)

Multiple Linear Regression Model

Predicted Values and Residuals Predicted Value for Observation with Residual for observation i = prediction error for observation i = Obtaining the predicting values and residuals in JMP: After using Fit Model to fit the multiple regression model, click the red triangle next to Response, click Save Columns and then click Predicted Values and Residuals. Columns with the predicted values and residuals will be created.

Cars that Have Unexpected Gas Mileage Given Their Weight, Horsepower, Cargo and Seating Gas guzzlers (highest residuals): 1. Mercedes Benz G500: Chevrolet Tracker: Jeep Wrangler SE: 8.39 Environment friendly (most negative residuals): 1.Volkwagen Phaeton: BMW 745i: Toyota Scion xA: -6.03

Interpretation of Regression Coefficients Gas mileage regression from car04.JMP

Partial Slopes vs. Marginal Slopes Multiple Linear Regression Model: The coefficient is a partial slope. It indicates the change in the mean of y that is associated with a one unit increase in while holding all other variables fixed. A marginal slope is obtained when we perform a simple regression with only one X, ignoring all other variables. Consequently the other variables are not held fixed.

Path Diagram X1X1 X2X2 Y

Partial vs. Marginal Slopes Example

Partial and Marginal Slopes of Length

Partial Slopes vs. Marginal Slopes: Another Example In order to evaluate the benefits of a proposed irrigation scheme in a certain region, suppose that the relation of yield Y to rainfall R is investigated over several years. Data is in rainfall.JMP.

Higher rainfall is associated with lower temperature.

Rainfall is estimated to be beneficial once temperature is held fixed. Multiple regression provides a better picture of the benefits of an irrigation scheme because temperature would be held fixed in an irrigation scheme.

Path Diagram Rainfall Temperature Yield

Inferences about Regression Coefficients Confidence intervals: confidence interval for : Degrees of freedom for t equals n-(K+1). Standard error of,, found on JMP output. Hypothesis Test: Decision rule for test: Reject H 0 if or where p-value for testing vs. is printed in JMP output under Prob>|t|.

Inference Examples Find a 95% confidence interval for ? Is seating of any help in predicting gas mileage once weight, horsepower and length have been taken into account? Carry out a test at the 0.05 significance level.

Checking the Assumptions

Plots for Checking Assumptions We can construct residual plots of each explanatory variable X k vs. the residuals. We save the residuals by clicking the red triangle next to Response after fitting the model and clicking Save Columns and then residuals. We then plot X k vs. the residuals using Fit Y by X (where Y=the residuals). We can plot a horizontal line at 0 by using Fit Y by X (it is a property of multiple linear regression that the least squares line for the regression of the residuals on any X k is a horizontal line. A useful summary of the residual plots for each explanatory variable is the Residual by Predicted plot that is automatically plotted after using Fit Model. The residual by predicted plot is a plot of the predicted values,, vs. the residuals

Checking Assumptions Linearity: –Check that in residual by predicted plot, the mean of the residuals for each range of the predicted values is about zero. –Check that in each residual plot, the mean of the residuals for each range of the explanatory variable is about zero. Constant Variance: Check that in the residual by predicted plot that for each range of the predicted values, the spread of the residuals is about the same. Normality: Make normal quantile plot of the residuals. Check that all the residuals are inside of the two confidence bands.

Residual by predicted plot suggests linearity and constant variance assumptions are okay. Plot of weight vs. residuals suggests linearity is okay. Plot of horsepower vs. residuals suggests linearity is okay.

Plot of residuals vs. seating suggests linearity is not perfect for seating. Residuals for small and high seating seem to have a mean that is smaller than 0. Plot of residuals vs. horsepower suggest linearity is okay.

Checking Normality There is a slight violation of normality. A few points are outside the lower confidence bound. We will discuss how to address violations of normality later in the course.

R-Squared (Coefficient of Determination) The coefficient of determination for multiple regression is defined as for simple linear regression: Represents percentage of variation in y that is explained by the multiple regression line. is between 0 and 1. The closer to 1, the better the fit of the regression equation to the data. For the gas mileage regression, RSquare Summary of Fit