We think you have liked this presentation. If you wish to download it, please recommend it to your friends in any social system. Share buttons are a little bit lower. Thank you!
Presentation is loading. Please wait.
Published byJada Jocelyn
Modified about 1 year ago
Copyright © 2009 Pearson Education, Inc. Chapter 29 Multiple Regression
Copyright © 2009 Pearson Education, Inc. Slide 1- 3 Just Do It The method of least squares can be expanded to include more than one predictor. The method is known as multiple regression. For simple regression we found the Least Squares solution, the one whose coefficients made the sum of the squared residuals as small as possible. For multiple regression, we’ll do the same thing, but this time with more coefficients.
Copyright © 2009 Pearson Education, Inc. Slide 1- 4 Just Do It (cont.) You should recognize most of the numbers in the following example (%Body Fat) of a multiple regression table:
Copyright © 2009 Pearson Education, Inc. Slide 1- 5 So What’s New? The meaning of the coefficients in the regression model has changed in a subtle but important way. Multiple regression is an extraordinarily versatile calculation, underlying many widely used Statistics methods. Multiple regression offers our first glimpse into statistical methods that use more than two quantitative variables.
Copyright © 2009 Pearson Education, Inc. Slide 1- 6 What Multiple Regression Coefficients Mean We said that height might be important in predicting body fat in men. What’s the relationship between %Body Fat and Height in men? Here’s the scatterplot:
Copyright © 2009 Pearson Education, Inc. Slide 1- 7 What Multiple Regression Coefficients Mean (cont.) It doesn’t look like Height tells us much about %Body Fat. Or does it? The coefficient of Height in the multiple regression model was statistically significant, so it did contribute to the multiple regression model. How can this be? The multiple regression coefficient of Height takes account of the other predictor (Waist size) in the regression model.
Copyright © 2009 Pearson Education, Inc. Slide 1- 8 What Multiple Regression Coefficients Mean (cont.) For example, when we restrict our attention to men with waist sizes between 36 and 38 inches (points in blue), we can see a relationship between %Body Fat and Height:
Copyright © 2009 Pearson Education, Inc. Slide 1- 9 What Multiple Regression Coefficients Mean (cont.) So, overall there’s little relationship between %Body Fat and Height, but when we focus on particular waist sizes there is a relationship. This relationship is conditional because we’ve restricted our set to only those men with a certain range of waist sizes. For men with that waist size, an extra inch of height is associated with a decrease of about 0.60% in body fat. If that relationship is consistent for each Waist size, then the multiple regression coefficient will estimate it.
Copyright © 2009 Pearson Education, Inc. Slide What Multiple Regression Coefficients Mean (cont.) The following is a partial regression plot shows the coefficient of Height in the regression model has a slope equal to the coefficient value in the multiple regression model:
Copyright © 2009 Pearson Education, Inc. Slide The Multiple Regression Model For a multiple regression with k predictors, the model is: The assumptions and conditions for the multiple regression model sound nearly the same as for simple regression, but with more variables in the model, we’ll have to make a few changes.
Copyright © 2009 Pearson Education, Inc. Slide Assumptions and Conditions Linearity Assumption: Straight Enough Condition: Check the scatterplot for each candidate predictor variable—the shape must not be obviously curved or we can’t consider that predictor in our multiple regression model. Independence Assumption: Randomization Condition: The data should arise from a random sample or randomized experiment. Also, check the residuals plot - the residuals should appear to be randomly scattered.
Copyright © 2009 Pearson Education, Inc. Slide Assumptions and Conditions (cont.) Equal Variance Assumption: Does the Plot Thicken? Condition: Check the residuals plot - the spread of the residuals should be uniform.
Copyright © 2009 Pearson Education, Inc. Slide Assumptions and Conditions (cont.) Normality Assumption: Nearly Normal Condition: Check a histogram of the residuals—the distribution of the residuals should be unimodal and symmetric, and the Normal probability plot should be straight.
Copyright © 2009 Pearson Education, Inc. Slide Assumptions and Conditions (cont.) Summary of the checks of conditions in order: 1. Check the Straight Enough Condition. 2. If the scatterplots are straight enough, fit a multiple regression model to the data. 3. Find the residuals and predicted values. 4. Make and check a scatterplot of the residuals against the predicted values.
Copyright © 2009 Pearson Education, Inc. Slide Assumptions and Conditions (cont.) Summary of the checks of conditions in order: 5. Think about how the data were collected. 6. If the conditions check out this far, feel free to interpret the regression model and use it for prediction. 7. If you wish to test hypotheses about the coefficients or about the overall regression, then make a histogram and Normal probability plot of the residuals to check the Nearly Normal Condition.
Copyright © 2009 Pearson Education, Inc. Slide Multiple Regression Inference I: I Thought I Saw an ANOVA Table… Now that we have more than one predictor, there’s an overall test we should consider before we do more inference on the coefficients. We ask the global question “Is this multiple regression model any good at all?” We test The F-statistic and associated P-value from the ANOVA table are used to answer our question.
Copyright © 2009 Pearson Education, Inc. Slide Multiple Regression Inference II: Testing the Coefficients Once we check the F-test and reject the null hypothesis, we can move on to checking the test statistics for the individual coefficients. For each coefficient, we test If the assumptions and conditions are met (including the Nearly Normal Condition), these ratios follow a Student’s t-distribution.
Copyright © 2009 Pearson Education, Inc. Slide Multiple Regression Inference II: Testing the Coefficients (cont.) We can also find a confidence interval for each coefficient:
Copyright © 2009 Pearson Education, Inc. Slide Multiple Regression Inference II: Testing the Coefficients (cont.) Keep in mind that the meaning of a multiple regression coefficient depends on all the other predictors in the multiple regression model. If we fail to reject the null hypothesis for a multiple regression coefficient, it does not mean that the corresponding predictor variable has no linear relationship to y. It means that the corresponding predictor contributes nothing to modeling y after allowing for all the other predictors.
Copyright © 2009 Pearson Education, Inc. Slide How’s That, Again? It looks like each coefficient in the multiple regression tells us the effect of its associated predictor on the response variable. But, that is not so. The coefficient of a predictor in a multiple regression depends as much on the other predictors as it does on the predictor itself.
Copyright © 2009 Pearson Education, Inc. Slide Comparing Multiple Regression Models How do we know that some other choice of predictors might not provide a better model? What exactly would make an alternative model better? These questions are not easy—there’s no simple measure of the success of a multiple regression model.
Copyright © 2009 Pearson Education, Inc. Slide Comparing Multiple Regression Models (cont.) Regression models should make sense. Predictors that are easy to understand are usually better choices than obscure variables. Similarly, if there is a known mechanism by which a predictor has an effect on the response variable, that predictor is usually a good choice for the regression model. The simple answer is that we can’t know whether we have the best possible model.
Copyright © 2009 Pearson Education, Inc. Slide Adjusted R 2 There is another statistic in the full regression table called the adjusted R 2. This statistic is a rough attempt to adjust for the simple fact that when we add another predictor to a multiple regression, the R 2 can’t go down and will most likely get larger. This fact makes it difficult to compare alternative regression models that have different numbers of predictors.
Copyright © 2009 Pearson Education, Inc. Slide Adjusted R 2 (cont.) The formula for R 2 is while the formula for adjusted R 2 is
Copyright © 2009 Pearson Education, Inc. Slide Adjusted R 2 (cont.) Because the mean squares are sums of squares divided by their degrees of freedom, they are adjusted for the number of predictors in the model. As a result, the adjusted R 2 value won’t necessarily increase when a new predictor is added to the multiple regression model. That’s fine, but adjusted R 2 no longer tells the fraction of variability accounted for by the model—it isn’t even bounded by 0 and 100%.
Copyright © 2009 Pearson Education, Inc. Slide Adjusted R 2 (cont.) Comparing alternative regression models is a challenge, especially when they have different numbers of predictors. Adjusted R 2 is one way to help you choose your model. But, don’t use it as the sole decision criterion when you compare different regression models.
Copyright © 2009 Pearson Education, Inc. Slide What Can Go Wrong? Interpreting Coefficients Don’t claim to “hold everything else constant” for a single individual. Don’t interpret regression causally. Be cautious about interpreting a regression model as predictive. Don’t think that the sign of a coefficient is special. If a coefficient’s t-statistic is not significant, don’t interpret it at all.
Copyright © 2009 Pearson Education, Inc. Slide What Else Can Go Wrong? Don’t fit a linear regression to data that aren’t straight. Watch out for the plot thickening. Make sure the errors are nearly Normal. Watch out for high-influence points and outliers.
Copyright © 2009 Pearson Education, Inc. Slide What have we learned? There are many similarities between simple and multiple regression: The assumptions and conditions are essentially the same. R 2 still gives us the fraction of the total variation in y accounted for by the model. s e is still the standard deviation of the residuals. The degrees of freedom follows the same rule: n minus the number of parameters estimated. The regression table produced by any statistics package shows a row for each coefficient, giving its estimate, a standard error, a t-statistic, and a P-value. If all conditions are met, we can test each coefficient against the null hypothesis.
Copyright © 2009 Pearson Education, Inc. Slide What have we learned? (cont.) We have also learned some new things: We can perform an overall test of whether the multiple regression model provides a better summary for y than its mean by using the F- distribution. We learned that R 2 may not be appropriate for comparing multiple regression models with different numbers of predictors. The adjusted R 2 is one approach to this problem.
Copyright © 2010 Pearson Education, Inc. Slide
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Chapter 12: More About Regression Section 12.1 Inference for Linear Regression.
T-tests, ANOVA and regression Methods for Dummies February 1 st 2006 Jon Roiser and Predrag Petrovic.
Things to do in Lecture 1 Outline basic concepts of causality Go through the ideas & principles underlying ordinary least squares (OLS) estimation Derive.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Lecture Slides Elementary Statistics Eleventh Edition and the Triola.
Copyright © 2010 Pearson Education, Inc. Slide The correlation between two scores X and Y equals 0.8. If both the X scores and the Y scores are converted.
SECTION 10.2 Estimating a Population Mean. Whats the difference between what we did in Section 10.1 and what we are beginning in Section 10.2? In reality,
Probability and Statistics Representation of Data Measures of Center for Data Simple Analysis of Data.
Copyright © 2010 Pearson Education, Inc. Slide A least squares regression line was fitted to the weights (in pounds) versus age (in months) of a.
1 Simple Linear Regression 1. review of least squares procedure 2. inference for least squares lines.
Least Squares Regression Fitting a Line to Bivariate Data.
X,Y scatterplot These are plots of X,Y coordinates showing each individual's or sample's score on two variables. When plotting data this way we are usually.
MULTIPLE REGRESSION ANALYSIS. ENGR. DIVINO AMOR P. RIVERA STATISTICAL COORDINATION OFFICER I NSO LA UNION CONTENTS Table for the types of Multiple Regression.
21-1 Copyright 2010 McGraw-Hill Australia Pty Ltd PowerPoint slides to accompany Croucher, Introductory Mathematics and Statistics, 5e Chapter 21 Hypothesis.
Agresti/Franklin Statistics, 1 of 141 Chapter 12 Multiple Regression Learn…. T o use Multiple Regression Analysis to predict a response variable using.
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Copyright © 2011 by Pearson Education, Inc. All rights reserved Statistics for the Behavioral and Social Sciences: A Brief Course Fifth Edition Arthur.
The. of and a to in is you that it he for.
SADC Course in Statistics Assumptions underlying regression analysis (Session 05)
Fitness Chances selected (%) a b d c strength post pre Age Activity r = 0.57 Linear Models and Effect Magnitudes for Research, Clinical and Practical.
Chapter 14 Inference for Distributions of Categorical Variables: Chi-Square Procedures AP Statistics Hamilton/Mann.
T-test Charmaine Bird. What is the T-test The t-test assesses whether the means of two groups are statistically different from each other. This analysis.
Chapter 15 ANOVA. Comparing Means for Several Populations When we wish to test for differences in means for only 1 or 2 populations, we use one- or two-sample.
Week 2 – PART III POST-HOC TESTS. POST HOC TESTS When we get a significant F test result in an ANOVA test for a main effect of a factor with more than.
Chapter 4 To accompany Quantitative Analysis for Management, Eleventh Edition, by Render, Stair, and Hanna Power Point slides created by Brian Peterson.
Presentation of Data Tables and graphs are convenient for presenting data. They present the data in an organized format, enabling the reader to find information.
1 SESSION 2 ANOVA and regression. 2 Only the starting point In ANOVA, the rejection of the null hypothesis leaves many questions unanswered. Further analysis.
T tests, ANOVAs and regression Tom Jenkins Ellen Meierotto SPM Methods for Dummies 2007.
© 2016 SlidePlayer.com Inc. All rights reserved.