Multiple Regression Analysis

Slides:



Advertisements
Similar presentations
Properties of Least Squares Regression Coefficients
Advertisements

Economics 20 - Prof. Anderson1 The Simple Regression Model y =  0 +  1 x + u.
The Simple Regression Model
The Multiple Regression Model.
Chapter 12 Simple Linear Regression
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 7. Specification and Data Problems.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
3.3 Omitted Variable Bias -When a valid variable is excluded, we UNDERSPECIFY THE MODEL and OLS estimates are biased -Consider the true population model:
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
Chapter 12 Simple Linear Regression
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
Assumption MLR.3 Notes (No Perfect Collinearity)
The Simple Linear Regression Model: Specification and Estimation
Chapter 10 Simple Regression.
2.5 Variances of the OLS Estimators
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Simple Linear Regression
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Chapter 4 Multiple Regression.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
CHAPTER 4 ECONOMETRICS x x x x x Multiple Regression = more than one explanatory variable Independent variables are X 2 and X 3. Y i = B 1 + B 2 X 2i +
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Economics 20 - Prof. Anderson
The Simple Regression Model
Lecture 2 (Ch3) Multiple linear regression
Ch. 14: The Multiple Regression Model building
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
1Prof. Dr. Rainer Stachuletz Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
8.1 Ch. 8 Multiple Regression (con’t) Topics: F-tests : allow us to test joint hypotheses tests (tests involving one or more  coefficients). Model Specification:
Chapter 11 Simple Regression
7.1 Multiple Regression More than one explanatory/independent variable This makes a slight change to the interpretation of the coefficients This changes.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin The Two-Variable Model: Hypothesis Testing chapter seven.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
Class 5 Multiple Regression CERAM February-March-April 2008 Lionel Nesta Observatoire Français des Conjonctures Economiques
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Multiple Regression Analysis: Estimation. Multiple Regression Model y = ß 0 + ß 1 x 1 + ß 2 x 2 + …+ ß k x k + u -ß 0 is still the intercept -ß 1 to ß.
Economics 20 - Prof. Anderson1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Quantitative Methods. Bivariate Regression (OLS) We’ll start with OLS regression. Stands for  Ordinary Least Squares Regression. Relatively basic multivariate.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
Regression Overview. Definition The simple linear regression model is given by the linear equation where is the y-intercept for the population data, is.
Lecture 6 Feb. 2, 2015 ANNOUNCEMENT: Lab session will go from 4:20-5:20 based on the poll. (The majority indicated that it would not be a problem to chance,
Multiple Regression.
The simple linear regression model and parameter estimation
Ch. 2: The Simple Regression Model
Multiple Regression Analysis: Estimation
The Simple Linear Regression Model: Specification and Estimation
The Simple Regression Model
Chapter 3: TWO-VARIABLE REGRESSION MODEL: The problem of Estimation
Multiple Regression Analysis
Quantitative Methods Simple Regression.
Chapter 6: MULTIPLE REGRESSION ANALYSIS
Multiple Regression.
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Simple Linear Regression
Multiple Regression Analysis: Estimation
Simple Linear Regression and Correlation
Multiple Regression Analysis: OLS Asymptotics
Multiple Regression Analysis: OLS Asymptotics
Regression Models - Introduction
Presentation transcript:

Multiple Regression Analysis y = b0 + b1x1 + b2x2 + . . . bkxk + u 1. Estimation Economics 20 - Prof. Anderson

Parallels with Simple Regression b0 is still the intercept b1 to bk all called slope parameters u is still the error term (or disturbance) Still need to make a zero conditional mean assumption, so now assume that E(u|x1,x2, …,xk) = 0 Still minimizing the sum of squared residuals, so have k+1 first order conditions Economics 20 - Prof. Anderson

Interpreting Multiple Regression Economics 20 - Prof. Anderson

A “Partialling Out” Interpretation Economics 20 - Prof. Anderson

“Partialling Out” continued Previous equation implies that regressing y on x1 and x2 gives same effect of x1 as regressing y on residuals from a regression of x1 on x2 This means only the part of xi1 that is uncorrelated with xi2 is being related to yi so we’re estimating the effect of x1 on y after x2 has been “partialled out” Economics 20 - Prof. Anderson

Simple vs Multiple Reg Estimate Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Goodness-of-Fit Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Goodness-of-Fit (continued) How do we think about how well our sample regression line fits our sample data? Can compute the fraction of the total sum of squares (SST) that is explained by the model, call this the R-squared of regression R2 = SSE/SST = 1 – SSR/SST Economics 20 - Prof. Anderson

Goodness-of-Fit (continued) Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson More about R-squared R2 can never decrease when another independent variable is added to a regression, and usually will increase Because R2 will usually increase with the number of independent variables, it is not a good way to compare models Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Assumptions for Unbiasedness Population model is linear in parameters: y = b0 + b1x1 + b2x2 +…+ bkxk + u We can use a random sample of size n, {(xi1, xi2,…, xik, yi): i=1, 2, …, n}, from the population model, so that the sample model is yi = b0 + b1xi1 + b2xi2 +…+ bkxik + ui E(u|x1, x2,… xk) = 0, implying that all of the explanatory variables are exogenous None of the x’s is constant, and there are no exact linear relationships among them Economics 20 - Prof. Anderson

Too Many or Too Few Variables What happens if we include variables in our specification that don’t belong? There is no effect on our parameter estimate, and OLS remains unbiased What if we exclude a variable from our specification that does belong? OLS will usually be biased Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Omitted Variable Bias Economics 20 - Prof. Anderson

Omitted Variable Bias (cont) Economics 20 - Prof. Anderson

Omitted Variable Bias (cont) Economics 20 - Prof. Anderson

Omitted Variable Bias (cont) Economics 20 - Prof. Anderson

Summary of Direction of Bias Corr(x1, x2) > 0 Corr(x1, x2) < 0 b2 > 0 Positive bias Negative bias b2 < 0 Economics 20 - Prof. Anderson

Omitted Variable Bias Summary Two cases where bias is equal to zero b2 = 0, that is x2 doesn’t really belong in model x1 and x2 are uncorrelated in the sample If correlation between x2 , x1 and x2 , y is the same direction, bias will be positive If correlation between x2 , x1 and x2 , y is the opposite direction, bias will be negative Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson The More General Case Technically, can only sign the bias for the more general case if all of the included x’s are uncorrelated Typically, then, we work through the bias assuming the x’s are uncorrelated, as a useful guide even if this assumption is not strictly true Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Variance of the OLS Estimators Now we know that the sampling distribution of our estimate is centered around the true parameter Want to think about how spread out this distribution is Much easier to think about this variance under an additional assumption, so Assume Var(u|x1, x2,…, xk) = s2 (Homoskedasticity) Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Variance of OLS (cont) Let x stand for (x1, x2,…xk) Assuming that Var(u|x) = s2 also implies that Var(y| x) = s2 The 4 assumptions for unbiasedness, plus this homoskedasticity assumption are known as the Gauss-Markov assumptions Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Variance of OLS (cont) Economics 20 - Prof. Anderson

Components of OLS Variances The error variance: a larger s2 implies a larger variance for the OLS estimators The total sample variation: a larger SSTj implies a smaller variance for the estimators Linear relationships among the independent variables: a larger Rj2 implies a larger variance for the estimators Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Misspecified Models Economics 20 - Prof. Anderson

Misspecified Models (cont) While the variance of the estimator is smaller for the misspecified model, unless b2 = 0 the misspecified model is biased As the sample size grows, the variance of each estimator shrinks to zero, making the variance difference less important Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Estimating the Error Variance We don’t know what the error variance, s2, is, because we don’t observe the errors, ui What we observe are the residuals, ûi We can use the residuals to form an estimate of the error variance Economics 20 - Prof. Anderson

Error Variance Estimate (cont) df = n – (k + 1), or df = n – k – 1 df (i.e. degrees of freedom) is the (number of observations) – (number of estimated parameters) Economics 20 - Prof. Anderson

The Gauss-Markov Theorem Given our 5 Gauss-Markov Assumptions it can be shown that OLS is “BLUE” Best Linear Unbiased Estimator Thus, if the assumptions hold, use OLS Economics 20 - Prof. Anderson