Multiple Regression Analysis: Estimation

Slides:



Advertisements
Similar presentations
Properties of Least Squares Regression Coefficients
Advertisements

Multiple Regression Analysis
The Simple Regression Model
CHAPTER 3: TWO VARIABLE REGRESSION MODEL: THE PROBLEM OF ESTIMATION
3.3 Omitted Variable Bias -When a valid variable is excluded, we UNDERSPECIFY THE MODEL and OLS estimates are biased -Consider the true population model:
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
Assumption MLR.3 Notes (No Perfect Collinearity)
8.4 Weighted Least Squares Estimation Before the existence of heteroskedasticity-robust statistics, one needed to know the form of heteroskedasticity -Het.
The Simple Linear Regression Model: Specification and Estimation
Chapter 10 Simple Regression.
Simple Linear Regression
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Chapter 4 Multiple Regression.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
The Simple Regression Model
Lecture 2 (Ch3) Multiple linear regression
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Ordinary Least Squares
3. Multiple Regression Analysis: Estimation -Although bivariate linear regressions are sometimes useful, they are often unrealistic -SLR.4, that all factors.
7.1 Multiple Regression More than one explanatory/independent variable This makes a slight change to the interpretation of the coefficients This changes.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
1 Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved. CH5 Multiple Regression Analysis: OLS Asymptotic 
Chapter Three TWO-VARIABLEREGRESSION MODEL: THE PROBLEM OF ESTIMATION
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin The Two-Variable Model: Hypothesis Testing chapter seven.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Class 5 Multiple Regression CERAM February-March-April 2008 Lionel Nesta Observatoire Français des Conjonctures Economiques
Quantitative Methods. Bivariate Regression (OLS) We’ll start with OLS regression. Stands for  Ordinary Least Squares Regression. Relatively basic multivariate.
1 AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Part II: Theory and Estimation of Regression Models Chapter 5: Simple Regression Theory.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Regression Overview. Definition The simple linear regression model is given by the linear equation where is the y-intercept for the population data, is.
Lecture 6 Feb. 2, 2015 ANNOUNCEMENT: Lab session will go from 4:20-5:20 based on the poll. (The majority indicated that it would not be a problem to chance,
Multiple Regression Analysis: Estimation
Announcements Reminder: Exam 1 on Wed.
The simple linear regression model and parameter estimation
Chapter 4: Basic Estimation Techniques
Regression Analysis AGEC 784.
Ch. 2: The Simple Regression Model
Chapter 9 Multiple Linear Regression
THE LINEAR REGRESSION MODEL: AN OVERVIEW
Chapter 5: The Simple Regression Model
Welcome to Econ 420 Applied Regression Analysis
The Simple Regression Model
Chapter 3: TWO-VARIABLE REGRESSION MODEL: The problem of Estimation
Multiple Regression Analysis
Ch. 2: The Simple Regression Model
Correlation and Simple Linear Regression
Inference about the Slope and Intercept
Chapter 6: MULTIPLE REGRESSION ANALYSIS
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Inference about the Slope and Intercept
Correlation and Simple Linear Regression
Simple Linear Regression
Simple Linear Regression and Correlation
Heteroskedasticity.
The Simple Regression Model
Multiple Regression Analysis: OLS Asymptotics
Multiple Regression Analysis: OLS Asymptotics
Tutorial 6 SEG rd Oct..
Linear Regression Summer School IFPRI
Introduction to Regression
MGS 3100 Business Analysis Regression Feb 18, 2016
Multiple Regression Analysis
Regression Models - Introduction
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Multiple Regression Analysis: Estimation y = b0 + b1x1 + b2x2 + . . . bkxk + u Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Motivation for Multiple Regression The model with two Independent Variables Since exper/avginc is correlative with edu/expend, it is useful to explicityly put it in the equation for the unbiasedness of the estimation on b1. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

does not exist in 3.4 any more. Why? How to interpret the parameters in equation 3.4? Ceteris Paribus effect does not exist in 3.4 any more. Why? Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

cons inc consmax Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

The key assumption about how u is related to x1 and x2 is as equation 3.5. What it means? Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

The Model with K Independent Variables Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Parallels with Simple Regression b0 is still the intercept b1 to bk all called slope parameters u is still the error term (or disturbance) Still need to make a zero conditional mean assumption, so now assume that E(u|x1,x2, …,xk) = 0 Still minimizing the sum of squared residuals, so have k+1 first order conditions Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Mechanics and Interpretation of Ordinary Least Squares Obtaining the OLS Estimates. The method of OLS chooses the estimates to minimize the sum of squared residuals, that is to make 3.10 as Small as possible. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

OLS First Order Condition SRF SSR OLS First Order Condition Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Interpreting the OLS Regression Equation Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Example 3.1 Determinants of College GPA Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Example 3.2 Hourly Wage Equation Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

On the Meaning of “Holding Other Factors Fixed” in Multiple Regression The power of multiple regression analysis is that it provides this ceteris paribus interpretation even though the data have not been collected in a ceteris paribus fashion. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Changing More than One Independent Variable Simutaneously Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

OLS Fitted Values and Residuals Normally, Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

A “Partialling Out” Interpretation of Multiple Regression Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Proof: Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Comparison of Simple and Multiple Regression Estimates SRF MRF Two cases when the parameter estimates of SRF and MRF are equal: Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Example 3.3 Participation in 401 (K) Pension Plans. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Goodness-of-Fit Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Example 3.4 Determinants of College GPA Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Example 3.5 Explaining Arrest Records Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

The Expected Value of the OLS Estimators The Assumptions for Multiple Regression Assumption MLR.1: the population model is linear in parameters as y = b0 + b1x1 + b2x2+…+ bkxk+ u Assumption MLR.2: we can use a random sample of size n, {(xi1, xi2 ,…, xik, yi): i=1, 2, …, n}, from the population model. Thus we can write the sample model yi = b0 + b1xi1 + b2xi2+…+ bkxik+ ui Assumption MLR.3: E(u| x1, x2 ,…, xk) = 0 Model Misspecification: cons = b0 + b1inc + b2inc2+ u Omitting an important factor; Measurement Error. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Assumption MLR.4: No Perfect Collinearity In the sample (and therefore in the population), none of the independent variables is constant, and there are no exact linear relationships among the independents variables. Assumption MLR.4 does allow the independent variables to be correlated; they just cannot be perfectly correlated. One variable is a constant multiple of another, the same variable measured in different units. One independent variable can be expressed as an exact linear function of the two or more of the other independent variable. The sample size, n, is too small in relation to the number of parameters being estimated. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Proof: Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Including Irrelevant Variables in a Regression Model What happens if we include variables in our specification that don’t belong? There is no effect on our parameter estimate, and OLS remains unbiased, but have undesirable effects on the variances of the OLS estimators. What if we exclude a variable from our specification that does belong? OLS will usually be biased Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Omitted Variable Bias Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Omitted Variable Bias (cont) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Omitted Variable Bias (cont) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Omitted Variable Bias (cont) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Omitted Variable Bias Summary Two cases where bias is equal to zero b2 = 0, that is x2 doesn’t really belong in model x1 and x2 are uncorrelated in the sample. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Summary of Direction of Bias Corr(x1, x2) > 0 Corr(x1, x2) < 0 b2 > 0 Positive bias Negative bias b2 < 0 Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

The More General Case Technically, can only sign the bias for the more general case if all of the included x’s are uncorrelated Typically, then, we work through the bias assuming the x’s are uncorrelated, as a useful guide even if this assumption is not strictly true Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Example: Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Variance of the OLS Estimators Now we know that the sampling distribution of our estimate is centered around the true parameter Want to think about how spread out this distribution is Much easier to think about this variance under an additional assumption, so Assume Var(u|x1, x2,…, xk) = s2 (Homoskedasticity) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Variance of OLS (cont) Let x stand for (x1, x2,…xk) Assumption MLR.5: Var(u|x) = s2 also implies that Var(y| x) = s2 The 4 assumptions for unbiasedness, plus this homoskedasticity assumption are known as the Gauss-Markov assumptions Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Variance of OLS (cont) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Proof: Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Components of OLS Variances The error variance: a larger s2 implies a larger variance for the OLS estimators The total sample variation: a larger SSTj implies a smaller variance for the estimators Linear relationships among the independent variables: a larger Rj2 implies a larger variance for the estimators Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Misspecified Models Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Misspecified Models (cont) While the variance of the estimator is smaller for the misspecified model, unless b2 = 0 the misspecified model is biased As the sample size grows, the variance of each estimator shrinks to zero, making the variance difference less important Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Estimating the Error Variance We don’t know what the error variance, s2, is, because we don’t observe the errors, ui What we observe are the residuals, ûi We can use the residuals to form an estimate of the error variance Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Error Variance Estimate (cont) df = n – (k + 1), or df = n – k – 1 df (i.e. degrees of freedom) is the (number of observations) – (number of estimated parameters) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

The Gauss-Markov Theorem Given our 5 Gauss-Markov Assumptions it can be shown that OLS is “BLUE” Best Linear Unbiased Estimator Thus, if the assumptions hold, use OLS Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.