Multiple Regression Analysis

Slides:



Advertisements
Similar presentations
Properties of Least Squares Regression Coefficients
Advertisements

Multiple Regression Analysis
The Simple Regression Model
1 Javier Aparicio División de Estudios Políticos, CIDE Otoño Regresión.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 7. Specification and Data Problems.
Economics 20 - Prof. Anderson1 Stationary Stochastic Process A stochastic process is stationary if for every collection of time indices 1 ≤ t 1 < …< t.
3.3 Omitted Variable Bias -When a valid variable is excluded, we UNDERSPECIFY THE MODEL and OLS estimates are biased -Consider the true population model:
Econ 140 Lecture 61 Inference about a Mean Lecture 6.
10 Further Time Series OLS Issues Chapter 10 covered OLS properties for finite (small) sample time series data -If our Chapter 10 assumptions fail, we.
Stationary Stochastic Process
8. Heteroskedasticity We have already seen that homoskedasticity exists when the error term’s variance, conditional on all x variables, is constant: Homoskedasticity.
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
Assumption MLR.3 Notes (No Perfect Collinearity)
The Simple Linear Regression Model: Specification and Estimation
1 Research Method Lecture 4 (Ch5) OLS Asymptotics ©
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Hypothesis Testing.
Economics Prof. Buckles1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
2.5 Variances of the OLS Estimators
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 7. Specification and Data Problems.
Simple Linear Regression
Prof. Dr. Rainer Stachuletz
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Chapter 4 Multiple Regression.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Econ 140 Lecture 71 Classical Regression Lecture 7.
Multiple Regression Analysis
4. Multiple Regression Analysis: Estimation -Most econometric regressions are motivated by a question -ie: Do Canadian Heritage commercials have a positive.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Statistical Background
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 3. Asymptotic Properties.
Economics 20 - Prof. Anderson
The Simple Regression Model
EC Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
Review of Probability and Statistics
1Prof. Dr. Rainer Stachuletz Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Economics Prof. Buckles
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Intermediate Econometrics
2-1 MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4)
1 SAMPLE MEAN and its distribution. 2 CENTRAL LIMIT THEOREM: If sufficiently large sample is taken from population with any distribution with mean  and.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Properties of OLS How Reliable is OLS?. Learning Objectives 1.Review of the idea that the OLS estimator is a random variable 2.How do we judge the quality.
1 Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved. CH5 Multiple Regression Analysis: OLS Asymptotic 
2.4 Units of Measurement and Functional Form -Two important econometric issues are: 1) Changing measurement -When does scaling variables have an effect.
Application 3: Estimating the Effect of Education on Earnings Methods of Economic Investigation Lecture 9 1.
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
405 ECONOMETRICS Domodar N. Gujarati Prof. M. El-Sakka
Chapter 2 Statistical Background. 2.3 Random Variables and Probability Distributions A variable X is said to be a random variable (rv) if for every real.
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
1 We will now look at the properties of the OLS regression estimators with the assumptions of Model B. We will do this within the context of the simple.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
5. Consistency We cannot always achieve unbiasedness of estimators. -For example, σhat is not an unbiased estimator of σ -It is only consistent -Where.
Review of Statistics.  Estimation of the Population Mean  Hypothesis Testing  Confidence Intervals  Comparing Means from Different Populations  Scatterplots.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
Economics 20 - Prof. Anderson1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
Regression Overview. Definition The simple linear regression model is given by the linear equation where is the y-intercept for the population data, is.
Multiple Regression Analysis: Estimation
Multiple Regression Analysis
Basic Econometrics Chapter 4: THE NORMALITY ASSUMPTION:
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Tutorial 1: Misspecification
Multiple Regression Analysis: OLS Asymptotics
Multiple Regression Analysis: OLS Asymptotics
Multiple Regression Analysis
Presentation transcript:

Multiple Regression Analysis y = b0 + b1x1 + b2x2 + . . . bkxk + u 3. Asymptotic Properties Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Consistency Under the Gauss-Markov assumptions OLS is BLUE, but in other cases it won’t always be possible to find unbiased estimators In those cases, we may settle for estimators that are consistent, meaning as n  ∞, the distribution of the estimator collapses to the parameter value Economics 20 - Prof. Anderson

Sampling Distributions as n  n1 < n2 < n3 n2 n1 b1 Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Consistency of OLS Under the Gauss-Markov assumptions, the OLS estimator is consistent (and unbiased) Consistency can be proved for the simple regression case in a manner similar to the proof of unbiasedness Will need to take probability limit (plim) to establish consistency Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Proving Consistency Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson A Weaker Assumption For unbiasedness, we assumed a zero conditional mean – E(u|x1, x2,…,xk) = 0 For consistency, we can have the weaker assumption of zero mean and zero correlation – E(u) = 0 and Cov(xj,u) = 0, for j = 1, 2, …, k Without this assumption, OLS will be biased and inconsistent! Economics 20 - Prof. Anderson

Deriving the Inconsistency Just as we could derive the omitted variable bias earlier, now we want to think about the inconsistency, or asymptotic bias, in this case Economics 20 - Prof. Anderson

Asymptotic Bias (cont) So, thinking about the direction of the asymptotic bias is just like thinking about the direction of bias for an omitted variable Main difference is that asymptotic bias uses the population variance and covariance, while bias uses the sample counterparts Remember, inconsistency is a large sample problem – it doesn’t go away as add data Economics 20 - Prof. Anderson

Large Sample Inference Recall that under the CLM assumptions, the sampling distributions are normal, so we could derive t and F distributions for testing This exact normality was due to assuming the population error distribution was normal This assumption of normal errors implied that the distribution of y, given the x’s, was normal as well Economics 20 - Prof. Anderson

Large Sample Inference (cont) Easy to come up with examples for which this exact normality assumption will fail Any clearly skewed variable, like wages, arrests, savings, etc. can’t be normal, since a normal distribution is symmetric Normality assumption not needed to conclude OLS is BLUE, only for inference Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Central Limit Theorem Based on the central limit theorem, we can show that OLS estimators are asymptotically normal Asymptotic Normality implies that P(Z<z)F(z) as n , or P(Z<z)  F(z) The central limit theorem states that the standardized average of any population with mean m and variance s2 is asymptotically ~N(0,1), or Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Asymptotic Normality Economics 20 - Prof. Anderson

Asymptotic Normality (cont) Because the t distribution approaches the normal distribution for large df, we can also say that Note that while we no longer need to assume normality with a large sample, we do still need homoskedasticity Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson Asymptotic Standard Errors If u is not normally distributed, we sometimes will refer to the standard error as an asymptotic standard error, since So, we can expect standard errors to shrink at a rate proportional to the inverse of √n Economics 20 - Prof. Anderson

Lagrange Multiplier statistic Once we are using large samples and relying on asymptotic normality for inference, we can use more that t and F stats The Lagrange multiplier or LM statistic is an alternative for testing multiple exclusion restrictions Because the LM statistic uses an auxiliary regression it’s sometimes called an nR2 stat Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson LM Statistic (cont) Suppose we have a standard model, y = b0 + b1x1 + b2x2 + . . . bkxk + u and our null hypothesis is H0: bk-q+1 = 0, ... , bk = 0 First, we just run the restricted model Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson LM Statistic (cont) With a large sample, the result from an F test and from an LM test should be similar Unlike the F test and t test for one exclusion, the LM test and F test will not be identical Economics 20 - Prof. Anderson

Asymptotic Efficiency Estimators besides OLS will be consistent However, under the Gauss-Markov assumptions, the OLS estimators will have the smallest asymptotic variances We say that OLS is asymptotically efficient Important to remember our assumptions though, if not homoskedastic, not true Economics 20 - Prof. Anderson