Generalized Regression Model Based on Greene’s Note 15 (Chapter 8)

Slides:



Advertisements
Similar presentations
The Simple Linear Regression Model Specification and Estimation Hill et al Chs 3 and 4.
Advertisements

Tests of Static Asset Pricing Models
Properties of Least Squares Regression Coefficients
Multiple Regression Analysis
Classical Linear Regression Model
General Linear Model With correlated error terms  =  2 V ≠  2 I.
The Simple Regression Model
CHAPTER 3: TWO VARIABLE REGRESSION MODEL: THE PROBLEM OF ESTIMATION
Part 12: Asymptotics for the Regression Model 12-1/39 Econometrics I Professor William Greene Stern School of Business Department of Economics.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Random effects estimation RANDOM EFFECTS REGRESSIONS When the observed variables of interest are constant for each individual, a fixed effects regression.
The General Linear Model. The Simple Linear Model Linear Regression.
8.4 Weighted Least Squares Estimation Before the existence of heteroskedasticity-robust statistics, one needed to know the form of heteroskedasticity -Het.
Maximum likelihood (ML) and likelihood ratio (LR) test
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Simple Linear Regression
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
12.3 Correcting for Serial Correlation w/ Strictly Exogenous Regressors The following autocorrelation correction requires all our regressors to be strictly.
Classical Linear Regression Model Normality Assumption Hypothesis Testing Under Normality Maximum Likelihood Estimator Generalized Least Squares.
Least Squares Asymptotics Convergence of Estimators: Review Least Squares Assumptions Least Squares Estimator Asymptotic Distribution Hypothesis Testing.
Chapter 15 Panel Data Analysis.
Review.
Topic4 Ordinary Least Squares. Suppose that X is a non-random variable Y is a random variable that is affected by X in a linear fashion and by the random.
Part 7: Regression Extensions [ 1/59] Econometric Analysis of Panel Data William Greene Department of Economics Stern School of Business.
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Economics Prof. Buckles
Maximum likelihood (ML)
Econometrics I Professor William Greene Stern School of Business
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Part 14: Generalized Regression 14-1/46 Econometrics I Professor William Greene Stern School of Business Department of Economics.
1 MADE WHAT IF SOME OLS ASSUMPTIONS ARE NOT FULFILED?
Inference issues in OLS
Part 5: Random Effects [ 1/54] Econometric Analysis of Panel Data William Greene Department of Economics Stern School of Business.
What does it mean? The variance of the error term is not constant
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
Generalised method of moments approach to testing the CAPM Nimesh Mistry Filipp Levin.
1 We will now look at the properties of the OLS regression estimators with the assumptions of Model B. We will do this within the context of the simple.
EC 532 Advanced Econometrics Lecture 1 : Heteroscedasticity Prof. Burak Saltoglu.
KNN Ch. 3 Diagnostics and Remedial Measures Applied Regression Analysis BUSI 6220.
CLASSICAL NORMAL LINEAR REGRESSION MODEL (CNLRM )
Part 4A: GMM-MDE[ 1/33] Econometric Analysis of Panel Data William Greene Department of Economics Stern School of Business.
MathematicalMarketing Slide 5.1 OLS Chapter 5: Ordinary Least Square Regression We will be discussing  The Linear Regression Model  Estimation of the.
Computacion Inteligente Least-Square Methods for System Identification.
Econometrics III Evgeniya Anatolievna Kolomak, Professor.
Estimation Econometría. ADE.. Estimation We assume we have a sample of size T of: – The dependent variable (y) – The explanatory variables (x 1,x 2, x.
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
Heteroscedasticity Chapter 8
Ch5 Relaxing the Assumptions of the Classical Model
Esman M. Nyamongo Central Bank of Kenya
REGRESSION DIAGNOSTIC II: HETEROSCEDASTICITY
Evgeniya Anatolievna Kolomak, Professor
Chapter 3: TWO-VARIABLE REGRESSION MODEL: The problem of Estimation
Basic Econometrics Chapter 4: THE NORMALITY ASSUMPTION:
The regression model in matrix form
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Chengyuan Yin School of Mathematics
Serial Correlation and Heteroscedasticity in
Econometrics Chengyaun yin School of Mathematics SHUFE.
Simple Linear Regression
OVERVIEW OF LINEAR MODELS
Econometrics I Professor William Greene Stern School of Business
Econometrics I Professor William Greene Stern School of Business
Econometrics Chengyuan Yin School of Mathematics.
Serial Correlation and Heteroscedasticity in
Econometrics I Professor William Greene Stern School of Business
Econometrics I Professor William Greene Stern School of Business
Presentation transcript:

Generalized Regression Model Based on Greene’s Note 15 (Chapter 8)

Generalized Regression Model Setting: The classical linear model assumes that E[  ] = Var[  ] =  2 I. That is, observations are uncorrelated and all are drawn from a distribution with the same variance. The generalized regression (GR) model allows the variances to differ across observations and allows correlation across observations.

Implications The assumption that Var[  ] =  2 I is used to derive the result Var[b] =  2 (XX) -1. If it is not true, then the use of s 2 (XX) -1 to estimate Var[b] is inappropriate. The assumption was used to derive most of our test statistics, so they must be revised as well. Least squares gives each observation a weight of 1/n. But, if the variances are not equal, then some observations are more informative than others. Least squares is based on simple sums, so the information that one observation might provide about another is never used.

GR Model The generalized regression model: y = X  + , E[  |X] = 0, Var[  |X] =  2 . Regressors are well behaved. We consider some examples with Trace  = n. (This is a normalization with no content.) Leading Cases –Simple heteroscedasticity –Autocorrelation –Panel data and heterogeneity more generally.

Least Squares Still unbiased. (Proof did not rely on  ) For consistency, we need the true variance of b, Var[b|X] = E[(b-β)(b-β)’|X] = (X’X) -1 E[X’εε’X] (X’X) -1 =  2 (X’X) -1 X  X (X’X) -1. Divide all 4 terms by n. If the middle one converges to a finite matrix of constants, we have the result, so we need to examine (1/n)X  X = (1/n)  i  j  ij x i x j. This will be another assumption of the model. Asymptotic normality? Easy for heteroscedasticity case, very difficult for autocorrelation case.

Robust Covariance Matrix Robust estimation: How to estimate Var[b|X] =  2 (X’X) -1 X  X (X’X) -1 for the LS b? The distinction between estimating  2  an n by n matrix and estimating  2 X  X =  2  i  j  ij x i x j For modern applied econometrics, –The White estimator –Newey-West.

The White Estimator

Newey-West Estimator

Generalized Least Squares A transformation of the model: P =  -1/2. P’P =  -1 Py = PX  + P  or y* = X*  +  *. Why? E[  *  *’|X*]= PE[  ’|X*]P’ = PE[  ’|X]P’ = σ 2 P  P’ = σ 2  -1/2  -1/2 = σ 2  0 = σ 2 I

Generalized Least Squares Aitken theorem. The Generalized Least Squares estimator, GLS. Py = PX  + P  or y* = X*  +  *. E[  *  *’|X*]= σ 2 I Use ordinary least squares in the transformed model. Satisfies the Gauss – Markov theorem. b* = (X*’X*) -1 X*’y*

Generalized Least Squares Efficient estimation of  and, by implication, the inefficiency of least squares b. = (X*’X*) -1 X*’y* = (X’P’PX) -1 X’P’Py = (X’Ω -1 X) -1 X’Ω -1 y ≠ b. is efficient, so by construction, b is not.

Asymptotics for GLS Asymptotic distribution of GLS. (NOTE: We apply the full set of results of the classical model to the transformed model). Unbiasedness Consistency - “well behaved data” Asymptotic distribution Test statistics

Unbiasedness

Consistency

Asymptotic Normality

Asymptotic Normality (Cont.)

Test Statistics (Assuming Known Ω) With known Ω, apply all familiar results to the transformed model. With normality, t and F statistics apply to least squares based on Py and PX With asymptotic normality, use Wald statistics and the chi-squared distribution, still based on the transformed model.

Generalized (Weighted) Least Squares:Heteroscedasticity

Autocorrelation  t =  t-1 + u t (‘First order autocorrelation.’ How does this come about?) Assume -1 <  < 1. Why? u t = ‘nonautocorrelated white noise’  t =  t-1 + u t (the autoregressive form) =  (  t-2 + u t-1 ) + u t =... (continue to substitute) = u t +  u t-1 +  2 u t-2 +  3 u t = (the moving average form) (Some observations about modeling time series.)

Autocorrelation

Autocovariances

Autocorrelation Matrix

Generalized Least Squares

The Autoregressive Transformation

Unknown  The problem (of course),  is unknown. For now, we will consider two methods of estimation: –Two step, or feasible estimation. Estimate  first, then do GLS. Emphasize - same logic as White and Newey-West. We don’t need to estimate . We need to find a matrix that behaves the same as (1/n)X  -1 X. –Properties of the feasible GLS estimator Maximum likelihood estimation of ,  2, and  all at the same time. –Joint estimation of all parameters. Fairly rare. Some generalities… –We will examine two applications: Harvey’s model of heteroscedasticity and Beach-MacKinnon on the first order autocorrelation model

Specification  must be specified first. A full unrestricted  contains n(n+1)/2 - 1 parameters. (Why minus 1? Remember, tr(  ) = n, so one element is determined.)  is generally specified in terms of a few parameters. Thus,  =  (  ) for some small parameter vector . It becomes a question of estimating . Examples:

Heteroscedasticity: Harvey’s Model Var[  i | X] =  2 exp(  z i ) Cov[  i,  j | X] = 0 e.g.: z i = firm size e.g.: z i = a set of dummy variables (e.g., countries) (The groupwise heteroscedasticity model.) [  2  ] = diagonal [exp(  +  z i )],  = log(  2 )

AR(1) Model of Autocorrelation

Two Step Estimation The general result for estimation when  is estimated. GLS uses [X  -1 X]X  -1 y which converges in probability to . We seek a vector which converges to the same thing that this does. Call it “Feasible GLS” or FGLS, based on [X X]X y The object is to find a set of parameters such that [X X]X y - [X  -1 X]X  -1 y  0

Feasible GLS

Two Step FGLS (Theorem 8.5) To achieve full efficiency, we do not need an efficient estimate of the parameters in , only a consistent one. Why?

Harvey’s Model Examine Harvey’s model once again. Methods of estimation: Two step FGLS: Use the least squares residuals to estimate , then use {X[ Ω(  )] -1 X} -1 X’[ Ω(  )] -1 y to estimate . Full maximum likelihood estimation. Estimate all parameters simultaneously. A handy result due to Oberhofer and Kmenta - the “zig-zag” approach. Examine a model of groupwise heteroscedasticity.

Harvey’s Model for Groupwise Heteroscedasticity Groupwise sample, y ig, x ig,… N groups, each with N g observations. Var[ε ig ] = σ g 2 Let d ig = 1 if observation i,g is in group j, 0 else. = group dummy variable. Var[ε ig ] = σ g 2 exp(θ 2 d 2 + … θ G d G ) Var 1 = σ g 2, Var 2 = σ g 2 exp(θ 2 ) and so on.

Estimating Variance Components OLS is still consistent: Est.Var 1 = e 1 ’e 1 /N 1 estimates σ g 2 Est.Var 2 = e 2 ’e 2 /N 2 estimates σ g 2 exp(θ 2 ) Estimator of θ 2 is ln[(e 2 ’e 2 /N 2 )/(e 1 ’e 1 /N 1 )] (1) Now use FGLS – weighted least squares Recompute residuals using WLS slopes (2) Recompute variance estimators Iterate to a solution… between (1) and (2)