2Introduction Assess the main ways of remedying autocorrelation Describe the problem of non-constant error termsAssess the main test for heteroskedasticityExamine the main remedies for heteroskedasticityIntroduce the multivariate approach to regression analysis
3Cochrane-Orcutt methodology This approach to remedying autocorrelation relies on the generalised difference equation approachIt is an iterative procedure, the estimates of the parameters converge towards their actual value.It assumes the autocorrelation follows a first order autoregressive process (simplest form)
5Procedure (Based on previous slide) Run regression and collect the error term (u).Run a second regression of u against u(t-1) to obtain an estimate of ρ.Form the generalised difference equation, to obtain new estimates of the parametersRe-run this equation, obtain u and repeat process until there is no significant difference from one iteration to the next.
6Restricted Version of the GDE The generalised difference equation can be written in the following form, which includes a restriction, the product of the coefficients on y(-1) and x is equal to the negative of the coefficient on x(-1) :
7Unrestricted version of the GDE In the unrestricted version of the equation in the previous slide, there are no restrictions on the values of the coefficients:
8The Common Factor TestTo determine whether the restricted or unrestricted versions are best, we can use the common factor test, under the null hypothesis that the restriction holds, the test statistic is:
9HeteroskedasticityThis occurs when a Gauss-Markov assumption is broken, it indicates non-constant variance of the error term.It is often caused by variables which have substantial differences in the values of the observations, i.e. GDP differing between the USA and CubaWith heteroskedasticity the estimator is no longer BLUE, as it is not the best.
10Tests for Heteroskedasticity Goldsfeldt-Quandt test (not used much now)Whites Test (used in E-views computer software).LM test (Similar to the LM test for heteroskedasticity)
11White’s TestThis test follows a similar pattern to the LM test for autocorrelation discussed earlier.Based on the same first equation as in slide 4, we estimate the model and collect the residual u.Square the residual to form the variance of the residualRun a secondary regression of the squared residual on the explanatory variable and this variable squared
12White’s TestThis secondary regression appears like the following:
13White’s TestCollect the R-squared statistic and multiply by the number of observations to obtain the test statistic.This follows a chi-squared distribution, the d of f equal the number of parameters in the secondary regression (i.e. 2 in above example)The null hypothesis is no heteroskedasticity.If there were 2 explanatory variables, you could include the cross product of these variables, in the secondary regression.
14Remedies for heteroskedasticity If the standard deviation of the residual is known, the heteroskedasticity can be removed by dividing the regression equation through by the standard deviation of the residual (weighted Least Squares)If this is not known, as is likely, we need to stipulate what the standard deviation is equal to. We can then divide the regression equation through by this variable to obtain a constant variance residual.
15Remedying Heteroskedasticity If we assume the error variance is proportional to x, the explanatory variable and we have the usual model:
16Remedying Heteroskedasticity If we divide our model through by the square root of x:
17Remedying Heteroskedasticity Taking the new error term, we can show it is no longer suffering from heteroskedasticity by proving it is now constant:
18Multivariate Regression Regressions with more than one explanatory variable are similar to the bi-variate case, we can interpret the coefficients and t-statistics in much the same way as before.We now have a potential problem with multicollinearity, when the explanatory variables are correlatedThe formula for the β contains an expression for the covariance between the explanatory variables.
19Multivariate Regression Analysis The regression coefficients are more likely to be accurate, the less closely related are the explanatory variablesThis allows testing for the significance of groups of variablesThe R-squared increases with the more explanatory variablesWe need to consider whether all relevant effects are included in the modelWhen estimating the t-tests, we need to account for the extra degrees of freedomThere is a trade off between the number of explanatory variables and the ease of interpreting the model
20ConclusionThe common factor test can be used to determine the best solution to the problem of autocorrelationIn the presence of heteroskedasticity, the estimator is no longer BLUE.White’s test is the most common test for heteroskedasticityWeighted Least Squares can be used to remedy the heteroskedasticity problem.