Presentation on theme: "Further Inference in the Multiple Regression Model Hill et al Chapter 8."— Presentation transcript:
Further Inference in the Multiple Regression Model Hill et al Chapter 8
The F-Test Used to test hypotheses on one or more parameters Unrestricted model: Restricted model
The F-statistic Are the differences in SSE significant? If the null hypothesis is true, then the statistic F has an F-distribution with J numerator degrees of freedom and T-K denominator degrees of freedom.
Omitted and irrelevant variables An omitted variable which is correlated with other variables in the regression will lead to bias. The omission of insignificant variables may lead to bias (remember all you have done is failed to reject a null) Including irrelevant variables will inflate the variances of the estimated parameters.
The RESET test: principle If we omit variables and they are correlated with existing variables, including a function of these variables may allow us to pick up some of the effect of the omitted variables. If we can artificially improve the model by including powers of the predictions of the model, then a better functional form may exist. Overall: if we can improve a model by including powers of the predictions the model is inadequate.
The RESET test: practice In both cases the null is of no mis-specification
The RESET test: example The linear model is mis-specified.
Collinear Economic Variables Explanatory variables move together in systematic ways. Attribute the increase in TR that is the consequence of two types of advertising. Identify the effects of increasing input quantities when technology is of the fixed proportions type.
The consequences of collinearity Exact collinearity renders OLS inoperable. Near exact leads to increased standard errors. R 2 may be high but individual coefficients are likely to be insignificant. Estimates will be sensitive to the addition of a few observations. Accurate prediction may still be possible.
Identifying and mitigating collinearity Identifying: –Large standard errors with high R 2. –Pairwise correlation coefficients in excess of 0.8 –Auxiliary regressions. Mitigating –Additional data. –Parameter restrictions