Presentation is loading. Please wait.

Presentation is loading. Please wait.

Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.

Similar presentations


Presentation on theme: "Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes."— Presentation transcript:

1 Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes

2  9.1 Introduction  9.2 Lags in the Error Term: Autocorrelation  9.3 Estimating an AR(1) Error Model  9.4 Testing for Autocorrelation  9.5 An Introduction to Forecasting: Autoregressive Models  9.6 Finite Distributed Lags  9.7 Autoregressive Distributed Lag Models

3 Figure 9.1

4  The model so far assumes that the observations are not correlated with one another.  This is believable if one has drawn a random sample, but less likely if one has drawn observations sequentially in time  Time series observations, which are drawn at regular intervals, usually embody a structure where time is an important component. Principles of Econometrics, 3rd Edition

5  If we cannot completely model this structure in the regression function itself, then the remainder spills over into the unobserved component of the statistical model (its error) and this causes the errors be correlated with one another. Principles of Econometrics, 3rd Edition

6 Three ways to view the dynamics:

7 Figure 9.2(a) Time Series of a Stationary Variable Assume Stationarity For now

8 Figure 9.2(b) Time Series of a Nonstationary Variable that is ‘Slow Turning’ or ‘Wandering’

9 Figure 9.2(c) Time Series of a Nonstationary Variable that ‘Trends’

10 9.2.1 Area Response Model for Sugar Cane

11 Assume also

12 OK OK, constant Not zero!!!

13 Aka autocorrelation coefficient

14

15 Figure 9.3 Least Squares Residuals Plotted Against Time

16 Constant variance Null expectation of x and y

17 The existence of AR(1) errors implies:  The least squares estimator is still a linear and unbiased estimator, but it is no longer best. There is another estimator with a smaller variance.  The standard errors usually computed for the least squares estimator are incorrect. Confidence intervals and hypothesis tests that use these standard errors may be misleading.

18  As with heteroskedastic errors, you can salvage OLS when your data are autocorrelated.  Now you can use an estimator of standard errors that is robust to both heteroskedasticity and autocorrelation proposed by Newey and West.  This estimator is sometimes called HAC, which stands for heteroskedasticity autocorrelated consistent.

19  HAC is not as automatic as the heteroskedasticity consistent (HC) estimator.  With autocorrelation you have to specify how far away in time the autocorrelation is likely to be significant  The autocorrelated errors over the chosen time window are averaged in the computation of the HAC standard errors; you have to specify how many periods over which to average and how much weight to assign each residual in that average.  That weighted average is called a kernel and the number of errors to average is called bandwidth.

20  Usually, you choose a method of averaging (Bartlett kernel or Parzen kernel) and a bandwidth (nw1, nw2 or some integer). Often your software defaults to the Bartlett kernel and a bandwidth computed based on the sample size, N.  Trade-off: Larger bandwidths reduce bias (good) as well as precision (bad). Smaller bandwidths exclude more relevant autocorrelations (and hence have more bias), but use more observations to increase precision (smaller variance).  Choose a bandwidth large enough to contain the largest autocorrelations. The choice will ultimately depend on the frequency of observation and the length of time it takes for your system to adjust to shocks.

21 Sugar cane example The two sets of standard errors, along with the estimated equation are: The 95% confidence intervals for β 2 are:

22

23 Now this transformed nonlinear model has errors that are uncorrelated over time

24 It can be shown that nonlinear least squares estimation of (9.24) is equivalent to using an iterative generalized least squares estimator called the Cochrane-Orcutt procedure. Details are provided in Appendix 9A.

25  The nonlinear least squares estimator only requires that the errors be stable (not necessarily stationary).  Other methods commonly used make stronger demands on the data, namely that the errors be covariance stationary.  Furthermore, the nonlinear least squares estimator gives you an unconditional estimate of the autocorrelation parameter,, and yields a simple t-test of the hypothesis of no serial correlation.  Monte Carlo studies show that it performs well in small samples as well.

26  But nonlinear least squares requires more computational power than linear estimation, though this is not much of a constraint these days.  Nonlinear least squares (and other nonlinear estimators) use numerical methods rather than analytical ones to find the minimum of your sum of squared errors objective function. The routines that do this are iterative.

27 We should then test the above restrictions

28 9.4.1Residual Correlogram

29 9.4.1Residual Correlogram (more practically…)

30 Figure 9.4 Correlogram for Least Squares Residuals from Sugar Cane Example

31 For this nonlinear model, then, the residuals should be uncorrelated

32 Figure 9.5 Correlogram for Nonlinear Least Squares Residuals from Sugar Cane Example

33 The other way to determine whether or not your residuals are autocorrelated is to use an LM (Lagrange multiplier) test. For autocorrelation, this test is based on an auxiliary regression where lagged OLS residuals are added to the original regression equation. If the coefficient on the lagged residual is significant then you conclude that the model is autocorrelated.

34 We can derive the auxiliary regression for the LM test:

35 Centered around zero, so the power of this test now would come from The lagged error, which is what we care about…

36 This was the Breusch-Godfrey LM test for autocorrelation and it is distributed chi-sq In STATA: regress la lp estat bgodfrey

37 Adding lagged values of y can serve to eliminate the error autocorrelation In general, p should be large enough so that vt is white noise

38 Figure 9.6 Correlogram for Least Squares Residuals from AR(3) Model for Inflation Helps decide How many lags To include in the model

39

40

41

42

43

44 This model is just a generalization of the ones previously discussed. In this model you include lags of the dependent variable (autoregressive) and the contemporaneous and lagged values of independent variables as regressors (distributed lags). The acronym is ARDL(p,q) where p is the maximum distributed lag and q is the maximum autoregressive lag

45

46

47

48

49 Figure 9.7 Correlogram for Least Squares Residuals from Finite Distributed Lag Model

50

51 Figure 9.8 Correlogram for Least Squares Residuals from Autoregressive Distributed Lag Model

52

53 Figure 9.9 Distributed Lag Weights for Autoregressive Distributed Lag Model

54 Slide 9-54 Principles of Econometrics, 3rd Edition  autocorrelation  autoregressive distributed lag models  autoregressive error  autoregressive model  correlogram  delay multiplier  distributed lag weight  dynamic models  finite distributed lag  forecast error  forecasting  HAC standard errors  impact multiplier  infinite distributed lag  interim multiplier  lag length  lagged dependent variable  LM test  nonlinear least squares  sample autocorrelation function  standard error of forecast error  total multiplier  form of LM test

55 Slide 9-55 Principles of Econometrics, 3rd Edition

56 Slide 9-56 Principles of Econometrics, 3rd Edition (9A.2) (9A.1)

57 Slide 9-57 Principles of Econometrics, 3rd Edition (9A.4) (9A.3)

58 Slide 9-58 Principles of Econometrics, 3rd Edition (9A.5) (9A.6)

59 Slide 9-59 Principles of Econometrics, 3rd Edition

60 Slide 9-60 Principles of Econometrics, 3rd Edition (9B.1)

61 Slide 9-61 Principles of Econometrics, 3rd Edition (9B.2)

62 Slide 9-62 Principles of Econometrics, 3rd Edition (9B.3)

63 Figure 9A.1: Principles of Econometrics, 3rd Edition Slide 9-63

64 Figure 9A.2: Principles of Econometrics, 3rd Edition Slide 9-64

65 The Durbin-Watson bounds test.   if the test is inconclusive. Principles of Econometrics, 3rd Edition Slide 9-65

66 Slide 9-66 Principles of Econometrics, 3rd Edition

67 Slide 9-67 Principles of Econometrics, 3rd Edition (9C.2) (9C.1)

68 Slide 9-68 Principles of Econometrics, 3rd Edition (9C.3)

69 Slide 9-69 Principles of Econometrics, 3rd Edition (9C.4)

70 Slide 9-70 Principles of Econometrics, 3rd Edition

71 Slide 9-71 Principles of Econometrics, 3rd Edition (9C.6) (9C.5)

72 Slide 9-72 Principles of Econometrics, 3rd Edition (9D.2) (9D.1)

73 Figure 9A.3: Exponential Smoothing Forecasts for two alternative values of α Principles of Econometrics, 3rd Edition Slide 9-73


Download ppt "Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes."

Similar presentations


Ads by Google