1 Ka-fu Wong University of Hong Kong Forecasting with Regression Models.

Slides:



Advertisements
Similar presentations
FINANCIAL TIME-SERIES ECONOMETRICS SUN LIJIAN Feb 23,2001.
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Exam 1 review: Quizzes 1-6.
The Simple Regression Model
The Multiple Regression Model.
Vector Autoregressive Models
Econ 427 lecture 24 slides Forecast Evaluation Byron Gangnes.
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Unit Roots & Forecasting
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 7: Box-Jenkins Models – Part II (Ch. 9) Material.
R. Werner Solar Terrestrial Influences Institute - BAS Time Series Analysis by means of inference statistical methods.
Vector Error Correction and Vector Autoregressive Models
10 Further Time Series OLS Issues Chapter 10 covered OLS properties for finite (small) sample time series data -If our Chapter 10 assumptions fail, we.
The Simple Linear Regression Model: Specification and Estimation
Economics Prof. Buckles1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Chapter 10 Simple Regression.
Modeling Cycles By ARMA
Simple Linear Regression
Time Series Model Estimation Materials for this lecture Read Chapter 15 pages 30 to 37 Lecture 7 Time Series.XLS Lecture 7 Vector Autoregression.XLS.
1 Ka-fu Wong University of Hong Kong Volatility Measurement, Modeling, and Forecasting.
CHAPTER 4 ECONOMETRICS x x x x x Multiple Regression = more than one explanatory variable Independent variables are X 2 and X 3. Y i = B 1 + B 2 X 2i +
1 Ka-fu Wong University of Hong Kong Pulling Things Together.
Prediction and model selection
Ordinary least squares regression (OLS)
Economics 20 - Prof. Anderson
1 Ka-fu Wong University of Hong Kong Some Final Words.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Demand Estimation & Forecasting
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
Regression and Correlation Methods Judy Zhong Ph.D.
Introduction to Linear Regression and Correlation Analysis
Correlation and Regression
Chapter 11 Simple Regression
Introduction to Linear Regression
Lecture 7: Forecasting: Putting it ALL together. The full model The model with seasonality, quadratic trend, and ARMA components can be written: Ummmm,
Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2.
10. Basic Regressions with Times Series Data 10.1 The Nature of Time Series Data 10.2 Examples of Time Series Regression Models 10.3 Finite Sample Properties.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
Copyright © 2005 by the McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Managerial Economics Thomas Maurice eighth edition Chapter 7.
Scatter Diagrams scatter plot scatter diagram A scatter plot is a graph that may be used to represent the relationship between two variables. Also referred.
Correlation & Regression Analysis
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Module 4 Forecasting Multiple Variables from their own Histories EC 827.
Byron Gangnes Econ 427 lecture 18 slides Multivariate Modeling (cntd)
1 Ka-fu Wong University of Hong Kong A Brief Review of Probability, Statistics, and Regression for Forecasting.
Analysis of financial data Anders Lundquist Spring 2010.
EC 827 Module 2 Forecasting a Single Variable from its own History.
Lecture 12 Time Series Model Estimation Materials for lecture 12 Read Chapter 15 pages 30 to 37 Lecture 12 Time Series.XLSX Lecture 12 Vector Autoregression.XLSX.
Forecasting. Model with indicator variables The choice of a forecasting technique depends on the components identified in the time series. The techniques.
1 Autocorrelation in Time Series data KNN Ch. 12 (pp )
Dr. Thomas Kigabo RUSUHUZWA
Time Series Econometrics
Financial Econometrics Lecture Notes 4
Regression Analysis AGEC 784.
Lecture 12 Time Series Model Estimation
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Linear Regression.
VAR models and cointegration
Ch8 Time Series Modeling
ECO 400-Time Series Econometrics VAR MODELS
Simple Linear Regression
CHAPTER 16 ECONOMIC FORECASTING Damodar Gujarati
6-1 Introduction To Empirical Models
VAR Models Gloria González-Rivera University of California, Riverside
Multivariate Modeling (intro)
Presentation transcript:

1 Ka-fu Wong University of Hong Kong Forecasting with Regression Models

2 Linear regression models Endogenous variable Exogenous variables Explanatory variables Rule, rather than exception: all variables are endogenous.

3 Conditional forecasting The h-step ahead forecast of y given some assumed h-step- ahead value of x T+h. Assumed h-step-ahead value of the exogenous variables Call it scenario analysis or contingency analysis – based on some assumed h-step-ahead value of the exogenous variables.

4 Uncertainty of Forecast Specification uncertainty / error: our models are only approximation (since no one knows the truth). E.g., we adopt an AR(1) model but the truth is AR(2). Almost impossible to account for via a forecast interval. Parameter uncertainty / sampling error: parameters are estimated from a data sample. The estimate will always be different from the truth. The difference is called sampling error. Can account for via a forecast interval if we do the calculation carefully. Innovation uncertainty: errors that cannot be avoided even if we know the true model and true parameter. This is, unavoidable. Often account for via a forecast interval using standard softwares.

5 Quantifying the innovation and parameter uncertainty Consider the very simple case in which x has a zero mean:

6 Density forecast that accounts for parameter uncertainty ~

7 Interval forecasts that do not acknowledge parameter uncertainty

8 Interval forecasts that do acknowledge parameter uncertainty The closer x T+h * is closer to its mean, the smaller is the prediction-error variance.

9 Unconditional Forecasting Models Forecast based on some other models of x, say, by assuming x to follow an AR(1).

10 h-step-ahead forecast without modeling x explicitly Based on unconditional forecasting models Standing at time T, with observations, (x 1,y 1 ), (x 2,y 2 ),…,(x T,y T ) 1-step-ahead: y t =  0 +  1 x t-1 +  t y T+1 =  0 +  1 x T +  t 2-step-ahead: y t =  0 +  1 x t-2 +  t y T+2 =  0 +  1 x T +  t … h-step-ahead: y t =  0 +  1 x t-h +  t y T+h =  0 +  1 x T +  t

11 h-step-ahead forecast without modeling x explicitly Based on unconditional forecasting models Special cases: The model contains only time trends and seasonal components. Because these components are perfectly predictable.

12 Distributed Lags y depends on a distributed lags of past x’s Parameters to be estimated:  0,  1,…,  Nx

13 Polynomial Distributed Lags Parameters to be estimated:  0, a, b, c

14 Rational Distributed Lags Example: A(L) = a 0 + a 1 L B(L) = b 0 + b 1 L b 0 y t + b 1 y t-1 = a 0 x t + a 1 x t-1 + b 0  t + b 1  t-1 y t = [- b 1 y t-1 + a 0 x t + a 1 x t-1 + b 0  t + b 1  t-1 ]/b 0 y t = [- b 1 /b 0 ] y t-1 + [a 0 /b 0 ] x t + [a 1 /b 0 ] x t-1 +  t + [b 1 /b 0 ]  t-1

15 Regression model with AR(1) disturbance

16 ARMA(p,q) models equivalent to model with only a constant regressor and ARMA(p,q) disturbances.

17 Transfer function models A transfer function is a mathematical representation of the relation between the input and output of a system.

18 Vector Autoregressions, VAR(p) allows cross-variable dynamics VAR(1) of two variables. The variable vector consists of two elements. Regressors consist of the variable vector lagged one period only. The innovations allowed to be correlated.

19 Estimation of Vector Autoregressions Run OLS regressions equation by equation. OLS estimation turns out to have very good statistical properties when each equation has the same regressors, as in standard VARs. Otherwise, a more complicated estimation procedure called seemingly unrelated regression, which explicitly accounts for correlation across equation disturbances, would be need to obtain estimates with good statistical properties.

20 The choice order Estimation of Vector Autoregressions Use AIC and SIC.

21 Forecast Estimation of Vector Autoregressions y 1,T, y 2,T y 1,T+1, y 2,T+1 y 1,T+1, Y 2,T+1 y 1,T+2, Y 2,T+2 y 1,T+2, y 2,T+2 y 1,T+3, Y 2,T+3 y 1,T+3, y 2,T+3 Given the parameters, or parameter estimates

22 Predictive Causality Two principles Cause should occur before effect. A causal series should contain information useful for forecasting that is not available in the other series. Predictive Causality in a VAR y 2 does not cause y 1 if φ 12 =0 In a bivariate VAR, noncausality in 1-step-ahead forecast will imply noncausality in h-step-ahead forecast.

23 Predictive Causality In VAR with higher dimension, noncausality in 1-step-ahead forecast need not imply noncausality in h-step-ahead forecast. Example: Variable i may 1-step-cause variable j Variable j may 1-step-cause variable k Variable i 2-step-causes variable k but does not 1-step-cause variable k.

24 Impulse response functions All univariate ARMA(p,q) processes can be written as: We can always normalize the innovations with a constant m:

25 Impulse response functions 1 unit increase in  t ’ is equivalent to one standard deviation increase in  t. 1 unit increase in  t ’ has b 0 ’ impact on y t 1 standard deviation increase in  t has b 0  impact on y t, b 1  impact on y t, etc. Impact of  t on y t :

26 AR(1)

27 VAR(1)

28 Normalizing the VAR by the Cholesky factor If y 1 is ordered first, Example: y 1 = GDP, y 2 = Price level An innovation to GDP has effects on current GDP and price level. An innovation to price level has effects only on current price level but not current GDP.

29 Features of Cholesky decomposition The innovations of the transformed system are in standard deviation units. The current innovations in the normalized representation have can non-unit coefficients. The first equation has only one current innovation, e 1,t. The second equation has both current innovations. The normalization yields a zero covariance between the innovations.

30 Normalizing the VAR by the Cholesky factor If y 2 is ordered first, Example: y 1 = GDP, y 2 = Price level An innovation to price level has effects on current GDP and price level. An innovation to GDP has effects only on current GDP but not current price level.

31 Impulse response functions With bivariate autoregression, we can compute four sets of impulse-response functions: y 1 innovations (  1,t ) on y 1 y 1 innovations (  1,t ) on y 2 y 2 innovations (  2,t ) on y 1 y 2 innovations (  2,t ) on y 2

32 Variance decomposition How much of the h-step-ahead forecast error variance of variable i is explained by innovations to variable j, for h=1,2,…. ? With bivariate autoregression, we can compute four sets of variance decomposition: y 1 innovations (  1,t ) on y 1 y 1 innovations (  1,t ) on y 2 y 2 innovations (  2,t ) on y 1 y 2 innovations (  2,t ) on y 2

33 Example: y 1 = Housing starts, y 2 = Housing completions (1968:01 – 1996:06) Observation #1: Seasonal pattern. Observation #2: Highly cyclical with business cycles. Observation #3: Completions lag starts. group fig112 starts comps freeze(Figure112) fig112.line(d)

34 Correlogram and Ljung-Box Statistics of housing starts (1968:01 to 1991:12) freeze(Table112) starts.correl(24)

35 Correlogram and Ljung-Box Statistics of housing starts (1968:01 to 1991:12)

36 Correlogram and Ljung-Box Statistics of housing completions (1968:01 to 1991:12) freeze(Table113) comps.correl(24)

37 Correlogram and Ljung-Box Statistics of housing starts (1968:01 to 1991:12)

38 Starts and completions, sample cross- correlations freeze(Figure115) fig112.cross(24) starts comps

39 VAR regression by OLS (1) equation Table114.ls starts c starts(-1) starts(-2) starts(-3) starts(-4) comps(-1) comps(-2) comps(-3) comps(-4)

40 VAR regression by OLS (1)

41 VAR regression by OLS (1)

42 VAR regression by OLS (1)

43 VAR regression by OLS (2) equation Table116.ls comps c starts(-1) starts(-2) starts(-3) starts(-4) comps(-1) comps(-2) comps(-3) comps(-4)

44 VAR regression by OLS (2)

45 VAR regression by OLS (2)

46 VAR regression by OLS (2)

47 Predictive causality test group tbl108 comps starts freeze(Table118) tbl108.cause(4)

48 Impulse response functions (response to one standard-deviation innovations) var fig1110.ls 1 4 starts comps freeze(Figure1110) fig1110.impulse(36,m)

49 Variance decomposition freeze(Figure1111) fig1110.decomp(36,m)

50 Starts: History, 1968: :12 Forecast, 1992: :06

51 Starts: History, 1968: :12 Forecast, 1992: :06

52 Completions: History, 1968: :12 Forecast, 1992: :06

53 Completions: History, 1968: :12 Forecast, and Realization, 1992: :06

54 End