Univariate Time series - 2 Methods of Economic Investigation Lecture 19.

Slides:



Advertisements
Similar presentations
FINANCIAL TIME-SERIES ECONOMETRICS SUN LIJIAN Feb 23,2001.
Advertisements

VAR Models Gloria González-Rivera University of California, Riverside
The Simple Regression Model
Vector Autoregressive Models
Using SAS for Time Series Data
Part II – TIME SERIES ANALYSIS C5 ARIMA (Box-Jenkins) Models
Non-stationary data series
Economics 20 - Prof. Anderson1 Time Series Data y t =  0 +  1 x t  k x tk + u t 2. Further Issues.
Model specification (identification) We already know about the sample autocorrelation function (SAC): Properties: Not unbiased (since a ratio between two.
Unit Roots & Forecasting
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 7: Box-Jenkins Models – Part II (Ch. 9) Material.
Time Series Building 1. Model Identification
Omitted Variable Bias Methods of Economic Investigation Lecture 7 1.
R. Werner Solar Terrestrial Influences Institute - BAS Time Series Analysis by means of inference statistical methods.
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
L18: CAPM1 Lecture 18: Testing CAPM The following topics will be covered: Time Series Tests –Sharpe (1964)/Litner (1965) version –Black (1972) version.
8.4 Weighted Least Squares Estimation Before the existence of heteroskedasticity-robust statistics, one needed to know the form of heteroskedasticity -Het.
1 Ka-fu Wong University of Hong Kong Modeling Cycles: MA, AR and ARMA Models.
Economics Prof. Buckles1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Modeling Cycles By ARMA
Prof. Dr. Rainer Stachuletz
Specific to General Modelling The traditional approach to econometrics modelling was as follows: 1.Start with an equation based on economic theory. 2.Estimate.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Prediction and model selection
Financial Econometrics
Economics 20 - Prof. Anderson
Inference about a Mean Part II
1 Ka-fu Wong University of Hong Kong Some Final Words.
Non-Seasonal Box-Jenkins Models
Econ 140 Lecture 191 Autocorrelation Lecture 19. Econ 140 Lecture 192 Today’s plan Durbin’s h-statistic Autoregressive Distributed Lag model & Finite.
Economics Prof. Buckles
Autocorrelation Lecture 18 Lecture 18.
Ordinary Least Squares
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
ARMA models Gloria González-Rivera University of California, Riverside
STAT 497 LECTURE NOTES 2.
Linear Stationary Processes. ARMA models. This lecture introduces the basic linear models for stationary processes. Considering only stationary processes.
TIME SERIES ANALYSIS Time Domain Models: Red Noise; AR and ARMA models LECTURE 7 Supplementary Readings: Wilks, chapters 8.
Instrumental Variables: Problems Methods of Economic Investigation Lecture 16.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Lecture 7: Forecasting: Putting it ALL together. The full model The model with seasonality, quadratic trend, and ARMA components can be written: Ummmm,
Autoregressive Integrated Moving Average (ARIMA) Popularly known as the Box-Jenkins methodology.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Definitions Random Signal Analysis (Review) Discrete Random Signals Random.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Maximum Likelihood Estimation Methods of Economic Investigation Lecture 17.
Chapter Three TWO-VARIABLEREGRESSION MODEL: THE PROBLEM OF ESTIMATION
Instrumental Variables: Introduction Methods of Economic Investigation Lecture 14.
K. Ensor, STAT Spring 2004 Memory characterization of a process How would the ACF behave for a process with no memory? What is a short memory series?
Problems with the Durbin-Watson test
STAT 497 LECTURE NOTE 11 VAR MODELS AND GRANGER CAUSALITY 1.
Time Series Basics Fin250f: Lecture 8.1 Spring 2010 Reading: Brooks, chapter
3.Analysis of asset price dynamics 3.1Introduction Price – continuous function yet sampled discretely (usually - equal spacing). Stochastic process – has.
STAT 497 LECTURE NOTE 9 DIAGNOSTIC CHECKS 1. After identifying and estimating a time series model, the goodness-of-fit of the model and validity of the.
Ch14: Linear Least Squares 14.1: INTRO: Fitting a pth-order polynomial will require finding (p+1) coefficients from the data. Thus, a straight line (p=1)
The Box-Jenkins (ARIMA) Methodology
AUTOCORRELATION 1 Assumption B.5 states that the values of the disturbance term in the observations in the sample are generated independently of each other.
Time Series Analysis PART II. Econometric Forecasting Forecasting is an important part of econometric analysis, for some people probably the most important.
Stationarity and Unit Root Testing Dr. Thomas Kigabo RUSUHUZWA.
4-1 MGMG 522 : Session #4 Choosing the Independent Variables and a Functional Form (Ch. 6 & 7)
EC 827 Module 2 Forecasting a Single Variable from its own History.
Dr. Thomas Kigabo RUSUHUZWA
Financial Econometrics Lecture Notes 4
VAR models and cointegration
Ch8 Time Series Modeling
CHAPTER 16 ECONOMIC FORECASTING Damodar Gujarati
Goodness of Fit The sum of squared deviations from the mean of a variable can be decomposed as follows: TSS = ESS + RSS This decomposition can be used.
Simple Linear Regression
VAR Models Gloria González-Rivera University of California, Riverside
Presentation transcript:

Univariate Time series - 2 Methods of Economic Investigation Lecture 19

Last Time  Concepts that are useful Stationarity Ergodicity Ergodic Theorem Autocovariance Generating Function Lag Operators  Time Series Processes AR(p) MA(q)

Today’s Class  Building up to estimation Wold Decomposition Estimating with exogenous, serially correlated errors Testing for Lag Length

Refresher  Stationarity: Some persistence but not too much  Ergodicity: Persistence dies out entirely over some finite period of time  Square Summability (assumption for MA process) with parameter θ such that  Invertibility (assumption for AR process) with parameter φ which has roots λ such that

ARMA process  In general can have a process with both AR and MA components  A general ARMA(p, q) process in our lag function notation this looks like:a(L)x t = b(L)ε t For example, we may have an ARMA(2, 1) x t – (φ 1 x t-1 + φ 2 x t-3 ) = ε t + θ 1 ε t-1 (1 - φ 1 L – φ 2 L 2 ) x t = (1 + θ 1 L) ε t  If the process is invertible then we can rewrite this as: x t =a(L) -1 b(L) ε t

Why Focus on ARMA processes  Define the range of ARMA processes (invertible AR lag polynomials, square summable MA lag polynomials) which can rely on convergence theorems  any time series that is covariance stationary, has a linear ARMA representation.

Information Sets  At time t-n Everything for time t-n and before is known Everything at time t is unknown Information set Ω t-n  Define E t-n (ε t ) = E[ε t | Ω t-n ] Distinct from E[ε t ] because we know previous values of ε’s up until t-n For example, suppose n = 1 and ε t = π ε t-1 +η,  E (ε t ) =0 for all t so it’s a mean zero process  E t-1 (ε t ) =π ε t-1

Recalling the CEF  Define the linear conditional expectation function CEF(a | b) which is the linear project, i.e. the fitted values of a regression of a on b. i.e. a = βb  This is distinct from the general expectations operator in that it is imposing a linear form of the conditional expectation function.

Wold Decomposition Theorem - 1  Formally the Wold Decomposition Theorem says that:  Any mean zero weakly stationary process {x t } can be represented in the form  This comes with some properties for each term…

Wold Decomposition Theorem - 2  Where  ε t ≡ x t − CEF(x t | x t-1, x t-2,...,x 0 ).  Properties of ε t CEF (ε t |x t−1, x t−2,... x 0 )=0, E(ε t x t−j ) = 0, E(ε t ) = 0, E(ε t 2 ) = σ 2 for all t, and E(ε t ε s ) = 0 for all t ≠s  The MA polynomial is invertible  The parameters θ is square summable  {θ j } and {ε s } are unique.  η t is linearly deterministic i.e. η t = CEF(η t |x t−1,....).

A note on the Wold Decomposition  Much of the properties come directly from our assumptions tha the process is weakly stationary  While it says mean zero process, remember we can de-mean our data so most processes can be represented in this format.

Uses of Wold Form  This theorem is extremely useful because it returns time-series processes back to our standard OLS model. Notice that we’ve relaxed some of the conditions for the Gauss-Markov theorem to hold.  the Wold MA(∞) representation is unique. if two time series have the same Wold representation, then they are the same time series This on true only up to second moments in linear forecasting 

Emphasis on Linearity  although CEF(ε t | x t−j ) = 0, can have E(ε t | x t−j ) ≠ 0 with nonlinear projections  If the true x t is not generated by linear combinations of past x t plus a shock, then the Wold shocks ( ε’s) will be different from the true shocks.  The uniqueness result only states that the Wold representation is the unique linear representation where the shocks are linear forecast errors.

Estimating with Serially Correlated Errors  Suppose that we have: Y t = βX t + ε t E[ε t | x t ] = 0, E[ε t 2 | x t ]=σ 2 E[ ε t ε t-k ] = γ k for k≠0 and so define E[ε t ε k ] = σ 2 Γ We could consistently estimate β but our standard errors would be incorrect making it difficult to do inference. Just a heteroskedasticity problem which we have already seen with random effects  Use feasible GLS to estimate weights and then re-estimate OLS to obtain efficient standard errors.

Endogenous Lagged Regressors  May be the case that either the dependent variable or the regressor should enter the estimating equation in lag values too  Suppose we were estimating t Y t = β 0 X t +β 1 X t-1 + … β k X t-k + ε t. We think that these X’s are correlated with Y up to some lag length k We think these X’s are correlated with each other (e.g. some underlying AR process) but we’re not sure how many lags to include

Naive Test  Include lags going very far back r >> k  test the longest lag coefficient β r = 0 and see if that is significant. If not, drop it and keep going.  Problems: Practically, the longer lags you take, the more data you make unusable because it doesn’t have enough time periods to construct the lags. doesn’t allow lag t-6 but exclude lag t-3. The theoretical issue is that we will reject the null 5 percent of the time, even if it’s true (or whatever the significance of the test is).

More sophisticated testing  Can be a bit more sophisticated comparing restricted and unrestricted models define p max as some long lag length greater than the expected relevant lag length In general, we do not test our p max but as before, as p  p max the sample size decreases. Define ε j = Y t = β 0 X t +β 1 X t-1 + … β j X j and let N be the sample size. We therefore could imagine trying to minimize the sum of squared residual:

Cost Functions  Intuition: c(. ) is a penalty for adding additional parameters thus we try to pick the best specification using that cost function to penalize inclusion of extra but irrelevant lags. Akaike (AIC): c(n) = 2  the AIC criterion is not well-founded in theory and will be biased in finite samples  the bias will tend to overstate the true lag length Bayesian: c(n) = log(n)  the BIC will converge to the true p.

Return to Likelihood Ratio Tests  The minimization problem is just likelihood ratio test  To see this, compare lag length j to lag length k. We can write:  Define constant LR test Constant: Declining in N

Next Time  Multivariate Time Series Testing for Unit Roots Cointegration  Returning to Causal Effects Impulse Response Functions Forecasting