Stationary Stochastic Process

Slides:



Advertisements
Similar presentations
Random Processes Introduction (2)
Advertisements

Autocorrelation Functions and ARIMA Modelling
The Spectral Representation of Stationary Time Series.
Nonstationary Time Series Data and Cointegration
ELG5377 Adaptive Signal Processing
Economics 20 - Prof. Anderson1 Stationary Stochastic Process A stochastic process is stationary if for every collection of time indices 1 ≤ t 1 < …< t.
Using SAS for Time Series Data
Nonstationary Time Series Data and Cointegration Prepared by Vera Tabakova, East Carolina University.
1 MF-852 Financial Econometrics Lecture 11 Distributed Lags and Unit Roots Roy J. Epstein Fall 2003.
Unit Roots & Forecasting
10 Further Time Series OLS Issues Chapter 10 covered OLS properties for finite (small) sample time series data -If our Chapter 10 assumptions fail, we.
STAT 497 APPLIED TIME SERIES ANALYSIS
Stationary process NONSTATIONARY PROCESSES 1 In the last sequence, the process shown at the top was shown to be stationary. The expected value and variance.
Economics Prof. Buckles1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 7. Specification and Data Problems.
Prof. Dr. Rainer Stachuletz
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Multiple Regression Analysis
Multiple Regression Analysis
1Prof. Dr. Rainer Stachuletz Testing for Unit Roots Consider an AR(1): y t =  +  y t-1 + e t Let H 0 :  = 1, (assume there is a unit root) Define 
QA-3 FRM-GARP Sep-2001 Zvi Wiener Quantitative Analysis 3.
1Prof. Dr. Rainer Stachuletz Fixed Effects Estimation When there is an observed fixed effect, an alternative to first differences is fixed effects estimation.
ARIMA Forecasting Lecture 7 and 8 - March 14-16, 2011
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 3. Asymptotic Properties.
FRM Zvi Wiener Following P. Jorion, Financial Risk Manager Handbook Financial Risk Management.
Economics 20 - Prof. Anderson
Modern methods The classical approach: MethodProsCons Time series regression Easy to implement Fairly easy to interpret Covariates may be added (normalization)
The Simple Regression Model
Review of Probability and Statistics
1Prof. Dr. Rainer Stachuletz Time Series Data y t =  0 +  1 x t  k x tk + u t 2. Further Issues.
1Prof. Dr. Rainer Stachuletz Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Economics Prof. Buckles
Chapter 7 Probability and Samples: The Distribution of Sample Means
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
Review of Probability.
Time Series Analysis.
Limits and the Law of Large Numbers Lecture XIII.
10. Basic Regressions with Times Series Data 10.1 The Nature of Time Series Data 10.2 Examples of Time Series Regression Models 10.3 Finite Sample Properties.
1 Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved. CH5 Multiple Regression Analysis: OLS Asymptotic 
Week 21 Stochastic Process - Introduction Stochastic processes are processes that proceed randomly in time. Rather than consider fixed random variables.
Random processes. Matlab What is a random process?
3.Analysis of asset price dynamics 3.1Introduction Price – continuous function yet sampled discretely (usually - equal spacing). Stochastic process – has.
Auto Regressive, Integrated, Moving Average Box-Jenkins models A stationary times series can be modelled on basis of the serial correlations in it. A non-stationary.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
NONSTATIONARY PROCESSES 1 In the last sequence, the process shown at the top was shown to be stationary. The expected value and variance of X t were shown.
Economics 20 - Prof. Anderson1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Components of Time Series Su, Chapter 2, section II.
Stationarity and Unit Root Testing Dr. Thomas Kigabo RUSUHUZWA.
© 2013 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Analysis of Financial Data Spring 2012 Lecture 4: Time Series Models - 1 Priyantha Wijayatunga Department of Statistics, Umeå University
Stochastic Process - Introduction
Nonstationary Time Series Data and Cointegration
Time Series Analysis and Its Applications
Ch9 Random Function Models (II)
Further Issues Using OLS with Time Series Data
Further Issues in Using OLS with Time Series Data
Chapter 3: Getting the Hang of Statistics
STOCHASTIC HYDROLOGY Random Processes
Chapter 3: Getting the Hang of Statistics
Further Issues Using OLS with Time Series Data
Eni Sumarminingsih, SSi, MM
Lecturer Dr. Veronika Alhanaqtah
Financial Econometrics Fin. 505
CH2 Time series.
Multiple Regression Analysis
Stationary Stochastic Process
Presentation transcript:

Stationary Stochastic Process A stochastic process is stationary if for every collection of time indices 1 ≤ t1 < …< tm the joint distribution of (xt1, …, xtm) is the same as that of (xt1+h, … xtm+h) for h ≥ 1 Thus, stationarity implies that the xt’s are identically distributed and that the nature of any correlation between adjacent terms is the same across all periods Prof. Dr. Rainer Stachuletz

Covariance Stationary Process A stochastic process is covariance stationary if E(xt) is constant, Var(xt) is constant and for any t, h ≥ 1, Cov(xt, xt+h) depends only on h and not on t Thus, this weaker form of stationarity requires only that the mean and variance are constant across time, and the covariance just depends on the distance across time Prof. Dr. Rainer Stachuletz

Weakly Dependent Time Series A stationary time series is weakly dependent if xt and xt+h are “almost independent” as h increases If for a covariance stationary process Corr(xt, xt+h) → 0 as h → ∞, we’ll say this covariance stationary process is weakly dependent Want to still use law of large numbers Prof. Dr. Rainer Stachuletz

Prof. Dr. Rainer Stachuletz An MA(1) Process A moving average process of order one [MA(1)] can be characterized as one where xt = et + a1et-1, t = 1, 2, … with et being an iid sequence with mean 0 and variance s2e This is a stationary, weakly dependent sequence as variables 1 period apart are correlated, but 2 periods apart they are not Prof. Dr. Rainer Stachuletz

Prof. Dr. Rainer Stachuletz An AR(1) Process An autoregressive process of order one [AR(1)] can be characterized as one where yt = ryt-1 + et , t = 1, 2,… with et being an iid sequence with mean 0 and variance se2 For this process to be weakly dependent, it must be the case that |r| < 1 Corr(yt ,yt+h) = Cov(yt ,yt+h)/(sysy) = r1h which becomes small as h increases Prof. Dr. Rainer Stachuletz

Prof. Dr. Rainer Stachuletz Trends Revisited A trending series cannot be stationary, since the mean is changing over time A trending series can be weakly dependent If a series is weakly dependent and is stationary about its trend, we will call it a trend-stationary process As long as a trend is included, all is well Prof. Dr. Rainer Stachuletz

Assumptions for Consistency Linearity and Weak Dependence A weaker zero conditional mean assumption: E(ut|xt) = 0, for each t No Perfect Collinearity Thus, for asymptotic unbiasedness (consistency), we can weaken the exogeneity assumptions somewhat relative to those for unbiasedness Prof. Dr. Rainer Stachuletz

Large-Sample Inference Weaker assumption of homoskedasticity: Var (ut|xt) = s2, for each t Weaker assumption of no serial correlation: E(utus| xt, xs) = 0 for t  s With these assumptions, we have asymptotic normality and the usual standard errors, t statistics, F statistics and LM statistics are valid Prof. Dr. Rainer Stachuletz

Prof. Dr. Rainer Stachuletz Random Walks A random walk is an AR(1) model where r1 = 1, meaning the series is not weakly dependent With a random walk, the expected value of yt is always y0 – it doesn’t depend on t Var(yt) = se2t, so it increases with t We say a random walk is highly persistent since E(yt+h|yt) = yt for all h ≥ 1 Prof. Dr. Rainer Stachuletz

Random Walks (continued) A random walk is a special case of what’s known as a unit root process Note that trending and persistence are different things – a series can be trending but weakly dependent, or a series can be highly persistent without any trend A random walk with drift is an example of a highly persistent series that is trending Prof. Dr. Rainer Stachuletz

Transforming Persistent Series In order to use a highly persistent series and get meaningful estimates and make correct inferences, we want to transform it into a weakly dependent process We refer to a weakly dependent process as being integrated of order zero, [I(0)] A random walk is integrated of order one, [I(1)], meaning a first difference will be I(0) Prof. Dr. Rainer Stachuletz