Analyzing stochastic time series Tutorial

Slides:



Advertisements
Similar presentations
FINANCIAL TIME-SERIES ECONOMETRICS SUN LIJIAN Feb 23,2001.
Advertisements

Autocorrelation Functions and ARIMA Modelling
Time Series Analysis -- An Introduction -- AMS 586 Week 2: 2/4,6/2014.
DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Exam 1 review: Quizzes 1-6.
Model Building For ARIMA time series
Part II – TIME SERIES ANALYSIS C5 ARIMA (Box-Jenkins) Models
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 7: Box-Jenkins Models – Part II (Ch. 9) Material.
Time Series Building 1. Model Identification
An Introduction to Time Series Ginger Davis VIGRE Computational Finance Seminar Rice University November 26, 2003.
STAT 497 APPLIED TIME SERIES ANALYSIS
How should these data be modelled?. Identification step: Look at the SAC and SPAC Looks like an AR(1)- process. (Spikes are clearly decreasing in SAC.
1 Alberto Montanari University of Bologna Simulation of synthetic series through stochastic processes.
Forecasting Purpose is to forecast, not to explain the historical pattern Models for forecasting may not make sense as a description for ”physical” behaviour.
CMPT 855Module Network Traffic Self-Similarity Carey Williamson Department of Computer Science University of Saskatchewan.
Data Sources The most sophisticated forecasting model will fail if it is applied to unreliable data Data should be reliable and accurate Data should be.
Understanding and Predicting Host Load Peter A. Dinda Carnegie Mellon University
ARIMA Forecasting Lecture 7 and 8 - March 14-16, 2011
Financial Econometrics
Economics 20 - Prof. Anderson
Modern methods The classical approach: MethodProsCons Time series regression Easy to implement Fairly easy to interpret Covariates may be added (normalization)
K. Ensor, STAT Spring 2005 The Basics: Outline What is a time series? What is a financial time series? What is the purpose of our analysis? Classification.
Lecture 16 Random Signals and Noise (III) Fall 2008 NCTU EE Tzu-Hsien Sang.
R. Werner Solar Terrestrial Influences Institute - BAS Time Series Analysis by descriptive statistic.
Seiji Armstrong Huy Luong Huy Luong Alon Arad Alon Arad Kane Hill Kane Hill.
BOX JENKINS METHODOLOGY
Box Jenkins or Arima Forecasting. H:\My Documents\classes\eco346\Lectures\chap ter 7\Autoregressive Models.docH:\My Documents\classes\eco346\Lectures\chap.
Traffic modeling and Prediction ----Linear Models
1 Chapters 9 Self-SimilarTraffic. Chapter 9 – Self-Similar Traffic 2 Introduction- Motivation Validity of the queuing models we have studied depends on.
References for M/G/1 Input Process
AR- MA- och ARMA-.
ARMA models Gloria González-Rivera University of California, Riverside
Time Series Analysis.
STAT 497 LECTURE NOTES 2.
Week 2ELE Adaptive Signal Processing 1 STOCHASTIC PROCESSES AND MODELS.
Linear Stationary Processes. ARMA models. This lecture introduces the basic linear models for stationary processes. Considering only stationary processes.
#1 EC 485: Time Series Analysis in a Nut Shell. #2 Data Preparation: 1)Plot data and examine for stationarity 2)Examine ACF for stationarity 3)If not.
It’s About Time Mark Otto U. S. Fish and Wildlife Service.
Data analyses 2008 Lecture Last Lecture Basic statistics Testing Linear regression parameters Skill.
K. Ensor, STAT Spring 2005 Model selection/diagnostics Akaike’s Information Criterion (AIC) –A measure of fit plus a penalty term for the number.
Introducing ITSM 2000 By: Amir Heshmaty Far. S EVERAL FUNCTION IN ITSM to analyze and display the properties of time series data to compute and display.
Module 3: Introduction to Time Series Methods and Models.
Introduction to Time Series Analysis
Week 21 Stochastic Process - Introduction Stochastic processes are processes that proceed randomly in time. Rather than consider fixed random variables.
ارتباطات داده (883-40) فرآیندهای تصادفی نیمسال دوّم افشین همّت یار دانشکده مهندسی کامپیوتر 1.
K. Ensor, STAT Spring 2004 Memory characterization of a process How would the ACF behave for a process with no memory? What is a short memory series?
Time Series Basics Fin250f: Lecture 8.1 Spring 2010 Reading: Brooks, chapter
3.Analysis of asset price dynamics 3.1Introduction Price – continuous function yet sampled discretely (usually - equal spacing). Stochastic process – has.
Lecture#10 Spectrum Estimation
MULTIVARIATE TIME SERIES & FORECASTING 1. 2 : autocovariance function of the individual time series.
Auto Regressive, Integrated, Moving Average Box-Jenkins models A stationary times series can be modelled on basis of the serial correlations in it. A non-stationary.
Statistics 349.3(02) Analysis of Time Series. Course Information 1.Instructor: W. H. Laverty 235 McLean Hall Tel:
Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Discrete-time Random Signals
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
ECE-7000: Nonlinear Dynamical Systems 2. Linear tools and general considerations 2.1 Stationarity and sampling - In principle, the more a scientific measurement.
Components of Time Series Su, Chapter 2, section II.
Ch16: Time Series 24 Nov 2011 BUSI275 Dr. Sean Ho HW8 due tonight Please download: 22-TheFed.xls 22-TheFed.xls.
Introduction to stochastic processes
Chapter 6 Random Processes
Analysis of Financial Data Spring 2012 Lecture 4: Time Series Models - 1 Priyantha Wijayatunga Department of Statistics, Umeå University
Stochastic Process - Introduction
Covariance, stationarity & some useful operators
Chapter 6: Autoregressive Integrated Moving Average (ARIMA) Models
Stochastic models - time series.
Machine Learning Week 4.
Stochastic models - time series.
STOCHASTIC HYDROLOGY Random Processes
Lecturer Dr. Veronika Alhanaqtah
CH2 Time series.
BOX JENKINS (ARIMA) METHODOLOGY
Presentation transcript:

Analyzing stochastic time series Tutorial Kingston University, Dept. Computing, Information Systems and Mathematics Analyzing stochastic time series Tutorial Malgorzata Kotulska Department of Biomedical Engineering & Instrumentation Wroclaw University of Technology, Poland

Outline Data motivated analysis - time series in the real life Probability and time series - stochastic vs dterministic Stationarity Correlations in time series Modelling linear time series with short-range correlations – ARIMA processes Time series with long correlations – Gaussian and non-Gaussian self-similar processes, fractional ARIMA

Time series – examples P. J. Brockwell, R. A. Davis, Introduction to Time Series and Forecasting, Springer, 1987

Ionic channels in cell membrane M. Kullman, M. Winterhalter, S. Bezrukov, Biophys. J.82 (2003) p.802

Nile river J. Beran, Statistics for long-memory processes, Chapman and Hall, 1994

Objectives of time series analysis Data description Data interpretation Data forecasting Control Modelling / Hypothesis testing Prediction

Time series

Time series – realization of a stochastic process {Xt} is a stochastic time series if each component takes a value according to a certain probability distribution function. A time series model specifies the joint distribution of the sequence of random variables.

White noise - example of a time series model

Gaussian white noise

Stochastic properties of the process STATIONARITY System does not change its properties in time Well-developed analytical methods of signal analysis and stochastic processes

WHEN A STOCHASTIC PROCESS IS STATIONARY? {Xt} is a strictly stationary time series if (X1,...,Xn)=d (X1+h,...,Xn+h), where n1, h – integer, =d means distribution equality Properties: The random variables are identically distributed. An idependent identically distributed (iid) sequence is strictly stationary.

Weak stationarity {Xt} is a weakly stationary time series if EXt =  and Var(Xt)=2 are independent of time t Cov(Xs, Xr) depends on (s-r) only, independent of t. Properties: E(Xt2) is time-invariant.

Quantitative method for stationarity Reverse Arrangement Test Weak stationarity: Testing if E(Xt2) is time-invariant

Quantile line method P{Xt  k(t)}=  PROPERTIES: A quantile of order , 0  1, is such a value k(t) that probability of the series taking value less than k(t) at time t equals . P{Xt  k(t)}=  PROPERTIES: Lines parallel to the time axis  stationarity Lines parallel to each other, not to the time axis  constant variance, a variable mean (or median) Lines not parallel to each other  a variable variance (or scale parameter)

Quantile lines of the raw time series Nonstationarity with a variable mean and variance

Methods for nonstationary time series Trend removal Segmentation of the series Specific analytical methods (e.g. ambiguity function, variograms for autocorrelation function)

Trend estimation Filters, e.g. moving average filter, FIR, IIR filters Polynomial (or other, e.g. log) estimation and removal Filters, e.g. moving average filter, FIR, IIR filters Differencing

Seasonal models Classical decomposition model seasonal component trend Stochastic process random noise Xt = mt + Yt + st

Backshift operator B

Detrended series P. J. Brockwell, R. A. Davis, Introduction to Time Series and Forecasting, Springer, 1987

Quantile lines of the differenced time series

(Sample) autocorrelation function

Range of correlations Independent data (e.g. WN) Short-range correlations Long-range correlations (correlated or anti-correlated structure)

ACF for Gaussian WN

Short-range correlations Markov processes (e.g. ionic channel conformational transition) ARMA (ARIMA) linear processes Nonlinear processes (e.g. GARCH process)

ARMA (ARIMA) models Time series is an ARMA(p,q) process if Xt is stationary and if for every t: Xt  1Xt-1 ...  pXt-p= Zt + 1Zt-1 +...+ pZt-p where Zt represents white noise with mean 0 and variance 2 Left side of the equation represents Autoregresive AR(p) part, and right side Moving Average MA(q) component. The polynomials (1- 1z-...- pzp) cannot have (1+ 1z+...+ pzq) common factors.

Examples The range of MA component estimated by ACF (the lag number within Bartlett’s limits ), the range of AR component by PACF Confidence band is

Exponential decay of ACF MA(1) sample ACF AR(1)

Stationary processes with long memory Qualitative features Relatively long periods with high or low level of observation values In short periods there seems to be cycles and local trends. Looking at long series – no particular cycles or persisting trends Overall the series looks stationary

Stationary processes with long memory Quantitative features The variance of the sample mean decays to zero at a slower rate. Instead of there is The sample autocorrelation function decays to zero in a power-law manner instead of exponentially Similarly, the periodogram (frequency analysis) shows a power-law

Classical processes with long correlations Fractional ARIMA processes (fARIMA) Self similar processes

fARIMA ARMA (p,q): ARIMA (p,d,q): fractional ARIMA (p,d,q):

Self-similar process A process X={X(t)}t  0 is called self-similar if for some H > 0 Self similarity in distribution sense, not for the sample path. Every change in time scale a>0 corresponds to the space scale aH. The bigger H the more dramatic change. Self similarity can be observed also on the quantile lines. (Parallel to t H). H =1 – /2 – self-similarity index, (HR +)

Frequency-domain analysis Periodogram Periodogram - estimated PSD by a Fourier transform of a sample autocorrelation function. Periodogram of a long memory time series depends on frequency according to power-law relationship (straight line on log-log plot). It means that if one doubles the frequency - PSD diminishes by the same fraction regardless of the frequency

Basic features of self similar process APPEARANCE. If an amplitude of a self-similar process is rescaled by r H, X (rt) looks like X (t), statistically indistinguishable. VARIANCE of the signal changes as Var (X(t))  t 2H CORRELATION - correlated or anticorrelated structuring H=0.5 no memory H>0.5 long memory H<0.5 antipersistent long correlations – „short memory” PERIODOGRAM - power-law dependance on frequency Na podst. Eke.

Nile Example J. Beran, Statistics for long-memory processes, Chapman and Hall, 1994

Methods R/S analysis by Hurst DFA – Detrended Fluctuation Analysis Exponent-based (correlogram, periodogram of the residuals) : H =1 – /2 other (for appropriate PDFs, e.g. Orey index)

Hurst exponent – the algorithm A series with N elements is divided into shorter series – n elements each

Hurst exponent

Classical self-similar processes

Fractional Brownian motions (fBm) Gaussian noise A series n (n=1,...,N) of uncorrelated and random variables Each n - Gaussian distribution N(0,) Brownian motion A sum of Gaussian white-noise sequence yn(Bm)= n(Bm)=  n1/2 Definition based on standard devation of Bm. Bm – the classic example of nonstationary time series. No memory. Fractional Brownian motions (fBm) n(fBm)   nH

Fractional Brownian motions (fBm) X={X(t)}t  0 is a nonstationary Gaussian process with mean zero and an autocovariance function: Fractional Gaussian noise (fGn) A stationary process of increments in fBm (differences between values separated by some step)

Gaussian or non-gaussian process

Fractional Lévy Stable Motion In the fractional Lévy stable motion (FLSM) the distribution is Lévy‑stable. =0.5 Stable distribution (solid), the attraction domain of stable distribution: Burr and Pareto distributions (broken & dotted)

Scaling properties of PDF -stable distributions have scaling properties – a sum of independent and identically distributed random variables maintains the same shape of the distribution. Similarly as the Gaussian distribution, also a stable distribution (CLT). Only a few -stable distributions have direct formulas for their probability density function. Usually only the characteristic function is given. The distinctive properties of -stable distributions are their long tails, infinite variance and, in some cases, infinite mean value.

fractional Levy-stable motion fLSM process is a self‑similar non-stationary process which can be represented as Z(u) is a symmetric Lévy -stable motion, and  is the stability index of stable distribution. The increment process of FLSM is stationary and it is called a fractional stable noise (FSN).

Memory of a self-similar process d = H  1/ For a Gaussian process =2 For d > 0 the memory is long – a long-range persistent process. Otherwise (d < 0) – a long-range antipersistent process („short memory”). The time series looks very rough

Summary Time series can be deterministic or stochastic. Visual distinction not always possible. Stochastic time series may tested analytically by statistical methods and an appropriate model attributed if the series is stationary. Otherwise a pre-processing needed. Random data in time series may be correlated. Correlations are called memory. Independent data (e.g. WN) do not have a memory. Each element assumes a value according to an independent probability density function.

Summary (2) In short-range memory the time series is correlated with a few elements back; the autocorrelation function shows an exponential decay. Typical models of time series with short memory are linear ARMA models - linear combination of previous elements and white noise components. The range of AR component estimated by ACF, the range of MA component by PACF. In long-range memory the series is correlated with vary far away elements. The decay of the autocorrelation function is slower than exponential. ACF decays according to power-law. Long-range time series can be typically modelled by self-similar processes (fBm, FLSM) or fARIMA linear models

Recommended text books P. J. Brockwell, R. A. Davis, Introduction to Time Series and Forecasting, Springer, 1987 J. Beran, Statistics for long-memory processes, Chapman and Hall, 1994 G.E.P.Box, G.M. Jenkins, Time series analysis: forecasting and control, Holden Day, 1970