Covariance, stationarity & some useful operators

Slides:



Advertisements
Similar presentations
Autocorrelation Functions and ARIMA Modelling
Advertisements

Dates for term tests Friday, February 07 Friday, March 07
Part II – TIME SERIES ANALYSIS C4 Autocorrelation Analysis © Angel A. Juan & Carles Serrat - UPC 2007/2008.
Part II – TIME SERIES ANALYSIS C5 ARIMA (Box-Jenkins) Models
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 7: Box-Jenkins Models – Part II (Ch. 9) Material.
Time Series Building 1. Model Identification
R. Werner Solar Terrestrial Influences Institute - BAS Time Series Analysis by means of inference statistical methods.
STAT 497 APPLIED TIME SERIES ANALYSIS
Stationary process NONSTATIONARY PROCESSES 1 In the last sequence, the process shown at the top was shown to be stationary. The expected value and variance.
1 Power 2 Econ 240C. 2 Lab 1 Retrospective Exercise: –GDP_CAN = a +b*GDP_CAN(-1) + e –GDP_FRA = a +b*GDP_FRA(-1) + e.
ARIMA Forecasting Lecture 7 and 8 - March 14-16, 2011
Modern methods The classical approach: MethodProsCons Time series regression Easy to implement Fairly easy to interpret Covariates may be added (normalization)
Non-Seasonal Box-Jenkins Models
K. Ensor, STAT Spring 2005 The Basics: Outline What is a time series? What is a financial time series? What is the purpose of our analysis? Classification.
Business Statistics - QBM117 Statistical inference for regression.
Review of Probability.
BOX JENKINS METHODOLOGY
ARMA models Gloria González-Rivera University of California, Riverside
Time Series Analysis.
STAT 497 LECTURE NOTES 2.
Linear Stationary Processes. ARMA models. This lecture introduces the basic linear models for stationary processes. Considering only stationary processes.
It’s About Time Mark Otto U. S. Fish and Wildlife Service.
1 Forecasting Formulas Symbols n Total number of periods, or number of data points. A Actual demand for the period (  Y). F Forecast demand for the period.
Data analyses 2008 Lecture Last Lecture Basic statistics Testing Linear regression parameters Skill.
Introducing ITSM 2000 By: Amir Heshmaty Far. S EVERAL FUNCTION IN ITSM to analyze and display the properties of time series data to compute and display.
Big Data at Home Depot KSU – Big Data Survey Course Steve Einbender Advanced Analytics Architect.
Week 21 Stochastic Process - Introduction Stochastic processes are processes that proceed randomly in time. Rather than consider fixed random variables.
K. Ensor, STAT Spring 2004 Memory characterization of a process How would the ACF behave for a process with no memory? What is a short memory series?
Time Series Basics Fin250f: Lecture 8.1 Spring 2010 Reading: Brooks, chapter
3.Analysis of asset price dynamics 3.1Introduction Price – continuous function yet sampled discretely (usually - equal spacing). Stochastic process – has.
MULTIVARIATE TIME SERIES & FORECASTING 1. 2 : autocovariance function of the individual time series.
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
1 Chapter 5 : Volatility Models Similar to linear regression analysis, many time series exhibit a non-constant variance (heteroscedasticity). In a regression.
Geology 6600/7600 Signal Analysis 09 Sep 2015 © A.R. Lowry 2015 Last time: Signal Analysis is a set of tools used to extract information from sequences.
Previously Definition of a stationary process (A) Constant mean (B) Constant variance (C) Constant covariance White Noise Process:Example of Stationary.
Random Processes Gaussian and Gauss-Markov processes Power spectrum of random processes and white processes.
Geology 5600/6600 Signal Analysis 14 Sep 2015 © A.R. Lowry 2015 Last time: A stationary process has statistical properties that are time-invariant; a wide-sense.
NONSTATIONARY PROCESSES 1 In the last sequence, the process shown at the top was shown to be stationary. The expected value and variance of X t were shown.
Computational Finance II: Time Series K.Ensor. What is a time series? Anything observed sequentially (by time?) Returns, volatility, interest rates, exchange.
Components of Time Series Su, Chapter 2, section II.
Ch16: Time Series 24 Nov 2011 BUSI275 Dr. Sean Ho HW8 due tonight Please download: 22-TheFed.xls 22-TheFed.xls.
Introduction to stochastic processes
(a) Measuring the memory in a time series with auto-correlation function (ACF)
Analysis of Financial Data Spring 2012 Lecture 4: Time Series Models - 1 Priyantha Wijayatunga Department of Statistics, Umeå University
Energy Consumption Forecast Using JMP® Pro 11 Time Series Analysis
An introduction to Dynamic Factor Analysis
Stochastic Process - Introduction
Multiple Random Variables and Joint Distributions
Eric Ward Mark Scheuerell Eli Holmes
Time Series Analysis.
Time Series Analysis and Its Applications
Chapter 6: Autoregressive Integrated Moving Average (ARIMA) Models
Computational Data Analysis
Statistics 153 Review - Sept 30, 2008
Econ 240 C Lecture 4.
Model Building For ARIMA time series
Sampling Distribution
Sampling Distribution
Machine Learning Week 4.
STOCHASTIC HYDROLOGY Random Processes
The Spectral Representation of Stationary Time Series
Time Series Analysis and Forecasting
Eni Sumarminingsih, SSi, MM
Lecturer Dr. Veronika Alhanaqtah
Time Series introduction in R - Iñaki Puigdollers
CH2 Time series.
BOX JENKINS (ARIMA) METHODOLOGY
Stationary Stochastic Process
Chap 7: Seasonal ARIMA Models
Presentation transcript:

Covariance, stationarity & some useful operators Mark Scheuerell FISH 507 – Applied Time Series Analysis 5 January 2017

Example of a time series MAR(1) Workshop - ESA 2007, San Jose, CA 5 Aug 2007

Topics for today Expectation, mean & variance Covariance & correlation Stationarity Autocovariance & autocorrelation Correlograms White noise Random walks Backshift & difference operators

Expectation, mean & variance MAR(1) Workshop - ESA 2007, San Jose, CA Expectation, mean & variance 5 Aug 2007 The expectation (E) of a variable is its mean value in the population E(x) ≡ mean of x = m E([x - m]2) ≡ mean of squared deviations about m ≡ variance = s2 Can estimate s2 from sample as

MAR(1) Workshop - ESA 2007, San Jose, CA Covariance 5 Aug 2007 If we have 2 variables (x, y) we can generalize variance to covariance Can estimate g from sample as

Graphical example of covariance MAR(1) Workshop - ESA 2007, San Jose, CA 5 Aug 2007

MAR(1) Workshop - ESA 2007, San Jose, CA Correlation 5 Aug 2007 Correlation is a dimensionless measure of the linear association between 2 variables x & y It is simply the covariance standardized by the standard deviations Can estimate g from sample as

The ensemble & stationarity Consider again the mean function for a time series: m(t) = E(xt) The expectation is taken across an ensemble (population) of all possible time series With only 1 sample, however, we must estimate the mean at each time point by the observation If E(xt) is constant across time, we say the time series is stationary in the mean

Stationarity of time series Stationarity is a convenient assumption that allows us to describe the statistical properties of a time series. In general, a time series is said to be stationary if there is no systematic change in the mean or variance, no systematic trend, and no periodic variations or seasonality

Which of these are stationary? MAR(1) Workshop - ESA 2007, San Jose, CA 5 Aug 2007

Autocovariance function (ACVF) For stationary ts, we can define the autocovariance function (ACVF) as a function of the time lag (k) Very “smooth” series have large ACVF for large k; “choppy” series have ACVF near 0 for small k Can estimate gk from sample as

Autocorrelation function (ACF) The autocorrelation function (ACF) is simply the ACVF normalized by the variance ACF measures the correlation of a time series against a time-shifted version of itself (& hence the term “auto”) Can estimate gk from sample as

Properties of the ACF The ACF has several important properties, including -1 ≤ rk ≤ 1, rk = r-k (ie, it’s an “even function”), rk of periodic function is itself periodic rk for sum of 2 indep vars is sum of rk for each

The correlogram The common graphical output for the ACF is called the correlogram, and it has the following features: x-axis indicates lag (0 to k); y-axis is autocorrelation rk (-1 to 1); lag-0 correlation (r0) is always 1 (it’s a ref point); If rk = 0, then sampling distribution of rk is approx. normal, with var = 1/n; Thus, a 95% conf interval is given by

The correlogram

Correlogram for deterministic trend

Correlogram for sine wave

Correlogram for trend + season

Correlogram for random sequence

Correlogram for real data

Partial autocorrelation function MAR(1) Workshop - ESA 2007, San Jose, CA Partial autocorrelation function 5 Aug 2007 The partial autocorrelation function (PACF) measures the linear correlation of a series xt and xt+k with the linear dependence of {xt-1,xt-2,…,xt-(k-1)} removed It is defined as

Revisiting the temperature ts MAR(1) Workshop - ESA 2007, San Jose, CA Revisiting the temperature ts 5 Aug 2007 Data from http://www.ncdc.noaa.gov/

MAR(1) Workshop - ESA 2007, San Jose, CA ACF of temperature ts MAR(1) Workshop - ESA 2007, San Jose, CA 5 Aug 2007

MAR(1) Workshop - ESA 2007, San Jose, CA PACF of temperature ts MAR(1) Workshop - ESA 2007, San Jose, CA 5 Aug 2007

Cross-covariance function (CCVF) Often we are interested in looking for relationships between 2 different time series We can extend the idea of autocovariance to examine the covariance between 2 different ts Define the cross-covariance function (CCVF) for x & y

Cross-correlation function (CCF) The cross-correlation function (CCF) is the CCVF normalized by standard deviations of x & y

CCF for sunspots and lynx

Iterative approach to model building Postulate general class of models As we will see later, ACF & PACF will be very useful here Identify candidate model Estimate parameters Diagnostics: is model adequate? Use model for forecasting or control No Yes

White noise (WN) A time series {wt : t = 1,2,3,…,n} is discrete white noise if the variables w1, w2, w3, …, wn are independent, and identically distributed with a mean of zero Note: At this point we are making no assumptions about the distributional form of {wt}! For example, wt might be distributed as DiscreteUniform({-2,-1,0,1,2}) Normal(0,1)

White noise (WN) A time series {wt : t = 1,2,3,…,n} is discrete white noise if the variables w1, w2, w3, …, wn are independent, and identically distributed with a mean of zero Gaussian WN has the following 2nd-order properties:

White noise

Random walk (RW) A time series {xt : t = 1,2,3,…,n} is a random walk if xt = xt-1 + wt, and wt is white noise RW has the following 2nd-order properties: Note: Random walks are NOT stationary!

Random walk (RW)

The backward shift operator (B) MAR(1) Workshop - ESA 2007, San Jose, CA The backward shift operator (B) 5 Aug 2007 Define the backward shift operator by Or, more generally as So, RW model can be expressed as

The difference operator (∇) MAR(1) Workshop - ESA 2007, San Jose, CA The difference operator (∇) 5 Aug 2007 Define the first difference operator as So, first differencing a RW model yields WN

The difference operator (∇) MAR(1) Workshop - ESA 2007, San Jose, CA The difference operator (∇) 5 Aug 2007 Differences of order d are then defined by For example, twice differencing a ts

Difference to remove trend/season MAR(1) Workshop - ESA 2007, San Jose, CA Difference to remove trend/season 5 Aug 2007 Differencing is a very simple means for removing a trend or seasonal effect The 1st-difference removes a linear trend, a 2nd- difference would remove a quadratic trend, etc. For seasonal data, using a 1st-difference with lag = period removes both trend & seasonal effects Pro: no parameters to estimate Con: no estimate of stationary process

First-difference to remove trend MAR(1) Workshop - ESA 2007, San Jose, CA 5 Aug 2007

First-difference* to remove season MAR(1) Workshop - ESA 2007, San Jose, CA First-difference* to remove season 5 Aug 2007 *At lag = 12

Topics for today Expectation, mean & variance Covariance & correlation Stationarity Autocovariance & autocorrelation Correlograms White noise Random walks Backshift & difference operators