EC 827 Module 2 Forecasting a Single Variable from its own History.

Slides:



Advertisements
Similar presentations
FINANCIAL TIME-SERIES ECONOMETRICS SUN LIJIAN Feb 23,2001.
Advertisements

Autocorrelation Functions and ARIMA Modelling
Dates for term tests Friday, February 07 Friday, March 07
Managerial Economics in a Global Economy
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Using SAS for Time Series Data
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
Unit Roots & Forecasting
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 7: Box-Jenkins Models – Part II (Ch. 9) Material.
Time Series Building 1. Model Identification
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
R. Werner Solar Terrestrial Influences Institute - BAS Time Series Analysis by means of inference statistical methods.
Objectives (BPS chapter 24)
10 Further Time Series OLS Issues Chapter 10 covered OLS properties for finite (small) sample time series data -If our Chapter 10 assumptions fail, we.
STAT 497 APPLIED TIME SERIES ANALYSIS
Chapter 10 Simple Regression.
Correlation and Simple Regression Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
Regression Analysis. Unscheduled Maintenance Issue: l 36 flight squadrons l Each experiences unscheduled maintenance actions (UMAs) l UMAs costs $1000.
Modeling Cycles By ARMA
Statistics II: An Overview of Statistics. Outline for Statistics II Lecture: SPSS Syntax – Some examples. Normal Distribution Curve. Sampling Distribution.
Chapter 12 Simple Regression
Data Sources The most sophisticated forecasting model will fail if it is applied to unreliable data Data should be reliable and accurate Data should be.
Class 5: Thurs., Sep. 23 Example of using regression to make predictions and understand the likely errors in the predictions: salaries of teachers and.
The Simple Regression Model
Prediction and model selection
ARIMA Forecasting Lecture 7 and 8 - March 14-16, 2011
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
Economics 20 - Prof. Anderson
Discovering and Describing Relationships
Introduction to Probability and Statistics Linear Regression and Correlation.
Autocorrelation Lecture 18 Lecture 18.
Introduction to Regression Analysis, Chapter 13,
Relationships Among Variables
Ordinary Least Squares
BOX JENKINS METHODOLOGY
ARMA models Gloria González-Rivera University of California, Riverside
Hypothesis Testing in Linear Regression Analysis
Chapter 9 Comparing More than Two Means. Review of Simulation-Based Tests  One proportion:  We created a null distribution by flipping a coin, rolling.
STAT 497 LECTURE NOTES 2.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
Evidence Based Medicine
Ch4 Describing Relationships Between Variables. Pressure.
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 8: Estimation & Diagnostic Checking in Box-Jenkins.
Linear Stationary Processes. ARMA models. This lecture introduces the basic linear models for stationary processes. Considering only stationary processes.
Chapter 6 Lecture 3 Sections: 6.4 – 6.5.
Lecture 7: Forecasting: Putting it ALL together. The full model The model with seasonality, quadratic trend, and ARMA components can be written: Ummmm,
Managerial Economics Demand Estimation & Forecasting.
LECTURE 14 TUESDAY, 13 OCTOBER STA 291 Fall
Byron Gangnes Econ 427 lecture 11 slides Moving Average Models.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Week 21 Stochastic Process - Introduction Stochastic processes are processes that proceed randomly in time. Rather than consider fixed random variables.
Copyright © 2014, 2011 Pearson Education, Inc. 1 Chapter 21 The Simple Regression Model.
Lecture 10: Correlation and Regression Model.
K. Ensor, STAT Spring 2004 Memory characterization of a process How would the ACF behave for a process with no memory? What is a short memory series?
Correlation – Recap Correlation provides an estimate of how well change in ‘ x ’ causes change in ‘ y ’. The relationship has a magnitude (the r value)
Time Series Analysis Lecture 11
Chapter 8: Simple Linear Regression Yang Zhenlin.
Review of Statistics.  Estimation of the Population Mean  Hypothesis Testing  Confidence Intervals  Comparing Means from Different Populations  Scatterplots.
Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Review and Summary Box-Jenkins models Stationary Time series AR(p), MA(q), ARMA(p,q)
Previously Definition of a stationary process (A) Constant mean (B) Constant variance (C) Constant covariance White Noise Process:Example of Stationary.
The Box-Jenkins (ARIMA) Methodology
Module 4 Forecasting Multiple Variables from their own Histories EC 827.
Univariate Time series - 2 Methods of Economic Investigation Lecture 19.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
The Law of Averages. What does the law of average say? We know that, from the definition of probability, in the long run the frequency of some event will.
Linear Regression.
Module 3 Forecasting a Single Variable from its own History, continued
CH2 Time series.
Presentation transcript:

EC 827 Module 2 Forecasting a Single Variable from its own History

Regression Refresher (oh what fun?) u Consider a simple regression, e.g. one based on the incomes of workers at a firm: u …for which a computer will calculate an optimal mathematical model:  where the  ’s are coefficients on education, age etc… and the  is a random error.

Coefficients  Let’s say that   (the coefficient on age) is 125. If we assume that the data for age is just some number of years (e.g. 32 years of age etc…), then the coefficient implies that, on average, a worker who is one year older will earn $125 more than a younger equivalent, all things being equal. u But… –is the relationship real? –does it matter at all?

Statistical Significance u To tell whether the coefficient should be believed, the computer will generally give us some measure of the reliability of the coefficient estimate, normally either: –the standard error –the t-stat –the p-value u Fortunately, they all tell the same story...

Statistical Significance II u …they all measure, given the natural variability of the data, how likely it is that the actual (as opposed to estimated) coefficient is really zero (and hence, there’s no real relationship between, e.g. age and income). u Generally, we use a 95% confidence level as our measure of “certainty”. If a coefficient passes one of three equivalent tests, we are reasonably confident there is a relationship:

Coefficient Tests I u If the coefficient is approximately twice the size of the standard error or u The t-stat is greater (in absolute value) than 1.96 or u The p-value is less (in absolute value) than 0.05 u …then we’re confident we’ve found a real relationship. Note: saying we’re confident there’s a relationship isn’t the same as saying we’re confident that our relationship is accurate.

Coefficient Tests II u As a general rule, if a coefficient passes one of these tests, we think it’s important as an explanation of the dependent variable (in our example, income). If it fails the test, we may want to discard it. u We can also test whole sets of variables by using F-tests. Again, the computer will calculate these scores for you. Their meaning is the same: they measure whether or not it appears the variables are important determinants of whatever we’re modeling.

Economic Significance u Be careful: all these tests just look for a mathematically meaningful relationship, not a practically important one. u For example, it would be possible that the coefficient on age was 1.25, not 125, but was very statistically significant. That would imply that an extra year of age raised average salaries by $1.25 per year. Statistically measurable? Yes. Relevant or important? No.

Time Series Data u Data that we are interested in forecasting are time series: observations that have a well defined temporal ordering u Notation: t-1 X t = forecast constructed at time = t-1 of the observation on X to be realized at time = t. No standard notation u Alternative data have no particular ordering: e.g. sample of heights of people in this class

Forecasting a Single Variable I u Any time series can be considered as the sum of two components: –Deterministic Component: that part of the time series for which a perfect forecast of the future value can be constructed »examples: constant, time trends, constant seasonal factors –Stochastic Component: that part of the time series for which is random (stochastic) for which predictions of future value may turn out to be in error.

Forecasting a Single Variable II u Assumption is that a history of the variable is available - a time series of observations u Require that information in that history have implication for current or future realizations of the variable. u Useful information available at the present for forecasting the future requires correlation between events at different points in time.

Correlation and Causation u Correlation simply implies a link between two data series, not that the link is “cause and effect”. u In a perfect world, we’d like to find the effects of causes u In the real world, the best we can hope for is to find potential causes of effects

Forecasts from Own History: An Example u Consider a coin tossing experiment: –Toss a coin N times and record heads (1) or tails (0) for each replication –Generate a time series: 0,1,1,0,... –No deterministic component to this time series –Does the information in the time series of outcomes provide any basis for forecasting the outcome of additional tosses? Why or why not?

Covariance Stationarity u Characteristic of the Stochastic Component of a time series –mean does not depend on time –variance does not depend on time and is finite –autocovariances (autocorrelations) depend only on the distance between observations and not the time of the observations.

Sample Correlation Coefficients u For two series, X i and Y i on which N observations are available: u correlation coefficient =

Autocorrelations: Definitions u Definition: An autocorrelation coefficient is the correlation between observations of a time series that are separated by a fixed time interval. –A First order autocorrelation is the correlation between observations in a time series and the same observations lagged one period. –A p th order autocorrelation is the correlation between observations in a time series and the same observations lagged p periods.

Autocorrelations: Formulae

Sample Autocorrelation Function I Coin Tossing Experiment Lag

Predictions of Outcomes of Future Coin Tosses u Outcome of any particular coin toss (head or tail) is not influenced by the outcome of any past toss (assuming a fair coin) u Coin tosses are independent events u Autocorrelations for time series of coin toss outcomes are zero (estimated autocorrelations are not significantly different from zero) u No useful forecasting info in time series

Sample Partial Autocorrelation Coefficients u Construct a linear regression of a variable on a constant and lagged observations on the dependent variable up to order p. u Estimated coefficients in this regression model are the estimated partial autocorrelation coefficients, i.e. the coefficient on the n’th lag would be correlation between the event n periods ago and today’s event, given all of the other events.

White Noise Variables u white noise process (stochastic variable) = –zero mean –constant finite variance –not serially correlated (uncorrelated with observations at different points in time) »autocorrelation coefficients of order  1 are all zero. u Coin tossing is such a white noise process (actually stronger = independent white noise)

Wold Representation Theorem u Any zero mean covariance stationary process X t can be written as an infinite sum of white noise processes:

Wold Theorem: Implementation u What does that mean? It means that everything that happens is a function of an infinite series of all past random events. True, but… so what? u The problem is to estimate the terms –impossible since there are an infinite number –trick is to find some model that approximates the Wold representation

Forecasting Without Infinite Information u What will tomorrow look like? Generally, tomorrow will look like today. u What’s more important: –How much will tomorrow look like today? –How will tomorrow respond to today’s shocks? –How will tomorrow be different from today? u First, know what question to ask… then worry about answers.

Autocorrelation Patterns I u Autoregressive (AR) patterns: “Today looks like previous days” –at least three components »deterministic component (e.g. constant, trend, constant seasonal factors) »second component depends on observed values of previous periods »third component is a new shock, independent of anything that has happened in the past

Example: AR(1) process

Sample Autocorrelation Function II AR(1) Variable; a 1 =.8 Lag

Sample Autocorrelation II u Note that in AR(1) process all observations prior to time = t are correlated with the outcome at t. –all previous observations have information that is useful in forecasting what will happen at t. –usefulness (size of autocorrelation coefficients) decreases as information becomes older (move to more distant past) u This is the general result for AR processes, although the autocorrelation patterns can differ.

Autocorrelation Patterns II u Moving Average (MA) Process: “Today is determined by yesterday’s shocks” –at least three components »the deterministic component of the series »the effect of the current shock on the series »the effect of one or more previous shocks whose influence still persists in the current observation.

MA(1) Process: Example

Sample Autocorrelation Function III Ma1 Variable c 1 =.8 Lag The shock occurs at t=1, with full strength, then persists with strength c 1 *e t-1 = (.8*1) =.8 at t=2, then vanishes

Sample Autocorrelations III u Note that in MA process only a limited number of past observations have information (autocorrelations) that are useful in forecasting the current outcome. –number of past observations that have potentially useful forecasting information is determined by the length of the MA process (here only one past observation)

Sample Autocorrelations IV u Size of first order autocorrelation of MA(1) process determined by value of c 1 (see Diebold, p. 158) u (don’t worry about remembering this. That’s what PCs are for)

Forecasts from AR Models I u How does anything that occurs today (time = t) carry forward to influence future observations? –Assume shock e t = 1.0 –AR(1) coefficient in (e.g.) Annual Inflation Model is 0.58 –Shock of 1.0 at time = t carries forward to generate an increase in X t+1 of 0.58 –Shock of 0.58 to X t+1 carries forward to generate and increase in X t+2 of 0.58*0.58 = 0.34

Forecasts from AR Models II u Effect of shock at time = t persists to affect observations at t+3, t+4, etc. u Size of the effect on future observations becomes smaller as long as absolute value of autoregressive coefficient is < 1.0 u Contrast: –Shocks in AR model carry forward to affect future observations indefinitely; –Shocks in MA model carry forward limited number of periods

Forecasts from AR Models III u Effect on future observations at time t+j of a shock at a particular time t, is measured by impulse response function

Impulse Response Function Annual Inflation AR(1) Model AR(1) Impulse Response Function

Forecasts from MA Models u How does anything that occurs today (time = t) carry forward to influence future observations? –Assume that we see a shock e t = 1.0. –What are implications for t+1, t+2, etc?

Transitory & Permanent Shocks u A transitory shock is one whose effect eventually dies off –go far enough out into the future and events at that time are not influenced by what is happening today. u A permanent shock is one whose effect continues to influence events, no matter how far off into the future. –all past events have a lasting effect on the present and the future