Linear Regression Basics III Violating Assumptions Fin250f: Lecture 7.2 Spring 2010 Brooks, chapter 4(skim) 4.1-2, 4.4, 4.5, 4.7, 4.9-13.

Slides:



Advertisements
Similar presentations
Heteroskedasticity Hill et al Chapter 11. Predicting food expenditure Are we likely to be better at predicting food expenditure at: –low incomes; –high.
Advertisements

Kin 304 Regression Linear Regression Least Sum of Squares
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 7. Specification and Data Problems.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
The General Linear Model. The Simple Linear Model Linear Regression.
Simple Linear Regression
A Short Introduction to Curve Fitting and Regression by Brad Morantz
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
Graduate School Gwilym Pryce
OUTLIER, HETEROSKEDASTICITY,AND NORMALITY
Linear Regression Basics II Fin250f: Lecture 7.1 Spring 2010 Brooks, chapter ,3.7, 3.8.
Feb 21, 2006Lecture 6Slide #1 Adjusted R 2, Residuals, and Review Adjusted R 2 Residual Analysis Stata Regression Output revisited –The Overall Model –Analyzing.
Regression Hal Varian 10 April What is regression? History Curve fitting v statistics Correlation and causation Statistical models Gauss-Markov.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
Statistical Analysis SC504/HS927 Spring Term 2008 Session 7: Week 23: 7 th March 2008 Complex independent variables and regression diagnostics.
1 MF-852 Financial Econometrics Lecture 6 Linear Regression I Roy J. Epstein Fall 2003.
Statistics 350 Lecture 10. Today Last Day: Start Chapter 3 Today: Section 3.8 Homework #3: Chapter 2 Problems (page 89-99): 13, 16,55, 56 Due: February.
Topic4 Ordinary Least Squares. Suppose that X is a non-random variable Y is a random variable that is affected by X in a linear fashion and by the random.
Quantitative Business Analysis for Decision Making Simple Linear Regression.
Slide Copyright © 2010 Pearson Education, Inc. Active Learning Lecture Slides For use with Classroom Response Systems Business Statistics First Edition.
Measures of Association Deepak Khazanchi Chapter 18.
Statistics 350 Lecture 17. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
Business Statistics - QBM117 Statistical inference for regression.
Correlation & Regression
Regression and Correlation Methods Judy Zhong Ph.D.
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Specification Error I.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Multiple Regression The Basics. Multiple Regression (MR) Predicting one DV from a set of predictors, the DV should be interval/ratio or at least assumed.
12.1 Heteroskedasticity: Remedies Normality Assumption.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
Model Selection and Validation. Model-Building Process 1. Data collection and preparation 2. Reduction of explanatory or predictor variables (for exploratory.
REGRESSION DIAGNOSTICS Fall 2013 Dec 12/13. WHY REGRESSION DIAGNOSTICS? The validity of a regression model is based on a set of assumptions. Violation.
Ch14: Linear Least Squares 14.1: INTRO: Fitting a pth-order polynomial will require finding (p+1) coefficients from the data. Thus, a straight line (p=1)
I271B QUANTITATIVE METHODS Regression and Diagnostics.
KNN Ch. 3 Diagnostics and Remedial Measures Applied Regression Analysis BUSI 6220.
Statistical Data Analysis 2010/2011 M. de Gunst Lecture 9.
CLASSICAL NORMAL LINEAR REGRESSION MODEL (CNLRM )
Curve Fitting Introduction Least-Squares Regression Linear Regression Polynomial Regression Multiple Linear Regression Today’s class Numerical Methods.
Estimation Kline Chapter 7 (skip , appendices)
Assumptions of Multiple Regression 1. Form of Relationship: –linear vs nonlinear –Main effects vs interaction effects 2. All relevant variables present.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Chapter 17 Simple Linear Regression and Correlation.
Maths Study Centre CB Open 11am – 5pm Semester Weekdays
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
1 Ka-fu Wong University of Hong Kong A Brief Review of Probability, Statistics, and Regression for Forecasting.
1 AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Part II: Theory and Estimation of Regression Models Chapter 5: Simple Regression Theory.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Statistics 350 Review. Today Today: Review Simple Linear Regression Simple linear regression model: Y i =  for i=1,2,…,n Distribution of errors.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
Regression Overview. Definition The simple linear regression model is given by the linear equation where is the y-intercept for the population data, is.
Quantitative Methods Residual Analysis Multiple Linear Regression C.W. Jackson/B. K. Gordor.
Statistical Data Analysis - Lecture /04/03
AP Statistics Chapter 14 Section 1.
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
Statistics in MSmcDESPOT
Simultaneous equation system
Chapter 12: Regression Diagnostics
Nonlinear Regression KNNL – Chapter 13.
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
No notecard for this quiz!!
Statistical Assumptions for SLR
مدلسازي تجربي – تخمين پارامتر
Chapter 4, Regression Diagnostics Detection of Model Violation
Simple Linear Regression
Tutorial 1: Misspecification
Regression III.
Regression and Correlation of Data
Presentation transcript:

Linear Regression Basics III Violating Assumptions Fin250f: Lecture 7.2 Spring 2010 Brooks, chapter 4(skim) 4.1-2, 4.4, 4.5, 4.7,

Outline  Violating assumptions  Parameter stability  Model building

OLS Assumptions  Error variances  Error correlations  Error normality  Functional forms and linearity  Omitting variables  Adding irrelevant variables

Error Variance

Error Variance Which is a bigger error? * * * * * * Y X

Error Correlations  Patterns in residuals  Plot residuals/residual diagnostics  Further modeling necessary If you can forecast u(t+1), need to work harder

Error Normality  Skewness and kurtosis in residuals  Testing Plots Bera-Jarque test  How can this impact results?

Bera-Jarque Test for Normality

Nonnormal Errors: Impact  For some theory: No  In practice can be big problem  Many extreme data points  Forecasting models work hard to fit these extreme outliers  Some solutions: Drop data points Robust forecast objectives (absolute errors)

Functional Forms  Y=a+bX  Actual function is nonlinear  Several types of diagnostics Higher order (squared) terms (RESET) Think about specific nonlinear models  Neural networks  Threshold models  Tricky: More later

Omitting Variables Leave out x(2) If it is correlated with x(1) this is a problem. Beta(1) will be biased and inconsistent. Forecast will not be optimal

Irrelevant Variables  Overfitting/data snooping Model fits to noise  Impacts standard errors for coefficients  Coefficients still consistent and unbiased

Parameter Stability  Known break point Chow test Predictive failure test  Unknown break Quant likelihood ratio test Recursive least squares

Chow Test

Predictive Failure

Unknown Breaks  Search for break  Look for maximum Chow level  Distribution is tricky Monte-carlo/bootstrap

Recursive/rolling estimation  Recursive Estimate (1,T1) move T1 to full sample T See if parameters converge  Rolling Roll bands (t-T,t) through data Watch parameters move through time  We’ll use some of these

Pure Out of Sample Tests  Estimate parameters over (1,T1)  Get errors over (T1+1,T)

Model Construction  General -> specific Less financial theory More statistics Problems: large unwieldy models  Simple -> general More theory at the start Problems: can leave out important stuff