OLS SHORTCOMINGS Preview of coming attractions. QUIZ What are the main OLS assumptions? 1.On average right 2.Linear 3.Predicting variables and error term.

Slides:



Advertisements
Similar presentations
Week 13 November Three Mini-Lectures QMM 510 Fall 2014.
Advertisements

Managerial Economics in a Global Economy
Welcome to Econ 420 Applied Regression Analysis
Structural Equation Modeling
There are at least three generally recognized sources of endogeneity. (1) Model misspecification or Omitted Variables. (2) Measurement Error.
Forecasting Using the Simple Linear Regression Model and Correlation
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Multicollinearity Multicollinearity - violation of the assumption that no independent variable is a perfect linear function of one or more other independent.
Chapter 13 Additional Topics in Regression Analysis
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Prof. Dr. Rainer Stachuletz
Additional Topics in Regression Analysis
Regression Hal Varian 10 April What is regression? History Curve fitting v statistics Correlation and causation Statistical models Gauss-Markov.
The Simple Regression Model
CHAPTER 4 ECONOMETRICS x x x x x Multiple Regression = more than one explanatory variable Independent variables are X 2 and X 3. Y i = B 1 + B 2 X 2i +
Intro to Statistics for the Behavioral Sciences PSYC 1900
Statistical Analysis SC504/HS927 Spring Term 2008 Session 7: Week 23: 7 th March 2008 Complex independent variables and regression diagnostics.
Further Inference in the Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Topic 3: Regression.
Analysis of Individual Variables Descriptive – –Measures of Central Tendency Mean – Average score of distribution (1 st moment) Median – Middle score (50.
Using Statistics in Research Psych 231: Research Methods in Psychology.
Linear Regression Models Powerful modeling technique Tease out relationships between “independent” variables and 1 “dependent” variable Models not perfect…need.
Simple Linear Regression Analysis
Ordinary Least Squares
Lecture 5 Correlation and Regression
Correlation & Regression
Chapter 8 Forecasting with Multiple Regression
Introduction to Linear Regression and Correlation Analysis
Chapter 11 Simple Regression
1 MADE WHAT IF SOME OLS ASSUMPTIONS ARE NOT FULFILED?
Hypothesis Testing in Linear Regression Analysis
Understanding Multivariate Research Berry & Sanders.
What does it mean? The variance of the error term is not constant
2-1 MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4)
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Specification Error I.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Chap 14-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 14 Additional Topics in Regression Analysis Statistics for Business.
Properties of OLS How Reliable is OLS?. Learning Objectives 1.Review of the idea that the OLS estimator is a random variable 2.How do we judge the quality.
1Spring 02 Problems in Regression Analysis Heteroscedasticity Violation of the constancy of the variance of the errors. Cross-sectional data Serial Correlation.
Chapter 16 Data Analysis: Testing for Associations.
Discussion of time series and panel models
3-1 MGMG 522 : Session #3 Hypothesis Testing (Ch. 5)
Assumption checking in “normal” multiple regression with Stata.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
EC 532 Advanced Econometrics Lecture 1 : Heteroscedasticity Prof. Burak Saltoglu.
Chap 6 Further Inference in the Multiple Regression Model
Chapter 6 Introduction to Multiple Regression. 2 Outline 1. Omitted variable bias 2. Causality and regression analysis 3. Multiple regression and OLS.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
1 Regression Review Population Vs. Sample Regression Line Residual and Standard Error of Regression Interpretation of intercept & slope T-test, F-test.
Multiple Regression The equation that describes how the dependent variable y is related to the independent variables: x1, x2, xp and error term e.
MULTICOLLINEARITY: WHAT HAPPENS IF THE REGRESSORS ARE CORRELATED?
Quantitative Methods. Bivariate Regression (OLS) We’ll start with OLS regression. Stands for  Ordinary Least Squares Regression. Relatively basic multivariate.
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
Regression Overview. Definition The simple linear regression model is given by the linear equation where is the y-intercept for the population data, is.
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
REGRESSION DIAGNOSTIC II: HETEROSCEDASTICITY
Econometric methods of analysis and forecasting of financial markets
Fundamentals of regression analysis
Chapter 3: TWO-VARIABLE REGRESSION MODEL: The problem of Estimation
Chapter 6: MULTIPLE REGRESSION ANALYSIS
REGRESSION DIAGNOSTIC I: MULTICOLLINEARITY
BEC 30325: MANAGERIAL ECONOMICS
Chapter 13 Additional Topics in Regression Analysis
Instrumental Variables Estimation and Two Stage Least Squares
BEC 30325: MANAGERIAL ECONOMICS
Lecturer Dr. Veronika Alhanaqtah
Presentation transcript:

OLS SHORTCOMINGS Preview of coming attractions

QUIZ What are the main OLS assumptions? 1.On average right 2.Linear 3.Predicting variables and error term uncorrelated 4.No serial correlation in error term 5.Homoscedasticity + Normality of error term

OLS assumptions consequences We know that: –We cannot know the error term => we look for estimators –We cannot know the coefficients => we look for estimators –Estimators of coefficients are OK. Even if heteroscedasticity –Estimators of coefficients are OK. Even if autocorrelation –BUT we cannot know if they are different from zero even => if H or A then error terms inappropriately estimated

OLS assumption consequences If autocorrelation: –Coefficients correctly estimated –Error terms incorrect –If big sample, we do not have to care (estimators are consistent <= asymptotic properties of OLS) If heteroscedasticity: –Coefficients correctly estimated –Error terms incorrect (estimators are not consisntent <= asymptotic properties of OLS) What can we do? –Fool-proof estimations: GENERALISED LEAST SQUARES

How do we get autocorrelation? What we need in the error term is white noise

How do we get autocorrelation? Positive autocorrelation (rare changes of signs)

How do we get autocorrelation? Negative autocorrelation (frequent changes of signs)

How do we get autocorrelation? Model misspecification can give it to you for free

How do we get heteroscedasticity What we need is error terms independent of SIZE of X.

Omitted variable consequences We estimate model of x1 on y In reality there is not only x1, but also x2 –Estimator of x1 in the first model is BIASED Example –Impact of gender on net wage

Omitted variable consequences Example – continued –Impact of gender on net wage, controlling for education

Outliers What is an outlier? –Atypical observation It fits the model, but event was „strange” –Wrong observation It does not fit the model –Really wrong (unemployment rate in Warsaw) –Something unexpected (a structural event, oil shock) What it does to your model? –Makes your standard error larger/smaller –Makes your estimates sensible/senseless What can you do with them? –Throw out => need to have a good reason!!! –Inquire, why is it so?

Multicollinearity What is multicollinearity –Your „exes” correlated among each other What it does –If perfectly, matrix does not invert => no model –If imperfectly, your estimators are not reliable => why? You never know if it is xi or xj that drives the result Your t statistics are inappropriately estimated (you may reject the null hypothesis too often) What can you do with that? –Nothing really... => change your model

Endogeneity What is endogeneity? –Your x and your ε are correlated IN PRINCIPLE (simultaneity) What it does to your model? –Your estimators are no longer consistent (even if sample veeeery big) Where does it come from? –Omitted variable problem? (omitted and included variables correlated) –Reverse causality

What about selection bias? Heckman Nobel Prize 2003 Say you have three types of answers in a survey –Yes –No –IDK What if you try to explain Yes/Know, but there is something important in IDK? Example from yesterday: –employed and Mincer equation versus –employed and unemployed population

How to model? Testing hypotheses: combined and in a combined way: –These are not equivalent What to do with insignificant variables –General to specific IS NOT the same as taking only important How to chose the right specification –Information criteria: Bayesian, Akaike –Adjusted R2 –YOUR APPROACH!

What is OLS model telling you? Estimated coefficients are nothing but correlations  You know the causality from your theory and not the model!  You cannot test if your relation is really causal Whatever test you pass, it doesn’t have to make sense  You can have a spurious regression  Think what you are doing!  You can have a problem of outliers  Look at your dots with caution! Any model is only meaningful, if economics behind it is  Statistical significance is not everything  Look at the size of your estimators and economic significance  Ask yourself reasonable questions  Research for a model sells well, but gives little satisfaction