HETEROSKEDASTICITY Chapter 8.

Slides:



Advertisements
Similar presentations
Heteroskedasticity Hill et al Chapter 11. Predicting food expenditure Are we likely to be better at predicting food expenditure at: –low incomes; –high.
Advertisements

Managerial Economics in a Global Economy
Multiple Regression Analysis
The Simple Regression Model
Lecture 8 (Ch14) Advanced Panel Data Method
Lecture 3 (Ch4) Inferences
Instrumental Variables Estimation and Two Stage Least Square
4.3 Confidence Intervals -Using our CLM assumptions, we can construct CONFIDENCE INTERVALS or CONFIDENCE INTERVAL ESTIMATES of the form: -Given a significance.
8. Heteroskedasticity We have already seen that homoskedasticity exists when the error term’s variance, conditional on all x variables, is constant: Homoskedasticity.
8.4 Weighted Least Squares Estimation Before the existence of heteroskedasticity-robust statistics, one needed to know the form of heteroskedasticity -Het.
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
The Simple Linear Regression Model: Specification and Estimation
Cross section and panel method
Economics Prof. Buckles1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Prof. Dr. Rainer Stachuletz
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Chapter 5 Heteroskedasticity. What is in this Chapter? How do we detect this problem What are the consequences of this problem? What are the solutions?
Econ 140 Lecture 181 Multiple Regression Applications III Lecture 18.
Multiple Regression Analysis
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1 Regression Analysis Regression used to estimate relationship between dependent variable (Y) and one or more independent variables (X). Consider the variable.
Econ 140 Lecture 191 Heteroskedasticity Lecture 19.
Further Inference in the Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Topic 3: Regression.
The Simple Regression Model
Review.
EC Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
Economics Prof. Buckles
Correlation and Regression Analysis
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
Multiple Linear Regression Analysis
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
Hypothesis Tests and Confidence Intervals in Multiple Regressors
Regression and Correlation Methods Judy Zhong Ph.D.
ECON 7710, Heteroskedasticity What is heteroskedasticity? What are the consequences? How is heteroskedasticity identified? How is heteroskedasticity.
Inference for regression - Simple linear regression
Hypothesis Testing in Linear Regression Analysis
Returning to Consumption
CHAPTER 14 MULTIPLE REGRESSION
Chapter 10 Hetero- skedasticity Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
12.1 Heteroskedasticity: Remedies Normality Assumption.
Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill Education.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Heteroskedasticity Adapted from Vera Tabakova’s notes ECON 4551 Econometrics II Memorial University of Newfoundland.
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
Regression Analysis © 2007 Prentice Hall17-1. © 2007 Prentice Hall17-2 Chapter Outline 1) Correlations 2) Bivariate Regression 3) Statistics Associated.
Principles of Econometrics, 4t h EditionPage 1 Chapter 8: Heteroskedasticity Chapter 8 Heteroskedasticity Walter R. Paczkowski Rutgers University.
Correlation & Regression Analysis
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
8-1 MGMG 522 : Session #8 Heteroskedasticity (Ch. 10)
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
4-1 MGMG 522 : Session #4 Choosing the Independent Variables and a Functional Form (Ch. 6 & 7)
High Speed Heteroskedasticity Review. 2 Review: Heteroskedasticity Heteroskedasticity leads to two problems: –OLS computes standard errors on slopes incorrectly.
Heteroscedasticity Chapter 8
The Simple Linear Regression Model: Specification and Estimation
HETEROSCEDASTICITY: WHAT HAPPENS IF THE ERROR VARIANCE IS NONCONSTANT?
Our theory states Y=f(X) Regression is used to test theory.
HETEROSCEDASTICITY: WHAT HAPPENS IF THE ERROR VARIANCE IS NONCONSTANT?
Simple Linear Regression
Heteroskedasticity.
Chapter 7: The Normality Assumption and Inference with OLS
Product moment correlation
Financial Econometrics Fin. 505
Presentation transcript:

HETEROSKEDASTICITY Chapter 8

I. Introduction Previously have assumed homoskedasticity. Variance of unobservable error, u, conditional on the explanatory variables is constant. Var(u|x1, x2,…, xk) = Var(u)= s2 Heteroskedasticity occurs if variance of u changes across different segments of the population (i.e. different values of x). Var(u|x1, x2,…, xk) ≠ Var(u) not equal to≠ s2 Example: Return to education, where variance in ability differs by educational attainment

I. Introduction Failure of homoskedasticity Does not bias coefficient estimates Interpretation of R sq and adj R sq unaffected (because unconditional variance) Does lead to bias in the estimate of variance (and consequently, standard error) standard t-statistics, p-values, CI, F-statistics will lead to wrong inference i.e. t-statistics no longer follows t distribution OLS is no longer BLUE (i.e. no longer most efficient estimator)

. . . Example of Heteroskedasticity f(y|x) y x1 x2 x3 x E(y|x) = b0 + b1x . x1 x2 x3 x Economics 20 - Prof. Anderson

II. Heteroskedasticity-Robust Infererence We can adjust standard errors so that they are valid in the presence of heteroskedasticity of unknown form. Convenient, because can get correct standard errors regardless of kind of heteroskedasticity. Outline: Variance with heteroskedasticity Estimate heteroskedasticity-robust standard errors Detect heteroskedasticity

II. Variance with Heteroskedasticity

II. Variance with Heteroskedasticity

III. Estimation with Heteroskedasticity Now have a consistent estimate of the variance under heteroskedasticity the square root can be used as a standard error for inference Call them heteroskedasticity robust s.e. or Huber/White/Eicker s.e. Easily computable Sometimes the variance is corrected for degrees of freedom by multiplying by n/(n – k – 1) before taking the square root. As n → ∞ it’s all the same, though

III. Estimation with Heteroskedasticity Why don’t we always calculate robust s.e.? These robust standard errors only have asymptotic justification as n → ∞ With small sample sizes, the t statistics formed with robust standard errors will not have a distribution close to the t, and inferences will not be correct In Stata, robust standard errors are easily obtained using the robust option of reg y x1 x2, robust

II. Heteroskedasticity-Robust Infererence Estimation Large samples, people often report just robust s.e., in small samples will report both. T-stat tstat=estimate-hypothesized value /(s.e.) F-stat No easy analytical formula—can’t use standard formula. Use STATA to give you F-stat for robust s.e.

III. Estimation with Heteroskedasticity Log Wage Equation Robust s.e. can either be larger or smaller than usual s.e. Here (by chance) estimates are significant regardless of using wrong or right s.e…..not always case.

IV. Testing for Heteroskedasticity Homoskedasticity implies Var(u|x1, x2,…, xk) = s2 To test for heteroskedasticity, just test the following: H0: Var(u|x1, x2,…, xk) = s2, which is equivalent to H0: E(u2|x1, x2,…, xk) = E(u2) = s2 since assume E(u)=0 If assume the relationship between u2 and xj is linear, can test as a linear restriction For u2 = d0 + d1x1 +…+ dk xk + v this means testing H0: d1 = d2 = … = dk = 0 using an F-test.

IV. Testing for Heteroskedasticity The Breusch-Pagan Test Don’t observe the error u (or u2 ), but can estimate it with the residuals from the OLS regression û2 = d0 + d1x1 +…+ dk xk + e Now can use the R û2 2 from this regression to construct an F statistic. Can use normal F-statistic formula if we assume (which we do) that the errors here satisfy homoskedasticity Var(e|x1, x2,…, xk) = su2 The F statistic is just the reported F statistic for overall significance of the regression F = [R û2 2 /k]/[(1 – R û2 2 )/(n – k – 1)], which is distributed Fk, n – k - 1

IV. Testing for Heteroskedasticity The Breusch-Pagan Test Ex: Heteroskedasticity in Housing Price Equations Economics 20 - Prof. Anderson

IV. Testing for Heteroskedasticity The White Test The Breusch-Pagan test will detect any linear forms of heteroskedasticity The White test allows for nonlinearities by using squares and crossproducts of all the x’s û 2 = d0 + d1x1 + d2 x2 + d3x21+ d4x22 +d5x1 x2 + e Use R û2 2 to form F-statistic again, assume latter equation satisfies homoskedasticity test whether all the xj, xj2, and xjxh are jointly significant

IV. Testing for Heteroskedasticity Alternative form of The White Test White Test can get unwieldy pretty quickly Alternative: Consider that the fitted values from OLS, ŷ, are a function of all the x’s Thus, ŷ2 will be a function of the squares and cross-products and so ŷ and ŷ2 can proxy for all of the xj, xj2, and xjxh. Regress û2 = a0 + a1 ŷ + a2 ŷ2 + t Use the R û2 2 to form an F statistic Note only testing for 2 restrictions now (H0: a1, a2 =0)

IV. Testing for Heteroskedasticity Notes Recall, transforming dependent variable into log can reduce heteroskedasticity.

V. Generalized Least Squares & Weighted Least Squares While it’s always possible to estimate robust standard errors for OLS estimates, OLS is not the most efficient estimator. If we can correctly specify the form of the heteroskedasticity, we can obtain more efficient estimates. Basic idea: Transform the model into one that has homoskedastic errors – called weighted least squares Economics 20 - Prof. Anderson

V. Weighted Least Squares Heteroskedasticity is known up to a Multiplicative Constant Suppose the heteroskedasticity can be modeled as: Var(u|x) =si2 =s2h(x) Means variance of error is proportional to level of x I.e. As income increases, variability in savings increases.

V. Weighted Least Squares Heteroskedasticity is known up to a Multiplicative Constant Suppose Var(u|x) = s2h(x)= s2 hi Since hi is just a function of x: E(ui/√hi|x) = 0, because E(ui|x) = 0 Moreover, Var(u|x)=E(u2|x) –E(u|x)E(u|x)= E(u2|x) Then Var(ui/√hi|x) = E((ui/√hi ) 2|x )=1/ hi E( ui 2|x ) = 1/ hi Var(ui|x)= 1/ hi s2 hi =s2 So, if we divided our whole equation by √hi we would have a model where the error is homoskedastic

V. Weighted Least Squares Heteroskedasticity is known up to a Multiplicative Constant Example: Transformed equation satisfies all G-M assumptions. Economics 20 - Prof. Anderson

V. Weighted Least Squares General Least Squares Estimator Estimating the transformed equation by OLS is called generalized least squares (GLS)…class of estimators GLS will be BLUE in this case Provides more efficient estimates than if used OLS in untransformed analysis. Can uses s.e. for t-statistics, p-values, CI, and resulting R2 is used for F-statistics However, R2 not very good goodness of fit measure (tells us how much of variation in transformed x explains variation in transformed y)

V. Weighted Least Squares GLS estimator for correcting heteroskedasticity is called WLS estimator. minimize the weighted sum of squared residuals (weighted by 1/hi ), which is inverse of the variance. Less weight is given to observations with a higher error variance; in contrast, OLS gives same weight to all observations because it is best when error variance is identical for all partitions of the population Economics 20 - Prof. Anderson

V. Weighted Least Squares Minimization problem: ∑uhati2 /hi Can easily perform WLS using “weight” option in STATA. Produces s.e. that can use for t-statistics, etc… Some regression packages even include option to calculate robust s.e. after weighting, in case specify form of heteroskedasticity incorrectly. Ex: Explain financial wealth in terms of income, age, gender, 401(k) eligibility Suspect heteroskedasticity, so use WLS, with weight 1/inc: Var(finan|inc)= s2inc Economics 20 - Prof. Anderson

Economics 20 - Prof. Anderson

V. Weighted Least Squares WLS is great if we know what Var(ui|xi) looks like, but in most cases won’t know form of heteroskedasticity One example where we do know form is: data is aggregated across some group or geographic region instead of at the individual level relationship between amount worker contributes to 401(k) and plan generosity

V. Weighted Least Squares What if only have averages for a firm? mcontribi = b0 + b1mearnsi + b2magei + b3mmaratei + mui If individual regression satisfies all G-M assumptions, it can be shown that the aggregated regression has Var(mui |x)=s2/si where si is the number of employees at firm i. Variance of error term decreases with firm size. Weight is then: 1/hi=mi similar issue if use per capita data at city, county, state or country level…error in aggregate equation has variance proportional to 1/size of population in that area.

VI. Heteroskedasticity must be estimated: Feasible GLS More typically, we don’t know the form of the heteroskedasticity In this case, you need to estimate h(xi) Since are estimating h(xi) and using the estimate to transform the equation, call it feasible GLS. Typically, we start with the assumption of a fairly flexible model, such as Var(u|x) = s2exp(d0 + d1x1 + …+ dkxk) , where h(xi)= exp(d0 + d1x1 + …+ dkxk) Since we don’t know the d, must estimate

VI. Heteroskedasticity must be estimated: Feasible GLS Can transform above as u2 = s2exp(d0 + d1x1 + …+ dkxk)v v is error term assume E(v|x) = 1 and E(v) = 1 Taking natural logs of both sides: ln(u2) = a0 + d1x1 + …+ dkxk + e… a0 now contains original intercept and log(s2) Assume E(e) = 1 and e is independent of x We don’t have u, so replace with its estimate, û. Now can estimate this by OLS to get estimate of h(xi)

VI. Heteroskedasticity must be estimated: Feasible GLS Now, an estimate of h is obtained as ĥ = exp(ĝ). Now can use WLS with weights 1/ ĥ Implementation: Run the original OLS model, save the residuals, û, square them and take the log Regress ln(û2) on all of the independent variables and get the fitted values, ĝ Do WLS using 1/exp(ĝ) as the weight Economics 20 - Prof. Anderson

VI. Heteroskedasticity must be estimated: Feasible GLS FGLS is not BLUE, because use ĥ instead of h, but it is more efficient than just running OLS. As we saw with White test, could save valuable degrees of freedom by regressing: ln(û2) on ŷ and ŷ2 instead ln(û2) of all of the independent variables When doing F tests with WLS, must make sure to use same weights in restricted and unrestricted model form the weights from the unrestricted model and use those weights to do WLS on the restricted model as well as the unrestricted model

VI. Heteroskedasticity must be estimated: Feasible GLS Example: OLS (w/o het rob s.e.) and WLS Under B-P test, Ru2 =.040 and get F-stat that reject null…have evidence of heteroskedasticity Use Feasible GLS procedure and get second set of estimates. Signs and story are similar, but magnitudes different Economics 20 - Prof. Anderson