Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.

Slides:



Advertisements
Similar presentations
Eigen Decomposition and Singular Value Decomposition
Advertisements

Multiple Regression Analysis
The Simple Regression Model
Component Analysis (Review)
Prediction with Regression
4.3 Confidence Intervals -Using our CLM assumptions, we can construct CONFIDENCE INTERVALS or CONFIDENCE INTERVAL ESTIMATES of the form: -Given a significance.
Lecture 9 Today: Ch. 3: Multiple Regression Analysis Example with two independent variables Frisch-Waugh-Lovell theorem.
The General Linear Model. The Simple Linear Model Linear Regression.
Visual Recognition Tutorial
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
The Simple Linear Regression Model: Specification and Estimation
Linear Methods for Regression Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
Chapter 13 Additional Topics in Regression Analysis
Point estimation, interval estimation
Additional Topics in Regression Analysis
Chapter 4 Multiple Regression.
Chapter 11 Multiple Regression.
Lecture 11 Multivariate Regression A Case Study. Other topics: Multicollinearity  Assuming that all the regression assumptions hold how good are our.
Stat 112: Lecture 9 Notes Homework 3: Due next Thursday
1 A MONTE CARLO EXPERIMENT In the previous slideshow, we saw that the error term is responsible for the variations of b 2 around its fixed component 
Christopher Dougherty EC220 - Introduction to econometrics (chapter 3) Slideshow: prediction Original citation: Dougherty, C. (2012) EC220 - Introduction.
1 PREDICTION In the previous sequence, we saw how to predict the price of a good or asset given the composition of its characteristics. In this sequence,
Separate multivariate observations
Objectives of Multiple Regression
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.
Multiple Linear Regression - Matrix Formulation Let x = (x 1, x 2, …, x n )′ be a n  1 column vector and let g(x) be a scalar function of x. Then, by.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Some matrix stuff.
Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.
Shrinkage Estimation of Vector Autoregressive Models Pawin Siriprapanukul 11 January 2010.
Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.
Statistics and Linear Algebra (the real thing). Vector A vector is a rectangular arrangement of number in several rows and one column. A vector is denoted.
Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.
Modern Navigation Thomas Herring
Chap 14-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 14 Additional Topics in Regression Analysis Statistics for Business.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Simple regression model: Y =  1 +  2 X + u 1 We have seen that the regression coefficients b 1 and b 2 are random variables. They provide point estimates.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.
Founded 1348Charles University
Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.
SEM Basics 2 Byrne Chapter 2 Kline pg 7-15, 50-51, ,
Trees Example More than one variable. The residual plot suggests that the linear model is satisfactory. The R squared value seems quite low though,
Chapter 5 Statistical Inference Estimation and Testing Hypotheses.
Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.
Sampling Design and Analysis MTH 494 Lecture-21 Ossam Chohan Assistant Professor CIIT Abbottabad.
The Simple Linear Regression Model: Specification and Estimation  Theory suggests many relationships between variables  These relationships suggest that.
Regression Analysis Part A Basic Linear Regression Analysis and Estimation of Parameters Read Chapters 3, 4 and 5 of Forecasting and Time Series, An Applied.
Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.
Estimating standard error using bootstrap
Charles University Charles University STAKAN III
Charles University Charles University STAKAN III
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
Ch3: Model Building through Regression
Fundamentals of regression analysis
Charles University Charles University STAKAN III
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
OVERVIEW OF LINEAR MODELS
Charles University Charles University STAKAN III
Charles University Charles University STAKAN III
Chapter 4, Regression Diagnostics Detection of Model Violation
Charles University Charles University STAKAN III
Chapter 7: The Normality Assumption and Inference with OLS
Charles University Charles University STAKAN III
Chapter 13 Additional Topics in Regression Analysis
Multicollinearity What does it mean? A high degree of correlation amongst the explanatory variables What are its consequences? It may be difficult to separate.
Introduction to Econometrics, 5th edition
Presentation transcript:

Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences Jan Ámos Víšek Econometrics Tuesday, – Charles University Ninth Lecture

Schedule of today talk We will have the only topic: (Multi)collinearity What is it ? How to recognize (multi)collinearity? What are the consequences of (multi)collinearity ? What remedies can be prescribed ? we have find replays to the following ones: Prior to answering these questions,

What happens if the design matrix is not of full rank ? What happens if the matrix is “nearly” singular ? How to recognize it ? We shall answer the first question, then the third one and, as the last but not least of course, the second one ! We shall see later why!

What happens if the design matrix is not of full rank ? ( Multi )collinearity Assumptions Then, let us write for sake of simplicity with some ‘s being zero. Assertions Then If the design matrix is not yet of full rank, we repeat the steps, we’ve just demonstrated, up to reaching the full-rank-matrix.

What happens if the design matrix is not of full rank ? The answer is simple: NOTHING, we just exclude the “dependent column” !! We didn’t yet answer the question: What is (multi)collinearity! Please be patient, we shall do it at the proper time! Continued

How to recognize it ? Nevertheless, it is better to start with: I.e., if one column of is “nearly” a linear combination of others. What happens if the matrix is “nearly” singular ? Now it seems natural to answer the question: “Assumptions” is real symmetric and regular Assertions spectral decomposition and.  matrices and, both regular : : - eigenvectors of, : - eigenvalues of

( Multi )collinearity Let us recall that is real symmetric and regular. Hence there are matrices and, both regular so that and ( : - eigenvectors of, : - eigenvalues of All ‘s are positive. preliminary considerations ) Regularity of   is positive definite positive definite  How to recognize it ?

Is it really so, or not? is “nearly” singular, some ‘s are “nearly” zero Conclusion : So, we have found: Regularity of  all. Spectral decomposition  Singularity of  some. with

Consider instead of the matrix.. But The eigenvalues can be arbitrarily large. But..... How to recognize it ? E.g. one column of is still “nearly” a linear combination of others. is still “nearly” singular - no change !!! E.g. instead of giving FDI in millions of $, we’ll give it in thousands of $, etc.. Continued

How to recognize it ? Continued But their ratio is stable, i.e. Condition number ( index podmíněnosti ) : eigenvalues of. So, we can define:.

How to recognize it ? Statistical packages usually don’t offer directly. the condition number Factor analysis ( A demonstration in STATISTICA should be given. ) Notice: The matrix is ( up to the multiplication by ) empirical covariance matrix of the data. Factor analysis finds the spectral decomposition Continued

If some column(s) of the matrix is (are) “nearly” a linear combination of other columns, we call it ( Multi )collinearity In some textbooks the case when one column of is just a linear combination of others, is called also (multi)collinearity or perfect (multi)collinearity, e.g. Jan Kmenta. The round parentheses indicate that sometimes we speak about collinearity, sometimes about multicollinearity ( two dimension- al  multidimensional case ?? ). Definition: The words “collinearity” and “multicolinearity” means the same !! (multi)collinearity.

How to recognize it ? ( Multi )collinearity In the Second Lecture  Let us consider the models Instead of the condition number the packages sometimes offer the coefficients of determination of following regression models. the j-th column of the matrix X Continued something else, e.g. “redundancy”. It is usually a table of and their coefficients of determinations. Recalling:

How to recognize it ? ( Multi )collinearity If the coefficient of the j-th model is (very) large, j-th explanatory variable can be very closely approximated by a linear combination of some other explanatory variables  collinearity. What about to use the determinant of the matrix, to diagnose the collinearity? ( There are cases when it fails !!) which it offers is usually called “redundancy”. The branch (or table) of statistical package Continued

How to recognize it ? ( Multi )collinearity We can assume that if matrix is nearly singular, its deter- minant is nearly zero. Considering once again the matrix The determinant of the matrix as an indicator of collinearity definitely failed !! Continued However the “level of collinearity” does not change but the determinant can be made arbitrarily large.

How to recognize it ? ( Multi )collinearity. Putting, we have. Really, ( the j-th column of ) is the empirical covariance matrix of data. Nevertheless, let us recall that the matrix Continued

How to recognize it ? ( Multi )collinearity Making the “trick” with multiplying all elements of the matrix by a, we arrive at. The critical values were derived under the assumption of nor- mality of disturbances and hence it may be “biased”, pretty well. Farrar-Glauber test Continued The determinant of the correlation matrix of data can serve as an indicator of collinearity.

What happens if the matrix is “nearly” singular ? :.. First of all, let us find what is. Then Let us verify that ( Multi )collinearity Continued

What happens if the matrix is “nearly” singular ? ( Multi )collinearity. Continued

What happens if the matrix is “nearly” singular ? ( Multi )collinearity Assuming, the matrices are “approximately of the same magnitude”.. So we have The smaller eigenvalue is, the larger contribution to !!! Continued Assertion

(Multi)collinearity can cause that the variance of Conclusion the estimates of regression coefficients can be pretty large. Remark :  “decrease” or “increase” of by The “increase” or “decrease” of by  “decrease” or “increase” of by for all What is a remedy ? To consider normed data !! Of course, the interpretation of coefficients What is a remedy ? need not be straightforward !!!

What is a remedy for given level of the the condition number ? The question should be: Condition number > 100  (at least) one column of has to be excluded Condition number (10 (30), 100)  a special treatment (see below) is to be applied Condition number < 10 (30)  everything is O.K. – there is nothing to be done

First possible treatment of collinearity Ridge regression (hřebenová regrese) Let be iid. r.v’s,. Lemma Assumptions Assertions Bias of is and the matrix A.E.Hoerl, R.W.Kennard 1970 of the mean quadratic deviations ( MSE ) has the form.

Proof of previous lemma Bias since,. Two preliminary computations – for. Putting and, this is the bias. Secondly, let us find. Firstly

Proof - continued Finally hence, We have

A biased estimator 90% confidence interval, although contain- is rather wide. ing the true value, 90% confidence interval is much shorter and contains the true An unbiased estimator (has a pretty large variance) value, too. UNBIASED OR BIASED? William Shakespeare

Let be iid. r.v’s,, Lemma Assumptions Assertions Then is positive definite matrix. has full rank and. Proof is long and rather technical, hence it will be omitted. Let ‘s and ‘s be eigenvalues and eigenvectors of, Assumptions Assertions Then. respectively. Assertion Proof is only a “computation”.

Let us compare An example If (minimal), then the corresponding contribution to is, while for this con- tribution to is only.

Another possibility of treating collinearity Regression with ( linear ) constraints ( regrese s ( lineárními ) ohraničeními ) An observation Assuming random  for we have. It indicates that a theory, similar to the theory for ridge-regression-estimator, can be derived.

Let be matrix of type. Assertions Then for all and any matrix of type there is and a matrix of type and a one-to-one. Another possibility of treating regression with ( linear ) constraints Lemma mapping such that for any we have Proof of type so that is regular and Assumptions

Another possibility of treating regression with ( linear ) constraints Proof - continued. and for any put Then Let. If is regular As linearly independent rows of create regular matrix (of type ) i.e. is one-to-one and

Another possibility of treating regression with ( linear ) constraints Proof - continued. i.e. is on Finally for any we have for and This is residual for transformed data and unrestricted parameters This is residual for original data but restricted parameters Remember : is “on”

Another possibility of treating regression with ( linear ) constraints “Remarks” at the bottom of previous slide

Are there any realistic example of regression with ( linear ) constraints Combining forecasts of time series Bates, J. M., C. W. J. Granger (1969): The combination of forecasts. Operational Research Quarterly, 20, Granger, C. W. J. (1989): Invited review: Combining forecasts -twenty years later. Journal of Forecasting, 8, Clemen, R. T. (1986): Linear constraints and efficiency of combined forecasts. Journal of Forecasting, 6,

Assumptions Assertions 1) Prior density of for the fixed variance of disturbances is 2) prior density of variance of disturbances is i.e. -distribution with parameters c and d..,, (Of course, and are assumed to be known.) Then the posterior mean value of is Notice that for we obtain nearly the same estimator as on the previous slide. It may be of interest... Bayesian estimate

What is to be learnt from this lecture for exam ? Collinearity – what is it, how to recognize it, consequences. Ridge regression – optimality of bias. Regression with some constraints - random constraints, - deterministic constraints. All what you need is on