Presentation is loading. Please wait.

Presentation is loading. Please wait.

Founded 1348Charles University 1. FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of.

Similar presentations


Presentation on theme: "Founded 1348Charles University 1. FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of."— Presentation transcript:

1 Founded 1348Charles University 1

2 FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences Jan Ámos Víšek Econometrics Tuesday, 12.30 – 13.50 Charles University Ninth Lecture (summer term) 2

3 Schedule of today talk A motivation for robust studies Huber’s versus Hampel’s approach Prohorov distance - qualitative robustness Influence function - quantitative robustness gross-error sensitivity local shift sensitivity rejection point Breakdown point Recalling linear regression model Scale and regression equivariance The weighted least squares 3

4 Introducing robust estimators continued Schedule of today talk Maximum likelihood(-like) estimators - M-estimators Other types of estimators - L-estimators -R-estimators - minimum distance - minimum volume Advanced requirement on the point estimators 4

5 AN EXAMPLE FROM READING THE MATH Having explained what is the limit, an example was presented: To be sure that the students they were asked to solve the exercise : The answer was as follows: really understand what is in question, 5

6 The Weighted Least Squares The reasons for weighting (down) the residuals of observations. An example – diagonal elements of hat matrix Assuming intercept in model in the first column (and row) and, respectively has From see the next slide for geometry of situation 6

7 The Weighted Least Squares continued The i-th diagonal element of hat matrix 7

8 The Weighted Least Squares continued Moreover, is idempotent, i.e. “mean value” of the diagonal element of is. For the case of random regressors - Chatterjee, S., A. S. Hadi (1988): Sensitivity Analysis in Linear Regression. New York: J. Wiley & Sons, gave an approximation of 95% upper quantile. p larger then approx to quantile Denote this as the 1. approx. to critical values 8

9 The Weighted Least Squares continued D.A. Belsley, E. Kuh, R.E. Welsch (1980): Regression Diagnostics: Identifying Influential Data and Sources of Collinearity. New York: J. Wiley & Sons. Theorem Assumptions Assertions If has i.i.d. rows with p-dimensional normal d.f. ( where ), then. Of course, if rows of the matrix are independent, the rows of can’t be independent. But the correlation is of order, i.e. for large n we can employ the result. Then More precise analysis is in: 9

10 continued The Weighted Least Squares The 2. approx. to critical values The 1. approx. to critical values Number of observations 10

11 The Weighted Least Squares continued The 2. approx. to critical values The 1. approx. to critical values Number of observations If the diagonal term of the hat matrix is larger than or even, we should search whether it is outlier or leverage point. It can be reason to for weighing it down !! 11

12 The Weighted Least Squares continued Odhad metodou vážených nejmenší čtverců Let. Then the weighted least squares are given as follows: Putting, the normal equations are and finally. 12

13 Why the robust methods should be also used? Fisher, R. A. (1922): On the mathematical foundations of theoretical statistics. Philos. Trans. Roy. Soc. London Ser. A 222, pp. 309--368. 13

14 Continued Why the robust methods should be also used? 0.930.800.50 0.830.40 0 ! is asymptotically infinitely larger than 14

15 Standard normal density Student density with 5 degree of freedom Is it easy to distinguish between normal and student density? 15

16 Continued Why the robust methods should be also used? New York: J.Wiley & Sons Huber, P.J.(1981): Robust Statistics. 16

17 0.001.002.05.876.948 1.016 2.035 Continued Why the robust methods should be also used? So, only 5% of contamination makes two times better than. Is 5% of contamination much or few? E.g. Switzerland has 6% of errors in mortality tables, see Hampel et al.. Hampel, F.R., E.M. Ronchetti, P. J. Rousseeuw, W. A. Stahel (1986): Robust Statistics - The Approach Based on Influence Functions. New York: J.Wiley & Sons. 17

18 Conclusion: We have developed efficient monoposts which however work only on special F1 circuits. A proposal: Let us use both. If both work, bless the God. We are on F1 circuit. If not, let us try to learn why. What about to utilize, if necessary, a comfortable sedan. It can “survive” even the usual roads. 18

19 Huber’s approach One of possible frameworks of statistical problems is to consider a parameterized family of distribution functions. Let us consider the same structure of parameter space but instead of each distribution function let us consider a whole neighborhood of d.f.. Huber’s proposal: Finally, let us employ usual statistical technique for solving the problem in question. 19

20 continued - an example Huber’s approach Let us look for an (unbiased, consistent, etc.) esti- mator of location with minimal (asymptotic) variance for family., i.e. consider instead of single d.f. the family. Let us look for an (unbiased, consistent, etc.) estimator of location with minimal (asymptotic) variance for family of families. Finally, solve the same problem as at the beginning of the task. For each let us define 20

21 Hampel’s approach The information in data is the same as information in empirical d.f.. An estimate of a parameter of d.f. can be then considered as a functional. has frequently a (theoretical) counterpart. An example: 21

22 continued Hampel’s approach Expanding the functional at in direction to, we obtain: where is e.g. Fréchet derivative - details below. Message: Hampel’s approach is an infinitesimal one, employing “differential calculus” for functionals. Local properties of can be studied through the properties of. 22

23 Qualitative robustness Let us consider a sequence of “green” d.f. which coincide with the red one, up to the distance from the Y-axis. Does the “green” sequence converge to the red d.f. ? 23

24 Let us consider Kolmogorov-Smirnov distance, i.e. continued Qualitative robustness K-S distance of any “green” d.f. from the red one is equal to the length of yellow segment. The “green” sequence does not converge in K-S metric to the red d.f. ! CONCLUSION: Independently on n, unfortunately. 24

25 continued Qualitative robustness Prokhorov distance Now, the sequence of the green d.f. converges to the red one. We look for a minimal length, we have to move the green d.f. - to the left and up - to be above the red one. In words: CONCLUSION: 25

26 Conclusion : For practical purposes we need something “stronger” than qualitative robustness. DEFINITION E.g., the arithmetic mean is qualitatively robust at normal d.f. !?! In words: Qualitative robustness is the continuity with respect to Prohorov distance. i.i.d. Qualitative robustness 26

27 Quantitative robustness The influence function is defined where the limit exists. Influence function 27

28 continued Quantitative robustness Characteristics derived from influence function Gross-error sensitivity Local shift sensitivity Rejection point 28

29 Breakdown point (The definition is here only to show that the description of breakdown which is below, has good mathematical basis. ) Definition – please, don’t read it in the sense that the estimate tends (in absolute value ) to infinity or to zero. is the smallest (asymptotic) ratio which can destroy the estimate In words obsession (especially in regression – discussion below) 29

30 An introduction - motivation Robust estimators of parameters Let us have a family and data. Of course, we want to estimate. Maximum likelihood estimators : What can cause a problem? 30

31 What can cause a problem? Robust estimators of parameters Consider normal family with unit variance: An example (notice that does not depend on ). So we solve the extremal problem 31

32 A proposal of a new estimator Robust estimators of parameters Maximum likelihood-like estimators : Once again: What caused the problem in the previous example? So what about 32

33 Robust estimators of parameters quadratic part linear part 33

34 The most popular estimators Robust estimators of parameters maximum likelihood-like estimators M-estimators based on order statistics L-estimators based on rank statistics R-estimators 34

35 Robust estimators of parameters The less popular estimators but still well known. Robust estimators of parameters based on minimazing distance between empirical d.f. and theoretical one. Minimal distance estimators based on minimazing volume containing given part of data and applying “classical” (robust) method. Minimal volume estimators 35

36 Robust estimators of parameters The classical estimator, e.g. ML-estimator, has typically a formula to be employed for evaluating it. Algorithms for evaluating robust estimators Extremal problems (by which robust estimators are defined) have not (typically) a solution in the form of closed formula. To find an algorithm how to evaluate an approximation to the precise solution. Firstly To find a trick how to verify that the appro- ximation is tight to the precise solution. Secondly 36

37 High breakdown point obsession (especially in regression – discussion below) Hereafter let us have in mind that we speak implicitly about 37

38 Recalling the model Put ( if intercept ),. and where. Linear regression model 38

39 So we look for a model “reasonably” explaining data. Linear regression model Recalling the model graphically 39

40 This is a leverage point and this is an outlier. Linear regression model Recalling the model graphically 40

41 Formally it means: If for data the estimate is, than for data the estimate is Equivariance in scale If for data the estimate is, than for data the estimate is Equivariance in regression Scale equivariant Affine equivariant We arrive probably easy to an agreement that the estimates of parameters of model should not depend on the system of coordinates. Equivariance of regression estimators 41

42 Unbiasedness Consistency Asymptotic normality Low Gross-error sensitivity Reasonably high efficiency Low local shift sensitivity Finite rejection point Controllable breakdown point Scale- and regression-equivariance Algorithm with acceptable complexity and reliability of evaluation Heuristics, the estimator is based on, is to really work Advanced (modern?) requirement on the point estimator Still not exhaustive 42

43 What is to be learnt from this lecture for exam ? All what you need is on http://samba.fsv.cuni.cz/~visek/ Break down point Weigted least squares M-estimators and minimal distance estimators Main reasons for constructing robust estimators - influence of outliers in estimating mean and variance Influence function and indicators of robustness based on it


Download ppt "Founded 1348Charles University 1. FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of."

Similar presentations


Ads by Google