Fig.1 shows the behaving of the derivatives of G and γ 1 with respect to z r in the case of the standard normal distribution. The comparison of the two.

Slides:



Advertisements
Similar presentations
Chapter 3 Properties of Random Variables
Advertisements

+ Validation of Simulation Model. Important but neglected The question is How accurately does a simulation model (or, for that matter, any kind of model)
9.1 confidence interval for the population mean when the population standard deviation is known
Chap 8-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 8 Estimation: Single Population Statistics for Business and Economics.
USING DECISION SUPPORT SYSTEM TECHNIQUE FOR HYDROLOGICAL RISK ASSESSMENT CASE OF OUED MEKERRA IN THE WESTERN OF ALGERIA M. A. Yahiaoui Université de Bechar.
Psych 5500/6500 The Sampling Distribution of the Mean Fall, 2008.
Introduction to Summary Statistics
Central Limit Theorem.
Chapter 10 Simple Regression.
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Section 7-3 Estimating a Population Mean:  Known Created by.
Evaluating Hypotheses
WFM 5201: Data Management and Statistical Analysis
Chapter 8 Estimation: Single Population
FURTHER APPLICATIONS OF INTEGRATION 9. In chapter 6, we looked at some applications of integrals:  Areas  Volumes  Work  Average values.
Chap 6-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 6 Continuous Random Variables and Probability Distributions Statistics.
Standard Scores & Correlation. Review A frequency curve either normal or otherwise is simply a line graph of all frequency of scores earned in a data.
The Normal Distribution Unimodal Symmetrical Abscissa – the different values of x Ordinate – density, probability, or frequency of each value of x Thus,
1 Frequency Distributions. 2 Density Function We’ve discussed frequency distributions. Now we discuss a variation, which is called a density function.
Review of normal distribution. Exercise Solution.
Chapter 6 The Integral Sections 6.1, 6.2, and 6.3
Statistics. Question Tell whether the following statement is true or false: Nominal measurement is the ranking of objects based on their relative standing.
Copyright © Cengage Learning. All rights reserved. 10 Inferences Involving Two Populations.
Percentile Approximation Via Orthogonal Polynomials Hyung-Tae Ha Supervisor : Prof. Serge B. Provost.
Chapter 6 Probability. Introduction We usually start a study asking questions about the population. But we conduct the research using a sample. The role.
Combined Uncertainty P M V Subbarao Professor Mechanical Engineering Department A Model for Propagation of Uncertainty ….
Introduction to Statistical Quality Control, 4th Edition
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Two Functions of Two Random.
QBM117 Business Statistics Probability and Probability Distributions Continuous Probability Distributions 1.
Measures of Variability OBJECTIVES To understand the different measures of variability To determine the range, variance, quartile deviation, mean deviation.
Warsaw Summer School 2014, OSU Study Abroad Program Variability Standardized Distribution.
Introduction to Descriptive Statistics Objectives: 1.Explain the general role of statistics in assessment & evaluation 2.Explain three methods for describing.
Copyright 2011 by W. H. Freeman and Company. All rights reserved.1 Introductory Statistics: A Problem-Solving Approach by Stephen Kokoska Chapter 8: Confidence.
Integrals  In Chapter 2, we used the tangent and velocity problems to introduce the derivative—the central idea in differential calculus.  In much the.
Variables and Random Variables àA variable is a quantity (such as height, income, the inflation rate, GDP, etc.) that takes on different values across.
Measures of central tendency are statistics that express the most typical or average scores in a distribution These measures are: The Mode The Median.
Interval Estimation and Hypothesis Testing Prepared by Vera Tabakova, East Carolina University.
Virtual University of Pakistan
© Copyright McGraw-Hill 2000
Numerical Measures of Variability
Integration by Substitution Antidifferentiation of a Composite Function.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Section 7-5 Estimating a Population Variance.
Sampling distributions rule of thumb…. Some important points about sample distributions… If we obtain a sample that meets the rules of thumb, then…
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Mean, Variance, Moments and.
Chapter SixteenChapter Sixteen. Figure 16.1 Relationship of Frequency Distribution, Hypothesis Testing and Cross-Tabulation to the Previous Chapters and.
CY1B2 Statistics1 (ii) Poisson distribution The Poisson distribution resembles the binomial distribution if the probability of an accident is very small.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
Answering Descriptive Questions in Multivariate Research When we are studying more than one variable, we are typically asking one (or more) of the following.
Estimating a Population Variance
Chapter 4 Exploring Chemical Analysis, Harris
CHAPTER- 3.2 ERROR ANALYSIS. 3.3 SPECIFIC ERROR FORMULAS  The expressions of Equations (3.13) and (3.14) were derived for the general relationship of.
Chapter 6: Descriptive Statistics. Learning Objectives Describe statistical measures used in descriptive statistics Compute measures of central tendency.
CHAPTER 3 – Numerical Techniques for Describing Data 3.1 Measures of Central Tendency 3.2 Measures of Variability.
Statistics for Business and Economics 8 th Edition Chapter 7 Estimation: Single Population Copyright © 2013 Pearson Education, Inc. Publishing as Prentice.
Theoretical distributions: the Normal distribution.
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 10 Descriptive Statistics Numbers –One tool for collecting data about communication.
Variety of characteristic
Normal Distribution and Parameter Estimation
Descriptive measures Capture the main 4 basic Ch.Ch. of the sample distribution: Central tendency Variability (variance) Skewness kurtosis.
CENTRAL MOMENTS, SKEWNESS AND KURTOSIS
Descriptive Statistics
Statistics in Applied Science and Technology
1.2 Introduction to Graphing Equations
Chapter 7 Estimation: Single Population
Chapter 5 Continuous Random Variables and Probability Distributions
Interval Estimation and Hypothesis Testing
Statistics: The Interpretation of Data
Hypothesis Testing: The Difference Between Two Population Means
Moments Presented By Miss. Amunekar Shubhangi Mahadev.
Presentation transcript:

Fig.1 shows the behaving of the derivatives of G and γ 1 with respect to z r in the case of the standard normal distribution. The comparison of the two derivatives shows that G is generally less sensitive to contaminations than γ 1. Indeed, when attention is focused on the middle part of the distribution, the interval where a contamination determines an increase of kurtosis is narrower for the G index; on the other hand, when attention is focused on the tails of the distribution, the intervals where a contamination determines an increase of kurtosis are larger for the G index. The conjoint use of the two indexes enables a better evaluation of the nonnormality. Figure 1: Derivatives of G and γ 1 for the standard normal distribution the standard normal distribution Fig.1 shows the behaving of the derivatives of G and γ 1 with respect to z r in the case of the standard normal distribution. The comparison of the two derivatives shows that G is generally less sensitive to contaminations than γ 1. Indeed, when attention is focused on the middle part of the distribution, the interval where a contamination determines an increase of kurtosis is narrower for the G index; on the other hand, when attention is focused on the tails of the distribution, the intervals where a contamination determines an increase of kurtosis are larger for the G index. The conjoint use of the two indexes enables a better evaluation of the nonnormality. Figure 1: Derivatives of G and γ 1 for the standard normal distribution the standard normal distribution Influence Function Approach to Sensitiveness of Kurtosis Indexes The kurtosis index attributed to Gini is given by G = σ 2 /δ 2, where σ is the standard deviation and δ is the mean deviation of the distribution. It is well known that always G≥ 1, that for the normal distribution G = π /2 and that values of G smaller (larger) than π /2 indicate platikurtosis (leptokurtosis). With reference to the originary Faleschini’s procedure, we study the partial derivatives of the kurtosis index with respect to the frequencies of each modality. As the partial derivatives of σ 2 and δ with respect to the frequency n r ( r =1,2,..., s ) are and, and, where is the size and is the arithmetic mean of the data, it follows that the partial derivative of the index G with respect to the frequency n r is By substituting, setting the derivative equal to zero and dividing for the nonnegative quantity G, we obtain the equation, whose roots are. In the case of the standard normal distribution, as G = π /2, the range of the derivative is divided in the following sections: (-∞, -2,009), (- 2,009, -0,498), (-0,498, 0,498), (0,498, 2,009), (2,009, ∞)., whose roots are. In the case of the standard normal distribution, as G = π /2, the range of the derivative is divided in the following sections: (-∞, -2,009), (- 2,009, -0,498), (-0,498, 0,498), (0,498, 2,009), (2,009, ∞). The kurtosis index attributed to Gini is given by G = σ 2 /δ 2, where σ is the standard deviation and δ is the mean deviation of the distribution. It is well known that always G≥ 1, that for the normal distribution G = π /2 and that values of G smaller (larger) than π /2 indicate platikurtosis (leptokurtosis). With reference to the originary Faleschini’s procedure, we study the partial derivatives of the kurtosis index with respect to the frequencies of each modality. As the partial derivatives of σ 2 and δ with respect to the frequency n r ( r =1,2,..., s ) are and, and, where is the size and is the arithmetic mean of the data, it follows that the partial derivative of the index G with respect to the frequency n r is By substituting, setting the derivative equal to zero and dividing for the nonnegative quantity G, we obtain the equation, whose roots are. In the case of the standard normal distribution, as G = π /2, the range of the derivative is divided in the following sections: (-∞, -2,009), (- 2,009, -0,498), (-0,498, 0,498), (0,498, 2,009), (2,009, ∞)., whose roots are. In the case of the standard normal distribution, as G = π /2, the range of the derivative is divided in the following sections: (-∞, -2,009), (- 2,009, -0,498), (-0,498, 0,498), (0,498, 2,009), (2,009, ∞). Let’s consider - only for simplicity - a variable T having a cumulative distribution function F(t), such that E (T) =0 and var( T )= σ 2.If we contaminate the variable T at the point x through a quantity ε (0<ε<1), the cumulative distribution function of the new variable is where H(t) equals 0 for t < x and 1 for t ≥ x. Moreover its mean is, its variance is as Fiori and Zenga (2005) derived by performing some computations, and its mean deviation is as Fiori and Zenga (2005) derived by performing some computations, and its mean deviation is As the first integral can be split in two parts (from -∞ to εx and from εx to ∞), while the second one equals, the mean deviation can be written as Noting that the integral of the constant ε x from -∞ to ∞ equals ε x and the integral of t from -∞ to ∞ equals zero (as T is expressed in terms of deviation from its mean), the previous expression becomes where F 1 (t) is the incomplete first moment of T. At this point we can use the following relationship between F(t) and F 1 (t) in order to derive It is just the case to note that The obtained expressions of and enables to derive the influence function of G Preliminarily we verify that and, and, where equals 1 for x < 0 and 0 for x ≥ 0; then we find, by means of de L’Hospital’s rule, that where equals 1 for x < 0 and 0 for x ≥ 0; then we find, by means of de L’Hospital’s rule, that In the case of symmetrical variables, in which F (0)=1/2, after considering the standardized value, the influence function becomes At last, by setting the influence function equal to zero and dividing it for the nonnegative quantity G, we obtain the equation which is the same as the one obtained by means of Faleschini’s procedure. Let’s consider - only for simplicity - a variable T having a cumulative distribution function F(t), such that E (T) =0 and var( T )= σ 2.If we contaminate the variable T at the point x through a quantity ε (0<ε<1), the cumulative distribution function of the new variable is where H(t) equals 0 for t < x and 1 for t ≥ x. Moreover its mean is, its variance is as Fiori and Zenga (2005) derived by performing some computations, and its mean deviation is as Fiori and Zenga (2005) derived by performing some computations, and its mean deviation is As the first integral can be split in two parts (from -∞ to εx and from εx to ∞), while the second one equals, the mean deviation can be written as Noting that the integral of the constant ε x from -∞ to ∞ equals ε x and the integral of t from -∞ to ∞ equals zero (as T is expressed in terms of deviation from its mean), the previous expression becomes where F 1 (t) is the incomplete first moment of T. At this point we can use the following relationship between F(t) and F 1 (t) in order to derive It is just the case to note that The obtained expressions of and enables to derive the influence function of G Preliminarily we verify that and, and, where equals 1 for x < 0 and 0 for x ≥ 0; then we find, by means of de L’Hospital’s rule, that where equals 1 for x < 0 and 0 for x ≥ 0; then we find, by means of de L’Hospital’s rule, that In the case of symmetrical variables, in which F (0)=1/2, after considering the standardized value, the influence function becomes At last, by setting the influence function equal to zero and dividing it for the nonnegative quantity G, we obtain the equation which is the same as the one obtained by means of Faleschini’s procedure. Faleschini L. (1948) Su alcune proprietà dei momenti impiegati nello studio della variabilità, asimmetria e curtosi, Statistica, 8, Fiori A.M., Zenga M. (2005) L’interpretazione della curtosi e la sua curva di influenza in un’intuizione di L. Faleschini, Statistica, anno LXV, 2005, 2, Hampel F.R. (1974) The influence curve and its role in robust estimation, Journal of the American Statistical Association, 69, Ruppert D. (1987) What is kurtosis? An influence function approach, The American Statistician, 41, 1, 1-5. Faleschini L. (1948) Su alcune proprietà dei momenti impiegati nello studio della variabilità, asimmetria e curtosi, Statistica, 8, Fiori A.M., Zenga M. (2005) L’interpretazione della curtosi e la sua curva di influenza in un’intuizione di L. Faleschini, Statistica, anno LXV, 2005, 2, Hampel F.R. (1974) The influence curve and its role in robust estimation, Journal of the American Statistical Association, 69, Ruppert D. (1987) What is kurtosis? An influence function approach, The American Statistician, 41, 1, 1-5. Fiori and Zenga (2005) recently analysed the meaning of the kurtosis index introduced by K. Pearson, starting from a Faleschini’s pioneeristic work (1948) and arriving to identify its answer with respect to the frequencies of each modality. The authors achieved the same results as those achieved by Faleschini, using the more modern methodology of the influence function suggested by Hampel (1974) and utilized in this context by Ruppert (1987).The aim of this paper is to go through these results and to compare the kurtosis index attributed to Gini with Pearson’s one. The sensitiveness of kurtosis index attributed to Gini The influence function in the continuous case Francesco Campobasso, Dipartimento di Scienze Statistiche Carlo Cecchi, Università degli Studi di Bari, Italy, Introduction REFERENCES Final remarks