Presentation on theme: "Chapter 3 Properties of Random Variables Moments and Expectation."— Presentation transcript:
Chapter 3 Properties of Random Variables Moments and Expectation
Review Experiment Random Variable Observation Realization Parameter Sample Statistic
Moments One way to quantify the location and some measures of the shape of the pdf.
First moment about the origin
i th moment about the origin continuous random variable discrete random variable
i th central moment about the mean,
Expected value of a random variable X X continuous X discrete
Expected value of function of X, g(X) X continuous X discrete
Expected value and the first moment about the origin Comparing to You can see that the expected value of the random value x is the first moment about the origin.
Rules for finding expected values
Measures of central tendency Arithmetic mean Geometric mean Median Mode Weighted Mean
Mean, x, or average value Mean of a r.v. X is its expected value. Sample estimate is the arithmetic average. Arithmetic mean of grouped data (k is number of groups, n is total number of observations, n i is the number of observations in group i, x i is the class mark of the i th group.
Geometric mean Used when the ratio of two consecutive observations is either constant or nearly constant. The logarithm of the population geometric mean would be the expected value of the logarithm of X.
Median, X md The observation such that half of the values in the sample lie on either side of X md. The median may not exist. Population median, md would be the value satisfying: X continuous X discrete
Mode, mo Most frequently occurring value. The sample or population may have none, one or more than one mode. Population mode is the value of X maximizing p x (x). X continuous X discrete
Weighted mean Used for describing the central tendancy of grouped data.
Measures of Dispersion Measures of the spread of the data Range Variance
Range Difference between the largest and smallest sample values. For a population this interval often ranges from - to or from 0 to. The sample range is a function of only 2 of the sample values, but does convey some idea of the spread of the data. Disadvantage of range: does not reflect frequency or magnitude of values that deviate from the mean. Occasionally use the relative range Relative range =
Variance, 2 Defined as the second moment about the mean. The average squared deviation from the mean. For a discrete population of size n: Sample estimate of x 2 is s x 2
Variance Two basic differences between population and sample variance. used instead of n-1 is used as the denominator rather than n to avoid a biased estimate for s x 2 Variance of grouped data
Rules for finding the Variance
Units of Variance Units of the variance are the same as units on X 2. Units on its positive square root, the standard deviation, x, are the same as the units of the random variable, X. A dimensionless measure of dispersion is the coefficient of variation, C v.
Measures of Symmetry Many distributions are not symmetrical Tailing off to the right or the left is skewing the distribution. Tailing to the right-positively skewed Tailing to the left-negatively skewed
Skewness 3 rd moment about the mean
Practical measurements of skewness One measure of absolute skew is to measure the difference between the mean and the mode. Not meaningful for comparison sake because it is dependent on units of measure.
Pearsons first coefficient of skewness Relative measure of skewness more useful for comparison. Population skewness Sample skewness
Measures of Peakedness (Flatness) Kurtosis refers to the extent of peakedness of a probability distribution in comparison to the normal distribution. Kurtosis is the 4 th moment about the mean. Calculate the coefficient of kurtosis,
Covariance Measure of the linear relationship between two jointly distributed random variables, X and Y. Covariance is the 1,1 central moment
Covariance If X and Y are independent: Sample statistic:
Correlation Coefficient Normalized covariance If X and Y are independent
Correlation Coefficient Measure of how two variables vary together. A value of equal to positive one implies that X and Y are perfectly related by Y=a+bX. Positive values indicate large (small) values of X tend to be paired with large (small) values of Y. Negative values indicate large (small) values of X tend to be paired with small (large) values of Y. Two values are uncorrelated ONLY IF (x,y)=0. Correlation does NOT equal cause and effect.
Correlation coefficient: Linear dependence and functional dependence