Presentation on theme: "Topic 2: Statistical Concepts and Market Returns"— Presentation transcript:
1 Topic 2: Statistical Concepts and Market Returns
2 Descriptive Statistics The arithmetic mean is the sum of the observations divided by the number of observations.The population mean is given by µThe sample mean looks at the arithmetic average of the sample of data.The median is the value of the middle item of a set of items that has been sorted by ascending or descending order.The mode is the most frequently occurring value in a distribution.The weighted mean allows us to place greater importance on different observations. For example, we may choose to give larger companies greater weight in our computation of an index. In this case, we would weight each observation based on its relative size.
3 Descriptive Statistics The geometric mean is most frequently used to average rates of change over time or to compute the growth rate of a variable.Geometric Mean Using Natural LogsThe geometric mean return allows us to compute the average return when there is compounding.
4 Descriptive Statistics Quartiles divide the data into quarters.Quintiles divide the data into fifths.Deciles divide the data into tenths.Percentiles divide the data into hundredths.Variance measures the average squared deviation from the mean.Population Standard DeviationSample VarianceSample Standard Deviation
5 Descriptive Statistics Often times, observations above the mean are good, the variance is not a good measure of risk. Semivariance looks at the average squared deviations below the mean.The coefficient of variation is the ratio of the standard deviation to their mean value.measure of relative dispersioncan compare the dispersion of data with different scalesSkewness measures the symmetry of a distribution.A symmetric distribution has a skewness of 0.Positive skewness indicates that the mean is greater than the median (more than half the deviations from the mean are negative)Negative skewness indicates that the mean is less than the median (less than half the deviations from the mean are negative)
6 Binomial Distribution Sometimes a random variable can only take on two values, success or failure. This is referred to as a Bernoulli random variable.A Bernoulli trial is an experiment that produces only two outcomes.Y = 1 for success and Y = 0 for failure.A binomial random variable X is defined as the number of successes in n Bernoulli trials.Binomial distribution assumesThe probability, p, of success is constant for all trialsThe trials are independent
11 Normal DistributionApproximately 50 percent of all observations fall in the interval μ ± (2/3)σ.Approximately 68 percent of all observations fall in the interval μ ± σ.Approximately 95 percent of all observations fall in the interval μ ± 2σ.Approximately 99 percent of all observations fall in the interval μ ± 3σ.Standard normal distribution has a mean of zero and a standard deviation of 1. We use Z to denote the standard normal random variable.The lognormal distribution is widely used for modeling the probability distribution of asset prices.
13 Statistical Inference In statistics we are often times interested in obtaining information about the value of some parameter of a population.To obtain this information we usually take a smaller subset of the population and try to draw some conclusions from this sample.Sampling distribution of a statistic is the distribution of all the distinct possible values that the statistic can assume when computed from samples of the same size randomly drawn from the same population.Cross-sectional data represent observations over individual units at a point in time, as opposed to time series data.Time series data is a set of observations on a variable’s outcomes in different time periods.Investment analysts commonly work with both time-series and cross-sectional data.
14 Central Limit TheoremThe central limit theorem states that for large sample sizes, for any underlying distribution for a random variable, the sampling distribution of the sample mean for that variable will be approximately normal, with mean equal to the population mean for that random variable and variance equal to the population variance of the variable divided by sample size.
15 Standard Error of the Sample Mean For a sample mean calculated from a sample generated from a population with standard deviation σ, the standard error of the sample mean iswhen we know σ.If the population standard deviation is unknown we have,In practice, the population variance is almost always unknown. To compute the sample standard deviation we use,
16 Point and Interval Estimates of the Population Mean An estimator is a formula for estimating a parameter. An estimate is a particular value that we calculate from a sample by using an estimator.An unbiased estimator is one whose expected value equals the parameter it is intended to estimate.An unbiased estimator is efficient if no other unbiased estimator of the same population parameter has a sampling distribution with smaller variance.A consistent estimator is one for which the probability of estimates is close to the value of the population parameter increases as sample size increases.A confidence interval is an interval for which we can assert with a given probability 1 − α, called the degree of confidence, that it will contain the parameter it is intended to estimate.
17 Confidence Intervals for the Population Mean For normally distributed population with known variance.For large sample, population variance unknown.
18 Confidence Intervals for the Population Mean Population variance unknown, t-DistributionThe t-distribution is a symmetrical probability distribution defined by a single parameter known as degrees of freedom (df).
19 Student’s t-Distribution versus the Standard Normal Distribution
20 Selection of Sample Size All else equal, a larger sample size decreases the width of the confidence interval.
21 Bias in SamplingSample selection bias is the error of distorting a statistical analysis due to how the samples are collected.Look-ahead bias occurs when information that was not available on the test date is used in the estimation.Time-period bias occurs when the test is based on a time period that may make the results time-period specific.Survivorship bias occurs if companies are excluded from the analysis because they have gone out of business or because of reasons related to poor performance.Data mining bias – Data mining is the practice of determining a model by extensive searching through a dataset for statistically significant patterns.An out-of-sample test uses a sample that does not overlap the time period(s) of the sample(s) on which a variable, strategy, or model, was developed.
22 Hypothesis TestingOften times we are interested in testing the validity of some statement.For example, Is the underlying mean return on this mutual fund different from the underlying mean return on its benchmark?Hypothesis testing is part of the branch of statistics known as statistical inference.A hypothesis is a statement about one or more populations.
23 Steps in Hypothesis Testing 1. Stating the hypotheses.2. Identifying the appropriate test statistic and its probability distribution.3. Specifying the significance level.4. Stating the decision rule.5. Collecting the data and calculating the test statistic.6. Making the statistical decision.
24 Null vs. Alternative Hypothesis The null hypothesis is the hypothesis to be tested.The alternative hypothesis is the hypothesis accepted when the null hypothesis is rejected.
25 Formulation of Hypotheses H0: θ = θ0 versus HA: θ ≠ θ0H0: θ ≤ θ0 versus HA: θ > θ0H0: θ ≥ θ0 versus HA: θ < θ0The first formulation is a two-sided test. The other two are one-sided tests.
26 Test StatisticA test statistic is a quantity, calculated based on a sample, whose value is the basis for deciding whether or not to reject the null hypothesis.In reaching a statistical decision, we can make two possible errors:We may reject a true null hypothesis (a Type I error), orWe may fail to reject a false null hypothesis (a Type II error).The level of significance of a test is the probability of a Type I error that we accept in conducting a hypothesis test, is denoted by α.The standard approach to hypothesis testing involves specifying a level of significance (probability of Type I error) only.The power of a test is the probability of correctly rejecting the null (rejecting the null when it is false).A rejection point (critical value) for a test statistic is a value with which the computed test statistic is compared to decide whether to reject or not reject the null hypothesis.
27 Test StatisticThe p-value is the smallest level of significance at which the null hypothesis can be rejected.The smaller the p-value, the stronger the evidence against the null hypothesis and in favor of the alternative hypothesis.
28 Hypothesis Tests Concerning the Mean Can test that the mean of a population is equal to or differs from some hypothesized value.Can test to see if the sample means from two different populations differ.
29 Tests Concerning a Single Mean A t-test is usually used to test a hypothesis concerning the value of a population mean.If the variance is unknown and the sample is large, or the sample is small but the population is normally distributed, or approximately normally distributed.
30 Tests Concerning a Single Mean If the population sampled is normally distributed with known variance σ2, then the test statistic for a hypothesis test concerning a single population mean, µ, is
31 Tests Concerning a Single Mean If the population sampled has unknown variance and the sample is large, in place of a t-test, an alternative statistic is
32 Rejection Points for a z-Test For α = 0.10 H0: θ = θ0 verus Ha: θ ≠ θ0Reject the null hypothesis if z > or if z <2. H0: θ ≤ θ0 verus Ha: θ > θ0Reject the null hypothesis if z > 1.283. H0: θ ≥ θ0 verus Ha: θ < θ0Reject the null hypothesis if z < -1.28
33 Rejection Points for a z-Test For α = 0.05 H0: θ = θ0 verus Ha: θ ≠ θ0Reject the null hypothesis if z > 1.96 or if z <2. H0: θ ≤ θ0 verus Ha: θ > θ0Reject the null hypothesis if z > 1.6453. H0: θ ≥ θ0 verus Ha: θ < θ0Reject the null hypothesis if z <
34 Rejection Points for a z-Test For α = 0.01 H0: θ = θ0 verus Ha: θ ≠ θ0Reject the null hypothesis if z > or if z <2. H0: θ ≤ θ0 verus Ha: θ > θ0Reject the null hypothesis if z > 2.333. H0: θ ≥ θ0 verus Ha: θ < θ0Reject the null hypothesis if z < -2.33
35 Rejection Points, 0.05 Significance Level, Two-Sided Test of the Population Mean Using a z-Test
36 Rejection Point, 0.05 Significance Level, One-Sided Test of the Population Mean Using a z-Test
37 Tests Concerning the Differences between Means Sometimes we are interested in testing whether the mean value differs between two groups.If reasonable to assumenormally distributedsamples are independentWe can combine observations from both samples to get a pooled estimate of the unknown population variance.
39 Test Statistic for a Test of Difference between 2 Population Means Normally distributed populations, population variances unknown, but assumed to be equal.Pooled Estimator of the Common Variancedegrees of freedom is n1 + n2 - 2
40 Test Statistic for a Test of Difference between 2 Population Means Normally distributed populations, population variances unequal and unknown.Degrees of freedom is given by
41 Mean Differences – Populations Not Independent If the samples are not independent, a test of mean difference is done using paired observations.H0: µd = µd0 versus HA: µd ≠ µd0H0: µd ≤ µd0 versus HA: µd > µd0H0: µd ≥ µd0 versus HA: µd < µd0
42 Mean Differences – Populations Not Independent To calculate the t-statistic, we first need to find the sample mean difference:The sample variance isThe standard deviation of the mean isThe test statistic, with n – 1 df is,
43 Hypothesis Tests Concerning Variance We examine two types:tests concerning the value of a single population variance andtests concerning the differences between two population variances.We can formulate hypotheses as follows:
44 Tests Concerning the Value of a Population Variance (Normal Dist) where,
45 Tests Concerning the Equality of Two Variances We can formulate hypotheses as follows:Suppose we have two samples, the first with n1 observations and the second with n2 observations
46 Nonparametric Inference A nonparametric test is not concerned with a parameter or makes minimal assumptions about the population being sampled.A nonparametric test is primarily used in three situations: when data do not meet distributional assumptions, when data are given in ranks, or when the hypothesis we are addressing does not concern a parameter.
47 The Spearman Rank Correlation Coefficient The Spearman rank correlation coefficient is calculated on the ranks of two variables within their respective samples.