Presentation is loading. Please wait.

Presentation is loading. Please wait.

All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.

Similar presentations


Presentation on theme: "All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer."— Presentation transcript:

1 All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer

2 Overview What are we studying? Probability What is probability? The mathematical language for quantifying uncertainty Why are we studying probability? Dan has his reasons…

3 Motivation These are the kind of questions that we hope to be able to answer.

4 Review Definition: Random Variable – a random variable X is a mapping from the sample space to the real line Example: Flip a fair coin twice. The sample space is every combination of heads and tails. Choose the random variable X to be the number of heads. Let our outcome be one head and one tail. X maps this outcome to 1. Note: The notation can be confusing. In the book, usually X denotes the map and X with a subscript denotes a real number. However, this is not always the case, so you must examine the context to be sure of the meaning.

5 Sequences of Random Variables Much of probability theory is concerned with large sequences of random variables. This study is sometimes known as large sample theory, limit theory, or asymptotic theory. What is a sequence of random variables? Simply a set of indexed set of random variables. We will be interested in sequences that have some interesting limiting behavior (i.e. we can say something about them as n gets large)

6 A Special Kind of Sequence of Random Variables As it turns out, a very common and particularly useful class of sequences of random variables is IID Definition: IID – Independent and identically distributed Independent – essentially, the value of one random variable doesn’t effect the value of any other; for instance, a coin doesn’t remember which side last landed up, so consecutive flips are said to be independent Identically distributed – each random variable X has associated with it a cumulative distribution function (CDF) which is derived from the probability measure. The CDF gives the probability of the value of the random variable being less than or equal to a certain value. When two or more random variables have the same CDF, we say they are identically distributed.

7 Statements about sequences of random variables Given a sequence of random variables, most likely IID, it would be useful to be able to make statements of the form: The average of all X i will be between two values with certain probability. Example: Flip a fair coin n times. The average number of heads per toss will be between.4 and.6 with probability greater than or equal to 70% if we flip 84 times (n=84). Or, what is the probability that the average of X i is less/greater than a certain value? These kind of statements can be made with the help of the Weak Law of Large Numbers (WLLN) and the Central Limit Theorem (CLT), respectively. So why not state them now? “Hold your horses” there, Makarand. The statement of the WLLN and the CLT make use of a few different types of convergence which must be discussed first.

8 Types of Convergence There are two main types of convergence Convergence in Probability (CIP) A sequence of random variables is said to converge in probability to X if the probability of it differing from X goes to zero as n gets large. Convergence in Distribution (CID) A sequence of random variables is said to converge in distribution to X if the limit of the corresponding CDFs is the CDF of X. There is also another type, called convergence in quadratic mean, which is used primarily because it is stronger than CIP or CID (it implies CIP and CID) and it can be computed relatively easily.

9 Weak Law of Large Numbers If a sequence of random variables are IID, then the sample average converges in probability to the expectation value. On the left we have information about many trials and on the right we have information about the relative likelihood of the different values a random variable can take on. In words, the WLLN says that the distribution of the sample average becomes more concentrated around the expectation as n gets large.

10 Example of using the WLLN Consider flipping a coin for which the probability of heads is p. Let X i denote the outcome of a single toss (either 0 or 1). Hence p = P(X i =1)=E(X i ). The first equality is a definition. The second equality is obtained by averaging over the distribution. The fraction of heads after n tosses is equal to the sample average. Note that the X i are IID. Therefore the WLLN can be applied. The WLLN says that the sample average converges to p = E(X i ) in probability. You may find yourself wondering, how many times must I flip this coin such that the sample average is between.4 and.6 with probability greater than or equal to 70%? The WLLN tells you that it is possible to find such an n. The inequalities that Justin presented on from Chapter 4 can be used to show that n=84 does the trick in this case, but I’ll spare you the details.

11 The Central Limit Theorem Given a sequence of random variables with a mean and a variance, the CLT says that the sample average has a distribution which is approximately Normal, and gives the new mean and variance. Notice that nothing at all need be assumed about the P, CDF, or PDF associated with X, which could have any distribution from which a mean and variance can be derived.

12 Example of using the CLT Suppose that the number of errors per computer program has a Poisson distribution with mean 5. We get 125 programs. Approximately what is the probability that the average number of errors per computer program is less than 5.5? In this case 125 is the sample size, which we hope is large enough to make a good approximation. The approximation we are making here is that the sample average will have a Normal distribution. Taking the sample size, mean and variance into account, it is possible to show that the question asked is equivalent to the probability of the standard Normal distribution being less than 2.5, which turns out to be approximately 0.9983.

13 Topics in Chapter 5 not covered in this presentation All proofs Slutzky’s theorem and related theorems The effect of adding sequences of random variables on their convergence behavior Multivariate central limit theorem CLT with IID random vectors instead of variables The delta method The effect of applying a smooth function to a sequence of random variables on its limiting behavior

14 Interesting problems 6 & 8

15 Bibliography Chapters 1-5 of All of Statistics by Larry Wasserman


Download ppt "All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer."

Similar presentations


Ads by Google