Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :

Slides:



Advertisements
Similar presentations
Random Processes Introduction (2)
Advertisements

Let X 1, X 2,..., X n be a set of independent random variables having a common distribution, and let E[ X i ] = . then, with probability 1 Strong law.
DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Chapter 3 Brownian Motion 報告者:何俊儒.
Random Variable A random variable X is a function that assign a real number, X(ζ), to each outcome ζ in the sample space of a random experiment. Domain.
1. (f) Use continuity corrections for discrete random variable LEARNING OUTCOMES At the end of the lesson, students will be able to (g) Use the normal.
Review of Basic Probability and Statistics
Chapter 5 Basic Probability Distributions
Review.
Section 5.7 Suppose X 1, X 2, …, X n is a random sample from a Bernoulli distribution with success probability p. Then Y = X 1 + X 2 + … + X n has a distribution.
Prof. Bart Selman Module Probability --- Part d)
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Probability Distributions
The moment generating function of random variable X is given by Moment generating function.
Introduction to Probability and Statistics Chapter 7 Sampling Distributions.
Week 51 Theorem For g: R  R If X is a discrete random variable then If X is a continuous random variable Proof: We proof it for the discrete case. Let.
1 Sampling Distribution Theory ch6. 2  Two independent R.V.s have the joint p.m.f. = the product of individual p.m.f.s.  Ex6.1-1: X1is the number of.
Approximations to Probability Distributions: Limit Theorems.
Chapter 21 Random Variables Discrete: Bernoulli, Binomial, Geometric, Poisson Continuous: Uniform, Exponential, Gamma, Normal Expectation & Variance, Joint.
Standard error of estimate & Confidence interval.
Probability and Statistics Dr. Saeid Moloudzadeh Axioms of Probability/ Basic Theorems 1 Contents Descriptive Statistics Axioms of Probability.
Chapter 1 Probability and Distributions Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
Limits and the Law of Large Numbers Lecture XIII.
All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.
4.2 Variances of random variables. A useful further characteristic to consider is the degree of dispersion in the distribution, i.e. the spread of the.
Chapter 7. Statistical Intervals Based on a Single Sample Weiqi Luo ( 骆伟祺 ) School of Software Sun Yat-Sen University : Office.
1 Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND STATISTICS FOR SCIENTISTS AND ENGINEERS Systems.
Lectures prepared by: Elchanan Mossel elena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Chapter 10 – Sampling Distributions Math 22 Introductory Statistics.
Convergence in Distribution
X = 2*Bin(300,1/2) – 300 E[X] = 0 Y = 2*Bin(30,1/2) – 30 E[Y] = 0.
Chapter 4 DeGroot & Schervish. Variance Although the mean of a distribution is a useful summary, it does not convey very much information about the distribution.
King Saud University Women Students
Chapter 5.6 From DeGroot & Schervish. Uniform Distribution.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
STA347 - week 51 More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
Chapter 7: Introduction to Sampling Distributions Section 2: The Central Limit Theorem.
One Random Variable Random Process.
Probability Review-1 Probability Review. Probability Review-2 Probability Theory Mathematical description of relationships or occurrences that cannot.
ENGG 2040C: Probability Models and Applications Andrej Bogdanov Spring Limit theorems.
Random Variable The outcome of an experiment need not be a number, for example, the outcome when a coin is tossed can be 'heads' or 'tails'. However, we.
Chapter 4-5 DeGroot & Schervish. Conditional Expectation/Mean Let X and Y be random variables such that the mean of Y exists and is finite. The conditional.
Week 121 Law of Large Numbers Toss a coin n times. Suppose X i ’s are Bernoulli random variables with p = ½ and E(X i ) = ½. The proportion of heads is.
Chapter 2. Conditional Probability Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Probability Distributions, Discrete Random Variables
1 Probability and Statistical Inference (9th Edition) Chapter 5 (Part 2/2) Distributions of Functions of Random Variables November 25, 2015.
Discrete Random Variables. Introduction In previous lectures we established a foundation of the probability theory; we applied the probability theory.
1 Sampling distributions The probability distribution of a statistic is called a sampling distribution. : the sampling distribution of the mean.
Probability and Moment Approximations using Limit Theorems.
Section 10.5 Let X be any random variable with (finite) mean  and (finite) variance  2. We shall assume X is a continuous type random variable with p.d.f.
Chapter 5 Probability Distributions 5-1 Overview 5-2 Random Variables 5-3 Binomial Probability Distributions 5-4 Mean, Variance and Standard Deviation.
Central Limit Theorem Let X 1, X 2, …, X n be n independent, identically distributed random variables with mean  and standard deviation . For large n:
Chapter 5 Special Distributions Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Chebyshev’s Inequality Markov’s Inequality Proposition 2.1.
1 Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND STATISTICS FOR SCIENTISTS AND ENGINEERS Systems.
Sums of Random Variables and Long-Term Averages Sums of R.V. ‘s S n = X 1 + X X n of course.
Conditional Expectation
Random Variables Lecture Lecturer : FATEN AL-HUSSAIN.
Week 61 Poisson Processes Model for times of occurrences (“arrivals”) of rare phenomena where λ – average number of arrivals per time period. X – number.
Jiaping Wang Department of Mathematical Science 04/22/2013, Monday
CHAPTER 2 RANDOM VARIABLES.
Lecture 13 Sections 5.4 – 5.6 Objectives:
Chapter 5 Some Important Discrete Probability Distributions
Distributions and expected value
Chebychev, Hoffding, Chernoff
Lecture 41 Section 14.1 – 14.3 Wed, Nov 14, 2007
Further Topics on Random Variables: 1
Presentation transcript:

Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :

School of Data & Computer Science  6.1 Introduction  6.2 The Law of Large Numbers  6.3 The Central Limit Theorem  6.4 The Correction for Continuity 2 Chapter 6: Large Random Samples

School of Data & Computer Science 6.1 Introduction …… Relative frequency: n(A)/n Number of experiments performed p Relative Frequency vs. Probability (the probability of getting the head)

School of Data & Computer Science  Theorem Markov Inequality. Suppose that X is a random variable such that Pr(X ≥ 0) = 1. Then for every real number t > 0, Proof For convenience, we shall assume that X has a discrete distribution for which the p.f. is f. Since X can have only nonnegative values, all the terms in the summations are nonnegative. Therefore, 6.2 The Law of Large Numbers 4

School of Data & Computer Science  Theorem Chebyshev Inequality. Let X be a random variable for which Var(X) exists. Then for every number t > 0, Proof Let Y = [X − E(X)] 2. Then Pr(Y ≥ 0) = 1 and E(Y) = Var(X). By applying the Markov inequality to Y, we obtain the following result: 6.2 The Law of Large Numbers 5

School of Data & Computer Science  Theorem Mean and Variance of the Sample Mean. Let X 1,..., X n be a random sample from a distribution with mean μ and variance σ 2. Let be the sample mean, then 6.2 The Law of Large Numbers 6

School of Data & Computer Science  Example Determining the Required Number of Observations. Suppose that a random sample is to be taken from a distribution for which the value of the mean μ is not known, but for which it is known that the standard deviation σ is 2 units or less.We shall determine how large the sample size must be in order to make the probability at least 0.99 that |X n − μ| will be less than 1 unit. Since σ 2 ≤ 2 2 = 4, for every sample size n, 6.2 The Law of Large Numbers 7

School of Data & Computer Science  Example Tossing a Coin. Suppose that a fair coin is to be tossed n times independently. For i = 1,..., n, let X i = 1 if a head is obtained on the ith toss, and let X i = 0 if a tail is obtained on the ith toss. We shall determine the number of times the coin must be tossed in order to make 6.2 The Law of Large Numbers 8 If n ≥ 84, this probability will be at least tosses would actually be sufficient to satisfy the specified probability requirement.

School of Data & Computer Science  Definition Convergence in Probability. Sequence Z 1, Z 2,... of random variables converges to b in probability if for every number ε > 0, This property is denoted by and is sometimes stated simply as Z n converges to b in probability. 6.2 The Law of Large Numbers 9

School of Data & Computer Science  Theorem Law of Large Numbers. Suppose that X 1,..., X n form a random sample from a distribution for which the mean is μ and for which the variance is finite. Let X n denote the sample mean. Then Proof Let the variance of each X i be σ 2. It then follows from the Chebyshev inequality that for every number ε > 0, 6.2 The Law of Large Numbers 10

School of Data & Computer Science  Homework Ex. 3, 4, The Law of Large Numbers 11

School of Data & Computer Science 6.3 The Central Limit Theorem 12

School of Data & Computer Science  The Central Limit Theorem (CLT) Let X 1, X 2, …, X n be a random sample from a distribution (may or may not be normal) with mean μ and variance σ 2. Then if n is sufficiently large, has approximately a normal distribution with T o also has approximately a normal distribution with The larger the value of n, the better the approximation 6.3 The Central Limit Theorem 13

School of Data & Computer Science  Theorem Central Limit Theorem. If the random variables X 1,..., X n form a random sample of size n from a given distribution with mean μ and variance σ 2 (0 < σ 2 <∞), then for each fixed number x, where Ф denotes the c.d.f. of the standard normal distribution. 6.3 The Central Limit Theorem 14

School of Data & Computer Science  An Example for Uniform Distribution 6.3 The Central Limit Theorem 15

School of Data & Computer Science  An Example for Triangular Distribution 6.3 The Central Limit Theorem 16

School of Data & Computer Science  Example Tossing a Coin. Suppose that a fair coin is tossed 900 times. We shall approximate the probability of obtaining more than 495 heads. For i = 1,..., 900, let X i = 1 if a head is obtained on the ith toss and let X i = 0 otherwise. Then E(X i ) = 1/2 and Var(X i ) = 1/4. Then will be approximately the normal distribution for which the mean is (900)(1/2) = 450, the variance is (900)(1/4) = 225, 6.3 The Central Limit Theorem 17

School of Data & Computer Science  Example Sampling from a Uniform Distribution. Suppose that a random sample of size n = 12 is taken from the uniform distribution on the interval [0, 1]. We shall approximate the value of E(X i ) = ½; V(X i ) = 1/12;  6.3 The Central Limit Theorem 18

School of Data & Computer Science  Example Poisson Random Variables. Suppose that X 1,..., X n form a random sample from the Poisson distribution with mean θ. Let X n be the average. Then μ = θ and σ 2 = θ. The central limit theorem says that has approximately the standard normal distribution. The probability that |X n − θ| is less than some small number c could be approximated using the standard normal c.d.f.: 6.3 The Central Limit Theorem 19

School of Data & Computer Science  Homework Ex. 3, 4, The Central Limit Theorem 20

School of Data & Computer Science  Approximating a Discrete Distribution by a Continuous Distribution 6.4 The Correction for Continuity 21 How to improve the approximation?

School of Data & Computer Science If the distribution of Y provides a good approximation to the distribution of X, then for all integers a and b, we can approximate the discrete probability This simple approximation has the following shortcoming: 1.Although Pr(X ≥ a) and Pr(X > a) will typically have different values for the discrete distribution of X, Pr(Y ≥ a) = Pr(Y > a) because Y has a continuous distribution. 2.Although Pr(X = x) > 0 for each integer x that is a possible value of X, Pr(Y = x) = 0 for all x. 6.4 The Correction for Continuity 22 Discrete cases Continuous cases

School of Data & Computer Science 6.4 The Correction for Continuity 23

School of Data & Computer Science 6.4 The Correction for Continuity 24

School of Data & Computer Science  Example Coin Tossing. Suppose that a fair coin is tossed 20 times and that all tosses are independent. What is the probability of obtaining exactly 10 heads? Let X denote the total number of heads obtained in the 20 tosses. According to the central limit theorem, the distribution of X will be approximately the normal distribution with mean 10 and standard deviation [(20)(1/2)(1/2)] 1/2 = If we use the correction for continuity, 6.4 The Correction for Continuity using the binomial distribution

School of Data & Computer Science  Homework Ex. 1, The Correction for Continuity 26