Probability and Moment Approximations using Limit Theorems.

Slides:



Advertisements
Similar presentations
Random Processes Introduction (2)
Advertisements

Evaluating Classifiers
Random Variables ECE460 Spring, 2012.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Chapter 3 Brownian Motion 報告者:何俊儒.
Random Variable A random variable X is a function that assign a real number, X(ζ), to each outcome ζ in the sample space of a random experiment. Domain.
Chap 8: Estimation of parameters & Fitting of Probability Distributions Section 6.1: INTRODUCTION Unknown parameter(s) values must be estimated before.
STAT 497 APPLIED TIME SERIES ANALYSIS
ELEC 303 – Random Signals Lecture 18 – Statistics, Confidence Intervals Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 10, 2009.
1 Chap 5 Sums of Random Variables and Long-Term Averages Many problems involve the counting of number of occurrences of events, computation of arithmetic.
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Evaluating Hypotheses
Tch-prob1 Chapter 4. Multiple Random Variables Ex Select a student’s name from an urn. S In some random experiments, a number of different quantities.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
C4: DISCRETE RANDOM VARIABLES CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics Longin Jan Latecki.
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
1 Sampling Distribution Theory ch6. 2  Two independent R.V.s have the joint p.m.f. = the product of individual p.m.f.s.  Ex6.1-1: X1is the number of.
Approximations to Probability Distributions: Limit Theorems.
Random Variable and Probability Distribution
Lecture II-2: Probability Review
Continuous Probability Distribution  A continuous random variables (RV) has infinitely many possible outcomes  Probability is conveyed for a range of.
R. Kass/S07 P416 Lec 3 1 Lecture 3 The Gaussian Probability Distribution Function Plot of Gaussian pdf x p(x)p(x) Introduction l The Gaussian probability.
Standard error of estimate & Confidence interval.
Sampling Distributions  A statistic is random in value … it changes from sample to sample.  The probability distribution of a statistic is called a sampling.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Simulation Output Analysis
All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.
1 As we have seen in section 4 conditional probability density functions are useful to update the information about an event based on the knowledge about.
Dept of Bioenvironmental Systems Engineering National Taiwan University Lab for Remote Sensing Hydrology and Spatial Modeling STATISTICS Random Variables.
CHAPTER 4 Multiple Random Variable
Convergence in Distribution
Lab 3b: Distribution of the mean
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
One Random Variable Random Process.
CpSc 881: Machine Learning Evaluating Hypotheses.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Mean, Variance, Moments and.
7 sum of RVs. 7-1: variance of Z Find the variance of Z = X+Y by using Var(X), Var(Y), and Cov(X,Y)
ENGG 2040C: Probability Models and Applications Andrej Bogdanov Spring Limit theorems.
Conditional Probability Mass Function. Introduction P[A|B] is the probability of an event A, giving that we know that some other event B has occurred.
Week 121 Law of Large Numbers Toss a coin n times. Suppose X i ’s are Bernoulli random variables with p = ½ and E(X i ) = ½. The proportion of heads is.
Multiple Discrete Random Variables. Introduction Consider the choice of a student at random from a population. We wish to know student’s height, weight,
Computer simulation Sep. 9, QUIZ 2 Determine whether the following experiments have discrete or continuous out comes A fair die is tossed and the.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Expected values of discrete Random Variables. The function that maps S into S X in R and which is denoted by X(.) is called a random variable. The name.
1 Probability and Statistical Inference (9th Edition) Chapter 5 (Part 2/2) Distributions of Functions of Random Variables November 25, 2015.
Discrete Random Variables. Introduction In previous lectures we established a foundation of the probability theory; we applied the probability theory.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
Joint Moments and Joint Characteristic Functions.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Expectations of Continuous Random Variables. Introduction The expected value E[X] for a continuous RV is motivated from the analogous definition for a.
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Basic Random Processes. Introduction Annual summer rainfall in Rhode Island is a physical process has been ongoing for all time and will continue. We’d.
Chebyshev’s Inequality Markov’s Inequality Proposition 2.1.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
Sums of Random Variables and Long-Term Averages Sums of R.V. ‘s S n = X 1 + X X n of course.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate accuracy.
Basic statistics Usman Roshan.
Expectations of Random Variables, Functions of Random Variables
Appendix A: Probability Theory
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
Chapter 2 Minimum Variance Unbiased estimation
Monte Carlo Approximations – Introduction
Econometric Models The most basic econometric model consists of a relationship between two variables which is disturbed by a random error. We need to use.
3.1 Sums of Random Variables probability of z = x + y
9. Limit Theorems.
Further Topics on Random Variables: 1
1/2555 สมศักดิ์ ศิวดำรงพงศ์
Presentation transcript:

Probability and Moment Approximations using Limit Theorems

Introduction Until now, the procedures employed were all based on knowledge of the PMF/PDF and the implementation of its summation/integration. In many situation the PMF/PDF may be unknown or the summation/integration may not easily carried out. It would be of great utility, therefore, to be able to approximate the desired quantities using much simpler methods. For RV that are the sum of a large number of independent and identically distributed (IID) RV this can be done. We will focus on two very powerful theorems in probability The law of large numbers The central limit theorem 2

Convergence and approximation of a Sum We need to understand the role that convergence play in approximating the behavior of a sum or terms. This sum may be evaluated in closed form to allow a comparison to it approximation. The error in the approximation of s N by, is given by 3 The rate of convergence depends upon particular sequence, but it will surely converge at infinity.

Law of Large Number We argued before that if a fair coin is tossed N times in succession, then the relative frequency(r.f) of heads is close to ½. If N  ∞ then we expect the r.f. to approach ½. Let’s prove it by modeling the repeated coin toss experiment as a sequence of N Bernoulli subexperiments. The overall experimental output by the random vector We next assume that the discrete RV X i are IID with marginal PMF 4

Law of Large Number The r.f. is given by the sample mean RV If N  ∞ consider the mean and variance 5 X i ’s are independent  uncorrelated X i ’s are identically distributed  Random Variable

Law of Large Number We know that for X i ~ Ber(p), the variance is var(X i ) = p(1 - p). Since p = ½ for a fair coin, Therefore the width of the PMF of must decrease as N increases and eventually go to zero. Since the variance is we must have that as 6 Degenerate RV

Law of Large Number Note that the sum of N IID Bernoulli RV is a binomial RV, thus And To find the PMF of we let and note that can take on values Thus the PMF becomes 7

Law of Large Number A N increases takes on values more densely in the interval [0, 1]. We do not obtain a PMF with all its mass concentrated at 0.5 as we might expect. 8 Appears Gaussian although it chances in amplitude and width for each N

Law of Large Number The preceding results say that for large enough N the sample mean RV will always yield a number, which in our case is ½. In general 9

Law of Large Number Note that since is the r.f. of heads and p is the prob. of heads we have shown that The prob of a head in a single coin toss can be interpreted as the value obtained as the r.f. of heads in a large number of independent and identical coin tosses. This justifies our use of the sample mean RV as an estimator of a moment since 10

Theorem: Law of Large Number If X 1, X 2,…,X N are IID RVs with mean E X [X] and var(X) = σ 2 < ∞, then Proof. Consider the probability of the sample mean RV deviating from the expected value by more than ε, where ε is a samall positive number. This probability is given by Since we have upon using Chebyshev’s inequality and taking the limit of both sides yields 11

Theorem: Law of Large Number Since a probability must be greater than or equal to zero, we have Which is a statement that the sample mean RV converge in probability to the expected value of a single RV. Convergence is probability does not mean all realizations will converge. 12

Central Limit Theorem(CLT): PDF of sum of uniformly distributed RV By law of large numbers the PMF/PDF of the sample mean RV decreases in width until all the probability is concentrates about the mean. Let’s consider a sum of IID RV when X i ~ U(-1/2, 1/2). If N = 2, then S 2 = X 1 + X 2 and the PDF of S 2 is easily found using a convolution integral. 13

CLT: PDF of sum of uniformly distributed RV More generally, we have To find p S3 (x) we must convolve p S2 (x) with p X (x) to yield p X (x) * p X (x) * p X (x) since p S2 (x) = p X (x) * p X (x). This is 14 contributes to the integral

CLT: PDF of sum of uniformly distributed RV Since p X (-x) = p X (x),we can express this is more convenient form as The integrand may be determined by plotting p S2 (x) and the right- shifted version p X (u - x) and multiplying these to function 15

CLT: PDF of sum of uniformly distributed RV We have chosen the mean and variance of the Gaussian approximation to match that of p S3 ( x ). var(X) = (b - a) 2 /12 for X i ~ U(-1/2, 1/2) and hence var ( X i ) = 1/12. 16

CLT: PDF of sum of uniformly distributed RV In case of a uniformly distributed RV over the interval (0,1) a repeated convolution produces 17 As N increases the PDF moves to the right since E[S N ] = NE X [X] = N/2 and variance also increases since var(S N ) = Nvar(X) = N/12 Because of this it is not possible to state that the PDF converges to any PDF.

Central Limit Theorem Normalization solves the problem We can assert that this standardized RV will converge to a N(0,1) 18 Central limit theorem states that the PDF of the standardized sum of a large number of continuous IID RV will converge to a Gaussian PDF.

Central Limit Theorem: applications There is no need to know the PDF of each RV or even if it is known, to determine the exact PDF of the sum, which may not be possible. Some applications are Polling Noise characterization Scattering effects modeling Kinetic theory of gases Economic modeling. 19

Central Limit Theorem Central limit theorem for continuous RVs If X 1, X 2,…,X N are continuous IID RVs, each with mean E X [X] and variance var( X ), and then as N  ∞. Equivalently, the CDF of the standardized sum converges to Φ(x) or 20

CLT: PDF of sum of squares of independent N(0,1) RV Let X i ~ N(0, 1) for i = 1,2,…,N and assume the X i ’s are independent. We wish to determine the approximate PDF of To apply the CLT we note that since the Xi’s are IID so are the X i 2 ’s. Then as N  ∞ we have But X 2 ~ so that E X [X 2 ] = 1 and var(X 2 ) = 2 and therefore 21

CLT: PDF of sum of squares of independent N(0,1) RV 22 N = 10N = 20