Presentation is loading. Please wait.

Presentation is loading. Please wait.

Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Section 9.1: Parameter estimation CIS 2033. Computational Probability.

Similar presentations


Presentation on theme: "Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Section 9.1: Parameter estimation CIS 2033. Computational Probability."— Presentation transcript:

1 Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Section 9.1: Parameter estimation CIS Computational Probability and Statistics Pei Wang

2 Parameters of distributions
After determining the family of distribution, the next step is to estimate the parameters Example 9.1: The number of defects on each chip is believed to follow Pois(λ) Since λ = E(X) is the expectation of a Poisson variable, it can be estimated with a sample mean X-bar

3 Method of moments

4 Method of moments (2) Special cases: μ’2 is Var(X), and m’2 is s2(n-1)/n

5 Method of moments (3) To estimate k parameters, we can equate the first k population and sample moments (or centralized version), i.e. μ1 = m1 , … …, μk = mk The left-hand sides of these equations depend on the parameters, while the right-hand sides can be computed from data The method of moments finds estimators by solving the above equations

6 Moments method example
The CPU time for 30 randomly chosen tasks of a certain type are (in seconds) If they are considered to be the value of a random variable X, what is the model?

7 Moments method example (2)
The histogram suggests Gamma distribution

8 Moments method example (3)

9 Moments method example (4)
From data, we compute and write two equations, Solving this system in terms of α and λ, we get the method of moment estimates

10 Method of maximum likelihood
Maximum likelihood estimator is the parameter value that maximizes the likelihood of the observed sample, L(x1, …, xn) L(x1, …, xn) is defined as p(x1, …, xn) for a discrete distribution, and f(x1, …, xn) for a continuous distribution When the variables X1, …, Xn are independent, L(x1, …, xn) is obtained by multiplying the marginal pmfs or pdfs

11 Maximum likelihood Maximum likelihood estimator is the parameter value that maximizes the likelihood L(θ) of the observed sample, x1, …, xn When the observations are independent of each other, L(θ) = pθ(x1)*...*pθ(xn) for a discrete variable fθ(x1)*...*fθ(xn) for a continuous variable Which is a function with θ as variable

12 Maximum likelihood (2) Here we consider two types of L(θ):
If the function always increases or decreases in its non-zero range, the maximum value is at the boundary of the range, i.e., the min or max of θ If the function first increases then decreases in its non-zero range, the maximum value is at where its derivative L’(θ) is zero

13 Example of Type 1 To estimate the θ in U(0, θ) given positive data x1, …, xn, L(θ) is θ-n when θ ≥ max(x1, …, xn), otherwise it is 0 So the best estimator for θ is max(x1, …, xn) since L(θ) is a decreasing function when θ ≥ max(x1, …, xn) Similarly, if x1, …, xn are generated by U(a, b), the maximum likelihood estimate is a = min(x1, …, xn), b = max(x1, …, xn)

14 Example of Type 2 If the distribution is Ber(p), and m of the n sample values are 1, L(p) = pm(1 – p)n–m L’(p) = mpm–1(1 – p)n–m – pm(n – m)(1 – p)n–m–1 = (m – np)pm–1(1 – p)n–m–1 L’(p) is 0 when p = m/n, which also covers the situation where p is 0 or 1 So the sample mean is a maximum likelihood estimator of p in Ber(p)

15 Exercise If a probability mass function is partially known, how to guess the missing values using sample data? Take the following die as an instance a 1 2 3 4 5 6 p(a) 0.1 0.2 ? count 12 10 19 23 9 27

16 Log-likelihood Log function turns multiplication into addition, and power into multiplication E.g. ln(f × g) = ln(f) + ln(g) ln(fg) = g × ln(f) Log-likelihood function and likelihood function reach maximum at the same value Therefore, ln(L(θ)) may be easier for getting maximum likelihood

17 Log-likelihood (2) E. g., L(p) = pm(1 – p)n–m ln(L(p)) = m(ln(p)) + (n – m)(ln(1 – p)) [ln(L(p))]’ = m/p – (n – m)/(1 – p) m/p – (n – m)/(1 – p) = 0 m/p = (n – m)/(1 – p) m – mp = np – mp p = m/n

18 Comparing estimators A good estimator should have lower bias and variant, but how to balance these two factors?

19 Mean squared error When both the bias and variance of estimators can be obtained, usually we prefer the one that has the smallest mean squared error (MSE) For estimator T of parameter θ, MSE(T) = E[(T − θ)2] = E[T2] −2θE[T] + θ2 = Var(T) + (E[T] − θ)2 = Var(T) + Bias(T)2 So, MSE summarizes variance and bias

20 MSE example Let T1 and T2 be two unbiased estimators for the same parameter θ based on a sample of size n, and it is known that Var(T1) = (θ + 1)(θ − n) / (3n) Var(T2) = (θ + 1)(θ − n) / [(n + 2)n] Since n + 2 > 3 when n > 1, MSE(T1) > MSE(T2) , so T2 is a better estimator for all values of θ

21 MSE example (2) Let T1 and T2 be two estimators for the same parameter, and it is known that Var(T1) = 5/n2, Bias(T1) = -2/n Var(T2) = 1/n2, Bias(T2) = 3/n < 1 + 9, MSE(T1) < MSE(T2) , so T1 is a better estimator for the parameter


Download ppt "Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Section 9.1: Parameter estimation CIS 2033. Computational Probability."

Similar presentations


Ads by Google