Presentation is loading. Please wait.

Presentation is loading. Please wait.

G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.

Similar presentations


Presentation on theme: "G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions."— Presentation transcript:

1 G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of r.v.s, expectation values, error propagation 3Catalogue of pdfs 4The Monte Carlo method 5Statistical tests: general concepts 6Test statistics, multivariate methods 7Significance tests 8Parameter estimation, maximum likelihood 9More maximum likelihood 10Method of least squares 11Interval estimation, setting limits 12Nuisance parameters, systematic uncertainties 13Examples of Bayesian approach 14tba

2 G. Cowan Lectures on Statistical Data Analysis 2 The method of least squares Suppose we measure N values, y 1,..., y N, assumed to be independent Gaussian r.v.s with Assume known values of the control variable x 1,..., x N and known variances The likelihood function is We want to estimate , i.e., fit the curve to the data points.

3 G. Cowan Lectures on Statistical Data Analysis 3 The method of least squares (2) The log-likelihood function is therefore So maximizing the likelihood is equivalent to minimizing Minimum defines the least squares (LS) estimator Very often measurement errors are ~Gaussian and so ML and LS are essentially the same. Often minimize  2 numerically (e.g. program MINUIT ).

4 G. Cowan Lectures on Statistical Data Analysis 4 LS with correlated measurements If the y i follow a multivariate Gaussian, covariance matrix V, Then maximizing the likelihood is equivalent to minimizing

5 G. Cowan Lectures on Statistical Data Analysis 5 Example of least squares fit Fit a polynomial of order p:

6 G. Cowan Lectures on Statistical Data Analysis 6 Variance of LS estimators In most cases of interest we obtain the variance in a manner similar to ML. E.g. for data ~ Gaussian we have and so or for the graphical method we take the values of  where 1.0

7 G. Cowan Lectures on Statistical Data Analysis 7 Two-parameter LS fit

8 G. Cowan Lectures on Statistical Data Analysis 8 Goodness-of-fit with least squares The value of the  2 at its minimum is a measure of the level of agreement between the data and fitted curve: It can therefore be employed as a goodness-of-fit statistic to test the hypothesized functional form  (x;  ). We can show that if the hypothesis is correct, then the statistic t =  2 min follows the chi-square pdf, where the number of degrees of freedom is n d = number of data points - number of fitted parameters

9 G. Cowan Lectures on Statistical Data Analysis 9 Goodness-of-fit with least squares (2) The chi-square pdf has an expectation value equal to the number of degrees of freedom, so if  2 min ≈ n d the fit is ‘good’. More generally, find the p-value: E.g. for the previous example with 1st order polynomial (line), whereas for the 0th order polynomial (horizontal line), This is the probability of obtaining a  2 min as high as the one we got, or higher, if the hypothesis is correct.

10 G. Cowan Lectures on Statistical Data Analysis 10 Goodness-of-fit vs. statistical errors

11 G. Cowan Lectures on Statistical Data Analysis 11 Goodness-of-fit vs. stat. errors (2)

12 G. Cowan Lectures on Statistical Data Analysis 12 LS with binned data

13 G. Cowan Lectures on Statistical Data Analysis 13 LS with binned data (2)

14 G. Cowan Lectures on Statistical Data Analysis 14 LS with binned data — normalization

15 G. Cowan Lectures on Statistical Data Analysis 15 LS normalization example

16 G. Cowan Lectures on Statistical Data Analysis 16 Using LS to combine measurements

17 G. Cowan Lectures on Statistical Data Analysis 17 Combining correlated measurements with LS

18 G. Cowan Lectures on Statistical Data Analysis 18 Example: averaging two correlated measurements

19 G. Cowan Lectures on Statistical Data Analysis 19 Negative weights in LS average

20 G. Cowan Lectures on Statistical Data Analysis 20 Wrapping up lecture 10 Considering ML with Gaussian data led to the method of Least Squares. Several caveats when the data are not (quite) Gaussian, e.g., histogram-based data. Goodness-of-fit with LS “easy” (but do not confuse good fit with small stat. errors) LS can be used for averaging measurements. Next lecture: Interval estimation


Download ppt "G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions."

Similar presentations


Ads by Google