G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.

Slides:



Advertisements
Similar presentations
Inference for Regression
Advertisements

CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Chap 9: Testing Hypotheses & Assessing Goodness of Fit Section 9.1: INTRODUCTION In section 8.2, we fitted a Poisson dist’n to counts. This chapter will.
Chapter 7: Statistical Applications in Traffic Engineering
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Statistical Data Analysis Stat 3: p-values, parameter estimation
G. Cowan RHUL Physics Profile likelihood for systematic uncertainties page 1 Use of profile likelihood to determine systematic uncertainties ATLAS Top.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Business Statistics - QBM117
G. Cowan Statistics for HEP / NIKHEF, December 2011 / Lecture 3 1 Statistical Methods for Particle Physics Lecture 3: Limits for Poisson mean: Bayesian.
1 Nuisance parameters and systematic uncertainties Glen Cowan Royal Holloway, University of London IoP Half.
G. Cowan RHUL Physics Comment on use of LR for limits page 1 Comment on definition of likelihood ratio for limits ATLAS Statistics Forum CERN, 2 September,
Statistics for HEP / LAL Orsay, 3-5 January 2012 / Lecture 1
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
G. Cowan 2011 CERN Summer Student Lectures on Statistics / Lecture 41 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability.
G. Cowan RHUL Physics Higgs combination note status page 1 Status of Higgs Combination Note ATLAS Statistics/Higgs Meeting Phone, 7 April, 2008 Glen Cowan.
Inference about a Mean Part II
Chapter 2 Simple Comparative Experiments
G. Cowan Statistical Data Analysis / Stat 4 1 Statistical Data Analysis Stat 4: confidence intervals, limits, discovery London Postgraduate Lectures on.
G. Cowan Lectures on Statistical Data Analysis Lecture 14 page 1 Statistical Data Analysis: Lecture 14 1Probability, Bayes’ theorem 2Random variables and.
Inferences About Process Quality
Chapter 9 Hypothesis Testing.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
BCOR 1020 Business Statistics
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
G. Cowan Discovery and limits / DESY, 4-7 October 2011 / Lecture 3 1 Statistical Methods for Discovery and Limits Lecture 3: Limits for Poisson mean: Bayesian.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 9 Hypothesis Testing.
8 - 1 © 2003 Pearson Prentice Hall Chi-Square (  2 ) Test of Variance.
Hypothesis Testing in Linear Regression Analysis
G. Cowan 2009 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability densities,
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
Topics: Statistics & Experimental Design The Human Visual System Color Science Light Sources: Radiometry/Photometry Geometric Optics Tone-transfer Function.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 1 page 1 Lectures on Statistical Data Analysis London Postgraduate Lectures on Particle Physics;
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005 Dr. John Lipp Copyright © Dr. John Lipp.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
Hypothesis Testing Errors. Hypothesis Testing Suppose we believe the average systolic blood pressure of healthy adults is normally distributed with mean.
G. Cowan RHUL Physics LR test to determine number of parameters page 1 Likelihood ratio test to determine best number of parameters ATLAS Statistics Forum.
G. Cowan Terascale Statistics School 2015 / Combination1 Statistical combinations, etc. Terascale Statistics School DESY, Hamburg March 23-27, 2015 Glen.
© Copyright McGraw-Hill 2004
Review of Statistics.  Estimation of the Population Mean  Hypothesis Testing  Confidence Intervals  Comparing Means from Different Populations  Scatterplots.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
ES 07 These slides can be found at optimized for Windows)
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Lecture 4 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Statistical Significance Hypothesis Testing.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Hypothesis Testing. Suppose we believe the average systolic blood pressure of healthy adults is normally distributed with mean μ = 120 and variance σ.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
LECTURE 33: STATISTICAL SIGNIFICANCE AND CONFIDENCE (CONT.)
Statistics for the LHC Lecture 3: Setting limits
Chapter 2 Simple Comparative Experiments
Lecture 4 1 Probability (90 min.)
Graduierten-Kolleg RWTH Aachen February 2014 Glen Cowan
Computing and Statistical Data Analysis / Stat 8
Lecture 3 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
LESSON 18: CONFIDENCE INTERVAL ESTIMATION
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Statistical Tests and Limits Lecture 2: Limits
Computing and Statistical Data Analysis / Stat 7
Introduction to Statistics − Day 4
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Computing and Statistical Data Analysis / Stat 10
Presentation transcript:

G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway, University of London Course web page:

G. Cowan Computing and Statistical Data Analysis / Stat 9 2 Example of least squares fit Fit a polynomial of order p:

G. Cowan Computing and Statistical Data Analysis / Stat 9 3 Variance of LS estimators In most cases of interest we obtain the variance in a manner similar to ML. E.g. for data ~ Gaussian we have and so or for the graphical method we take the values of  where 1.0

G. Cowan Computing and Statistical Data Analysis / Stat 9 4 Two-parameter LS fit

G. Cowan Computing and Statistical Data Analysis / Stat 9 5 Goodness-of-fit with least squares The value of the  2 at its minimum is a measure of the level of agreement between the data and fitted curve: It can therefore be employed as a goodness-of-fit statistic to test the hypothesized functional form  (x;  ). We can show that if the hypothesis is correct, then the statistic t =  2 min follows the chi-square pdf, where the number of degrees of freedom is n d = number of data points - number of fitted parameters

G. Cowan Computing and Statistical Data Analysis / Stat 9 6 Goodness-of-fit with least squares (2) The chi-square pdf has an expectation value equal to the number of degrees of freedom, so if  2 min ≈ n d the fit is ‘good’. More generally, find the p-value: E.g. for the previous example with 1st order polynomial (line), whereas for the 0th order polynomial (horizontal line), This is the probability of obtaining a  2 min as high as the one we got, or higher, if the hypothesis is correct.

G. Cowan Computing and Statistical Data Analysis / Stat 9 7 Goodness-of-fit vs. statistical errors

G. Cowan Computing and Statistical Data Analysis / Stat 9 8 Goodness-of-fit vs. stat. errors (2)

G. Cowan Computing and Statistical Data Analysis / Stat 9 9 LS with binned data

G. Cowan Computing and Statistical Data Analysis / Stat 9 10 LS with binned data (2)

G. Cowan Computing and Statistical Data Analysis / Stat 9 11 LS with binned data — normalization

G. Cowan Computing and Statistical Data Analysis / Stat 9 12 LS normalization example

G. Cowan Computing and Statistical Data Analysis / Stat 9 13 Using LS to combine measurements

G. Cowan Computing and Statistical Data Analysis / Stat 9 14 Combining correlated measurements with LS

G. Cowan Computing and Statistical Data Analysis / Stat 9 15 Example: averaging two correlated measurements

G. Cowan Computing and Statistical Data Analysis / Stat 9 16 Negative weights in LS average

G. Cowan Computing and Statistical Data Analysis / Stat 9 17 Interval estimation — introduction Often use +/  the estimated standard deviation of the estimator. In some cases, however, this is not adequate: estimate near a physical boundary, e.g., an observed event rate consistent with zero. In addition to a ‘point estimate’ of a parameter we should report an interval reflecting its statistical uncertainty. Desirable properties of such an interval may include: communicate objectively the result of the experiment; have a given probability of containing the true parameter; provide information needed to draw conclusions about the parameter possibly incorporating stated prior beliefs. We will look briefly at Frequentist and Bayesian intervals.

G. Cowan Computing and Statistical Data Analysis / Stat 9 18 Frequentist confidence intervals Consider an estimator for a parameter  and an estimate We also need for all possible  its sampling distribution Specify upper and lower tail probabilities, e.g.,  = 0.05,  = 0.05, then find functions u  (  ) and v  (  ) such that:

G. Cowan Computing and Statistical Data Analysis / Stat 9 19 Confidence interval from the confidence belt Find points where observed estimate intersects the confidence belt. The region between u  (  ) and v  (  ) is called the confidence belt. This gives the confidence interval [a, b] Confidence level = 1  = probability for the interval to cover true value of the parameter (holds for any possible true  ).

G. Cowan Computing and Statistical Data Analysis / Stat 9 20 Confidence intervals by inverting a test Confidence intervals for a parameter  can be found by defining a test of the hypothesized value  (do this for all  ): Specify values of the data that are ‘disfavoured’ by  (critical region) such that P(data in critical region) ≤  for a prespecified , e.g., 0.05 or 0.1. If data observed in the critical region, reject the value . Now invert the test to define a confidence interval as: set of  values that would not be rejected in a test of size  (confidence level is 1  ). The interval will cover the true value of  with probability ≥ 1 . Equivalent to confidence belt construction; confidence belt is acceptance region of a test.

Computing and Statistical Data Analysis / Stat 9 21 Relation between confidence interval and p-value Equivalently we can consider a significance test for each hypothesized value of , resulting in a p-value, p .. If p  < , then we reject . The confidence interval at CL = 1 –  consists of those values of  that are not rejected. E.g. an upper limit on  is the greatest value for which p  ≥  In practice find by setting p  =  and solve for . G. Cowan

Computing and Statistical Data Analysis / Stat 9 22 Confidence intervals in practice The recipe to find the interval [a, b] boils down to solving → a is hypothetical value of  such that → b is hypothetical value of  such that

G. Cowan Computing and Statistical Data Analysis / Stat 9 23 Meaning of a confidence interval

G. Cowan Computing and Statistical Data Analysis / Stat 9 24 Central vs. one-sided confidence intervals

G. Cowan Computing and Statistical Data Analysis / Stat 9 25 Intervals from the likelihood function In the large sample limit it can be shown for ML estimators: defines a hyper-ellipsoidal confidence region, If then (n-dimensional Gaussian, covariance V)

G. Cowan Computing and Statistical Data Analysis / Stat 9 26 Approximate confidence regions from L(  ) So the recipe to find the confidence region with CL = 1  is: For finite samples, these are approximate confidence regions. Coverage probability not guaranteed to be equal to  ; no simple theorem to say by how far off it will be (use MC). Remember here the interval is random, not the parameter.

G. Cowan Computing and Statistical Data Analysis / Stat 9 27 Example of interval from ln L(  ) For n=1 parameter, CL = 0.683, Q  = 1.