The asymmetric uncertainties on data points in Roofit

Slides:



Advertisements
Similar presentations
DO’S AND DONT’S WITH LIKELIHOODS Louis Lyons Oxford (CDF)
Advertisements

Point Estimation Notes of STAT 6205 by Dr. Fan.
27 th March CERN Higgs searches: CL s W. J. Murray RAL.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
I’ll not discuss about Bayesian or Frequentist foundation (discussed by previous speakers) I’ll try to discuss some facts and my point of view. I think.
Chap 9: Testing Hypotheses & Assessing Goodness of Fit Section 9.1: INTRODUCTION In section 8.2, we fitted a Poisson dist’n to counts. This chapter will.
Probability theory Much inspired by the presentation of Kren and Samuelsson.
Basics of Statistical Estimation. Learning Probabilities: Classical Approach Simplest case: Flipping a thumbtack tails heads True probability  is unknown.
. PGM: Tirgul 10 Parameter Learning and Priors. 2 Why learning? Knowledge acquisition bottleneck u Knowledge acquisition is an expensive process u Often.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
Additional Slides on Bayesian Statistics for STA 101 Prof. Jerry Reiter Fall 2008.
Machine Learning Queens College Lecture 3: Probability and Statistics.
The paired sample experiment The paired t test. Frequently one is interested in comparing the effects of two treatments (drugs, etc…) on a response variable.
1 Probability and Statistics  What is probability?  What is statistics?
Statistical Decision Theory
Overview course in Statistics (usually given in 26h, but now in 2h)  introduction of basic concepts of probability  concepts of parameter estimation.
Statistics for Engineer Week II and Week III: Random Variables and Probability Distribution.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
Why do Wouter (and ATLAS) put asymmetric errors on data points ? What is involved in the CLs exclusion method and what do the colours/lines mean ? ATLAS.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Likelihood Methods in Ecology November 16 th – 20 th, 2009 Millbrook, NY Instructors: Charles Canham and María Uriarte Teaching Assistant Liza Comita.
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
Likelihood function and Bayes Theorem In simplest case P(B|A) = P(A|B) P(B)/P(A) and we consider the likelihood function in which we view the conditional.
Bayesian statistics Probabilities for everything.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
Week 41 Estimation – Posterior mean An alternative estimate to the posterior mode is the posterior mean. It is given by E(θ | s), whenever it exists. This.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Exam 2: Rules Section 2.1 Bring a cheat sheet. One page 2 sides. Bring a calculator. Bring your book to use the tables in the back.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
Sampling and estimation Petter Mostad
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #24.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Lecture 4 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Conditional Observables Joe Tuggle BaBar ROOT/RooFit Workshop December 2007.
Ivo van Vulpen Why does RooFit put asymmetric errors on data points ? 10 slides on a ‘too easy’ topic that I hope confuse, but do not irritate you
Bayesian Estimation and Confidence Intervals Lecture XXII.
Lecture Slides Elementary Statistics Twelfth Edition
(Day 3).
The expected confident intervals for triple gauge coupling parameter
Bayesian Estimation and Confidence Intervals
MCMC Stopping and Variance Estimation: Idea here is to first use multiple Chains from different initial conditions to determine a burn-in period so the.
STATISTICAL INFERENCE
Bayesian approach to the binomial distribution with a discrete prior
Statistical methods in LHC data analysis introduction
Bayesian data analysis
Bayes Net Learning: Bayesian Approaches
Lecture 4 1 Probability (90 min.)
Maximum Likelihood Find the parameters of a model that best fit the data… Forms the foundation of Bayesian inference Slide 1.
Discrete Event Simulation - 4
10701 / Machine Learning Today: - Cross validation,
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Sampling Distributions (§ )
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
CS639: Data Management for Data Science
Why do Wouter (and ATLAS) put asymmetric errors on data points ?
Computing and Statistical Data Analysis / Stat 10
1/2555 สมศักดิ์ ศิวดำรงพงศ์
Applied Statistics and Probability for Engineers
Introductory Statistics
Presentation transcript:

The asymmetric uncertainties on data points in Roofit Few slides on ‘easy/trivial’ topic that will hopefully leave you confused di-muon invariant mass [MeV/c2] Candidates (50 MeV/c2) CERN seminar Flavio Archilli 14-2-2017 Ivo van Vulpen (Nikhef/UvA, ATLAS)

How are our asymmetric uncertainties on data points defined ? Not ± √n 4 Candidates (50 MeV/c2) di-muon invariant mass [MeV/c2] 2/20

statistics 3/20

You are responsible for how you summarize your measurement Solving statistics problems in general Discussions (in large experiments): - strong opinions, very outspoken ‘experts’ - use RooSomethingFancy: made by experts & debugged - let’s do what we did last time, let’s be conservative - … You are responsible for how you summarize your measurement The tools you use have assumptions, biases, pitfalls, … , so you know best how others (and you) should interpret your measurement 4/20

Six reasonable options Discuss in your group which one you prefer How should we represent this measurement ? 4 Candidates (50 MeV/c2) di-muon invariant mass [MeV/c2] Basis for this talk: http://www.nikhef.nl/~ivov/Statistics/PoissonError/BobCousins_Poisson.pdf 5/20

Option 1: assign NO uncertainty ± 0.00 The number of observed events is what it is: 4 The uncertainty is in the interpretation step, i.e. on the model-parameters that you extract from it 6/20

Comfortable territory: Poisson distribution Probability to observe n events when λ are expected Binomial with n∞, p0 en np=λ Number of observed events #observed λ hypothesis P(Nobs|λ=4.9) λ=4.90 4 7/20

Properties of the Poisson distribution the famous √n properties λ=1.00 (1) Mean: (2) Variance: (3) Most likely: first integer ≤ λ λ=4.90 λ=5.00 8/20

Option 2: the famous sqrt(n) +2.00 -2.00 Poisson variance for λ equal to measured number of events … but Poisson distribution is asymetric: 9/20

Just treat it like a normal measurement What you have: 1) construct the Likelihood λ as free parameter 2) Find value of λ that maximizes the Likelihood 3) Determine error interval: Δ(-2Log(Lik.)) = +1 10/20

Likelihood Likelihood Poisson: probability to observe n events when λ are expected Likelihood P(Nobs=4|λ) #observed λ hypothesis λ (hypothesis) fixed scan 11/20

Option 3: Likelihood 12/20 P(Nobs=4|λ) Log(Likelihood) +2.35 P(Nobs=4|λ) -1.68 λ (hypothesis) -2Log(L) Log(Likelihood) -1.68 +2.35 -2Log(P(Nobs=4|λ)) ΔL=+1 12/20 2.32 4.00 6.35

Bayesian: statement on value of λ What you have: What you want: Likelihood pdf for λ Probability to observe Nobs events … given a specific hypothesis (λ) What can we say about theory (λ) … given # of observed events (4) 13/20

Bayesian: statement on value of λ Choice of prior P(λ): (basis for religious wars) Here: assume all values for λ equally likely Likelihood Probability density function λ (hypothesis) λ (hypothesis) Posterior PDF for λ  Integrate to get confidence interval

Option 4 & 5: representing the Bayesian posterior Bayes central +3.16 68% 16% 16% -1.16 Bayes equal probability +2.40 68% 8.3% 23.5% -1.71 15/20

Option 6: Frequentist approach Find values of λ that are on border of being compatible with observed #events λ=7.16 If λ > 7.16 then probability to observe 4 events (or less) <16% 16% Note: also uses ‘data you didn’t observe’, i.e. a bit like definition of significance smallest λ (>n) for which P(n≤nobs|λ) ≤ 0.159 +3.16 λ=2.09 -1.91 largest λ (<n) for which P(n≥nobs|λ) ≤ 0.159 16% 16/20

The six options 8 7 -1.91 +3.16 6 +2.00 +2.35 +3.16 +2.40 5 4 3 -2.00 -1.68 -1.16 -1.71 2 1 Nothing Poisson Likelihood Bayesian central Bayesian equal prob Frequentist Δλ: 0.00 4.00 4.03 4.32 4.11 5.07 17/20

The six options Δλ: 0.00 4.00 4.03 4.32 4.11 5.07 RooFit default 18/20 -1.91 +3.16 6 +2.00 +2.35 +3.16 +2.40 5 4 3 -2.00 -1.68 -1.16 -1.71 2 1 Nothing Poisson Likelihood Bayesian central Bayesian equal prob Frequentist Δλ: 0.00 4.00 4.03 4.32 4.11 5.07 18/20

Discussions in other experiment: Example  discussion in CDF: https://www-cdf.fnal.gov/physics/statistics/notes/pois_eb.txt We feel it is important to have a relatively simple rule that is readily understood by readers. A reader does not want to have to work hard simply to understand what an error bar on a plot represents. <…> Since the use of +-sqrt(n) is so widespread, the argument in favour of an alternative should be convincing in order for it to be adopted. 19/20

Summary di-muon invariant mass [MeV/c2] Candidates (50 MeV/c2) +3.16 4 -1.91 - You now know how Roofit ‘Poisson errors’ are defined Note: choice has no impact on likelihood fits - Do you agree with RooFit default? What about empty bins then? - Perfect topic to confuse and irritate people over coffee  do it!