Confidence Intervals Lecture 2 First ICFA Instrumentation School/Workshop At Morelia, Mexico, November 18-29, 2002 Harrison B. Prosper Florida State University.

Slides:



Advertisements
Similar presentations
Estimation, Variation and Uncertainty Simon French
Advertisements

Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Practical Statistics for LHC Physicists Bayesian Inference Harrison B. Prosper Florida State University CERN Academic Training Lectures 9 April, 2015.
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
A Brief Introduction to Bayesian Inference Robert Van Dine 1.
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
What is Statistical Modeling
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Statistical Tools PhyStat Workshop 2004 Harrison B. Prosper1 Statistical Tools A Few Comments Harrison B. Prosper Florida State University PHYSTAT Workshop.
A/Prof Geraint Lewis A/Prof Peter Tuthill
Evaluating Hypotheses
Credibility of confidence intervals Dean Karlen / Carleton University Advanced Statistical Techniques in Particle Physics Durham, March 2002.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Introduction to Bayesian statistics Three approaches to Probability  Axiomatic Probability by definition and properties  Relative Frequency Repeated.
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Standard error of estimate & Confidence interval.
Lecture 9: p-value functions and intro to Bayesian thinking Matthew Fox Advanced Epidemiology.
Statistical Analysis of Systematic Errors and Small Signals Reinhard Schwienhorst University of Minnesota 10/26/99.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
Machine Learning Queens College Lecture 3: Probability and Statistics.
1 Probability and Statistics  What is probability?  What is statistics?
Statistical Decision Theory
Harrison B. Prosper Workshop on Top Physics, Grenoble Bayesian Statistics in Analysis Harrison B. Prosper Florida State University Workshop on Top Physics:
Dr. Gary Blau, Sean HanMonday, Aug 13, 2007 Statistical Design of Experiments SECTION I Probability Theory Review.
Bayesian Methods I: Parameter Estimation “A statistician is a person who draws a mathematically precise line from an unwarranted assumption to a foregone.
Week 71 Hypothesis Testing Suppose that we want to assess the evidence in the observed data, concerning the hypothesis. There are two approaches to assessing.
Practical Statistics for Particle Physicists Lecture 3 Harrison B. Prosper Florida State University European School of High-Energy Physics Anjou, France.
LECTURE 25 THURSDAY, 19 NOVEMBER STA291 Fall
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Making sense of randomness
Confidence Intervals First ICFA Instrumentation School/Workshop At Morelia, Mexico, November 18-29, 2002 Harrison B. Prosper Florida State University.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Cosmological Model Selection David Parkinson (with Andrew Liddle & Pia Mukherjee)
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Sampling and estimation Petter Mostad
Experience from Searches at the Tevatron Harrison B. Prosper Florida State University 18 January, 2011 PHYSTAT 2011 CERN.
Practical Statistics for Particle Physicists Lecture 2 Harrison B. Prosper Florida State University European School of High-Energy Physics Parádfürdő,
V7 Foundations of Probability Theory „Probability“ : degree of confidence that an event of an uncertain nature will occur. „Events“ : we will assume that.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #24.
1 27 March 2000CL Workshop, Fermilab, Harrison B. Prosper The Reverend Bayes and Solar Neutrinos Harrison B. Prosper Florida State University 27 March,
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Bayesian Estimation and Confidence Intervals Lecture XXII.
Univariate Gaussian Case (Cont.)
The asymmetric uncertainties on data points in Roofit
Bayesian Estimation and Confidence Intervals
ECO 173 Chapter 10: Introduction to Estimation Lecture 5a
Bayesian Generalized Product Partition Model
Multivariate Analysis Past, Present and Future
The Nathan Kline Institute, NY
Bayes Net Learning: Bayesian Approaches
Data Mining Lecture 11.
ECO 173 Chapter 10: Introduction to Estimation Lecture 5a
Incorporating systematic uncertainties into upper limits
LECTURE 05: THRESHOLD DECODING
Chi Square Distribution (c2) and Least Squares Fitting
Discrete Event Simulation - 4
Statistical NLP: Lecture 4
Counting Statistics and Error Prediction
Parametric Methods Berlin Chen, 2005 References:
Inference on the Mean of a Population -Variance Known
LECTURE 05: THRESHOLD DECODING
Sampling Distributions (§ )
CS639: Data Management for Data Science
Mathematical Foundations of BME Reza Shadmehr
Applied Statistics and Probability for Engineers
Presentation transcript:

Confidence Intervals Lecture 2 First ICFA Instrumentation School/Workshop At Morelia, Mexico, November 18-29, 2002 Harrison B. Prosper Florida State University

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper2 Recap of Lecture 1 To interpret a confidence level as a relative frequency requires the concept of a set of ensembles of experiments. To interpret a confidence level as a relative frequency requires the concept of a set of ensembles of experiments. Each ensemble has a coverage probability, which is simply the fraction of experiments in the ensemble with intervals that contain the value of the parameter pertaining to that ensemble. Each ensemble has a coverage probability, which is simply the fraction of experiments in the ensemble with intervals that contain the value of the parameter pertaining to that ensemble. The confidence level is the minimum coverage probability over the set of ensembles. The confidence level is the minimum coverage probability over the set of ensembles.

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper3 Count Parameter space Confidence Interval – General Algorithm Probability to be here is ≥ β Probability to be here is ≥ β )(Nl 

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper4 Coverage Probabilities θ Central Feldman-Cousins Mode-Centered N±√N

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper5 What’s Wrong With Ensembles? Nothing, if they are objectively real, such as: Nothing, if they are objectively real, such as: –The people in Morelia between the ages of 16 and 26 –Daily temperature data in Morelia during the last decade But the ensembles used in data analysis typically are not But the ensembles used in data analysis typically are not frequency interpretation repetition –There was only a single instance of Run I of DØ and CDF. But to determine confidence levels with a frequency interpretation we must embed the two experiments into ensembles, that is, we must decide what constitutes a repetition of the experiments. The problem is that reasonable people sometimes disagree about the choice of ensembles, but, because the ensembles are not real, there is generally no simple way to resolve disagreements. The problem is that reasonable people sometimes disagree about the choice of ensembles, but, because the ensembles are not real, there is generally no simple way to resolve disagreements.

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper Outline Deductive Reasoning Deductive Reasoning Inductive Reasoning Inductive Reasoning –Probability –Bayes’ Theorem Example Example Summary Summary

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper propositionsA Consider the propositions: A = (this is a baby) B B = (she cries a lot) Major premiseATRUEBTRUE Major premise: If A is TRUE, then B is TRUE Minor premiseATRUE Minor premise: A is TRUE Conclusion: BTRUE Conclusion: Therefore, B is TRUE Major premiseATRUEBTRUE Major premise: If A is TRUE, then B is TRUE Minor premiseBFALSE Minor premise: B is FALSE Conclusion: AFALSE Conclusion: Therefore, A is FALSE Aristotle, ~ 350 BC AB = A Deductive Reasoning

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper A A = (this is a baby) B B = (she cries a lot) Major premiseATRUEBTRUE Major premise: If A is TRUE, then B is TRUE Minor premiseAFALSE Minor premise: A is FALSE Conclusion: B? Conclusion: Therefore, B is ? Major premiseATRUEBTRUE Major premise: If A is TRUE, then B is TRUE Minor premiseBTRUE Minor premise: B is TRUE Conclusion: A? Conclusion: Therefore, A is ? AB = A Deductive Reasoning - ii

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper Major premiseATRUEBTRUE Major premise: If A is TRUE, then B is TRUE Minor premiseBTRUE Minor premise: B is TRUE Conclusion: Amore plausible Conclusion: Therefore, A is more plausible Major premiseATRUEBTRUE Major premise: If A is TRUE, then B is TRUE Minor premiseAFALSE Minor premise: A is FALSE Conclusion: Bless plausible Conclusion: Therefore, B is less plausible plausible Can “plausible” be made precise? Inductive Reasoning A A = (this is a baby) B B = (she cries a lot)

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper Bayes(1763), Laplace(1774), Boole(1854), Cox(1946) Jeffreys(1939), Cox(1946), Polya(1946), Jaynes(1957) Yes! In 1946, the physicist Richard Cox showed that inductive reasoning follows rules that are isomorphic to those of probability theory

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper Boolean Algebra The Rules of Inductive Reasoning: Boolean Algebra + Probability

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper12 Probability B A AB Conditional Probability A theorem

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper These rules together with Boolean algebra Bayesian Probability Theory are the foundation of Bayesian Probability Theory Product Rule Sum Rule Probability - ii Bayes’ Theorem

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper Bayes’ Theorem marginalization We can sum over propositions that are of no interest

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper15 Bayes’ Theorem: Example 1 Signal/Background Discrimination Signal/Background Discrimination –S = Signal –B = Background The probability P(S|Data), of an event being a signal given some event Data, can be approximated in several ways, for example, with a feed-forward neural network The probability P(S|Data), of an event being a signal given some event Data, can be approximated in several ways, for example, with a feed-forward neural network

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper posteriorprior marginalization likelihood I is the prior information Bayes’ Theorem: Continuous Propositions

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper17 Bayes’ Theorem: Model Comparison Standard Model SUSY Models M For each model M, we integrate over the parameters of the theory. M P(M|x,I) is the probability of M model M given data x.

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper Variance BayesianConfidenceInterval where Bayesian measures of uncertainty Also known as a Credible Interval

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper19 Example 1: Counting Experiment Experiment: Experiment: θ –To measure the mean rate θ of UHECRs above eV per unit solid angle. Assume the probability of N events to be a given by a Poisson distribution Assume the probability of N events to be a given by a Poisson distribution

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper20 Example 1: Counting Experiment - ii Apply Bayes’ Theorem Apply Bayes’ Theorem What should we take for the prior probability P(θ|I)? What should we take for the prior probability P(θ|I)? How about How about –P(θ|I) = dθ?

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper21 But why choose P(θ|I) = dθ? Why not P(θ 2 |I) = dθ? Why not P(θ 2 |I) = dθ? Or P(tan(θ)|I) = dθ? Or P(tan(θ)|I) = dθ? Choosing a prior probability in the absence of useful prior information is a difficult and controversial problem. Choosing a prior probability in the absence of useful prior information is a difficult and controversial problem. –The difficulty is to determine the variable in terms of which the prior probability density is constant. –In practice one considers a class of prior probabilities and uses it to assess the sensitivity of the results to the choice of prior.

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper22 Credible Intervals With P(θ|I) = dθ θ Bayesian N±√N

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper23 Credible Interval Widths Bayesian N±√N

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper24 Coverage Probabilities: Credible Intervals θ Bayesian Even though these are Bayesian intervals nothing prevents us from studying their frequency behavior with respect to some ensemble!

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper Model Data Likelihood l is the efficiency times branching fraction times integrated luminosity Prior information is the mean background count estimates are estimates Bayes’ Theorem: Measuring a Cross-section

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper Apply Bayes’ Theorem: nuisance parameters Then marginalize: that is, integrate over nuisance parameters priorlikelihood posterior Measuring a Cross-section - ii

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper27 What Prior? We can usually write down something sensible for the luminosity and background priors. But, again, what to write for the cross-section prior is controversial. In DØ and CDF, P(σ|I) = d σ as a matter of convention, we set P(σ|I) = d σ Rule: Always report the prior you have used!

November 21, 2002First ICFA Instrumentation School/Workshop Harrison B. Prosper28 Summary Confidence levels are probabilities; as such, their interpretation depends on the interpretation of probability adopted Confidence levels are probabilities; as such, their interpretation depends on the interpretation of probability adopted If one adopts a frequency interpretation the confidence level is tied to the set of ensembles one uses and is a property of that set. Typically, however, these ensembles do not objectively exist. If one adopts a frequency interpretation the confidence level is tied to the set of ensembles one uses and is a property of that set. Typically, however, these ensembles do not objectively exist. If one adopts a degree of belief interpretation the confidence level is a property of the interval calculated. A set of ensembles is not required for interpretation. However, the results necessarily depend on the choice of prior. If one adopts a degree of belief interpretation the confidence level is a property of the interval calculated. A set of ensembles is not required for interpretation. However, the results necessarily depend on the choice of prior.