G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 31 Statistical Tests and Limits Lecture 3 IN2P3 School of Statistics Autrans, France 17—21.

Slides:



Advertisements
Similar presentations
1 Statistical Methods in HEP, in particular... Nuisance parameters and systematic uncertainties Glen Cowan Royal Holloway, University of London
Advertisements

Using the Profile Likelihood in Searches for New Physics / PHYSTAT 2011 G. Cowan 1 Using the Profile Likelihood in Searches for New Physics arXiv:
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
G. Cowan Statistics for HEP / LAL Orsay, 3-5 January 2012 / Lecture 2 1 Statistical Methods for Particle Physics Lecture 2: Tests based on likelihood ratios.
Statistics In HEP 2 Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
G. Cowan RHUL Physics Profile likelihood for systematic uncertainties page 1 Use of profile likelihood to determine systematic uncertainties ATLAS Top.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
G. Cowan Statistics for HEP / NIKHEF, December 2011 / Lecture 3 1 Statistical Methods for Particle Physics Lecture 3: Limits for Poisson mean: Bayesian.
G. Cowan RHUL Physics Bayesian methods for HEP / DESY Terascale School page 1 Bayesian statistical methods for HEP Terascale Statistics School DESY, Hamburg.
1 Nuisance parameters and systematic uncertainties Glen Cowan Royal Holloway, University of London IoP Half.
G. Cowan Statistical methods for HEP / Freiburg June 2011 / Lecture 3 1 Statistical Methods for Discovery and Limits in HEP Experiments Day 3: Exclusion.
G. Cowan RHUL Physics Comment on use of LR for limits page 1 Comment on definition of likelihood ratio for limits ATLAS Statistics Forum CERN, 2 September,
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
G. Cowan 2011 CERN Summer Student Lectures on Statistics / Lecture 41 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability.
G. Cowan RHUL Physics Statistical Methods for Particle Physics / 2007 CERN-FNAL HCP School page 1 Statistical Methods for Particle Physics (2) CERN-FNAL.
G. Cowan Statistical Data Analysis / Stat 4 1 Statistical Data Analysis Stat 4: confidence intervals, limits, discovery London Postgraduate Lectures on.
G. Cowan Lectures on Statistical Data Analysis Lecture 14 page 1 Statistical Data Analysis: Lecture 14 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 13 page 1 Statistical Data Analysis: Lecture 13 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
G. Cowan Discovery and limits / DESY, 4-7 October 2011 / Lecture 3 1 Statistical Methods for Discovery and Limits Lecture 3: Limits for Poisson mean: Bayesian.
G. Cowan RHUL Physics Bayesian Higgs combination page 1 Bayesian Higgs combination using shapes ATLAS Statistics Meeting CERN, 19 December, 2007 Glen Cowan.
Graduierten-Kolleg RWTH Aachen February 2014 Glen Cowan
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
Statistical Decision Theory
G. Cowan SUSSP65, St Andrews, August 2009 / Statistical Methods 2 page 1 Statistical Methods in Particle Physics Lecture 2: Limits and Discovery.
G. Cowan 2009 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability densities,
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
G. Cowan Lectures on Statistical Data Analysis Lecture 1 page 1 Lectures on Statistical Data Analysis London Postgraduate Lectures on Particle Physics;
Practical Statistics for Particle Physicists Lecture 3 Harrison B. Prosper Florida State University European School of High-Energy Physics Anjou, France.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
G. Cowan RHUL Physics Bayesian Higgs combination page 1 Bayesian Higgs combination based on event counts (follow-up from 11 May 07) ATLAS Statistics Forum.
G. Cowan RHUL Physics page 1 Status of search procedures for ATLAS ATLAS-CMS Joint Statistics Meeting CERN, 15 October, 2009 Glen Cowan Physics Department.
G. Cowan, RHUL Physics Discussion on significance page 1 Discussion on significance ATLAS Statistics Forum CERN/Phone, 2 December, 2009 Glen Cowan Physics.
G. Cowan RHUL Physics LR test to determine number of parameters page 1 Likelihood ratio test to determine best number of parameters ATLAS Statistics Forum.
G. Cowan CERN Academic Training 2010 / Statistics for the LHC / Lecture 41 Statistics for the LHC Lecture 4: Bayesian methods and further topics Academic.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Lecture 4 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
G. Cowan CERN-JINR 2009 Summer School / Topics in Statistical Data Analysis page 1 Topics in Statistical Data Analysis for HEP Lecture 1: Bayesian Methods.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
G. Cowan, RHUL Physics Statistics for early physics page 1 Statistics jump-start for early physics ATLAS Statistics Forum EVO/Phone, 4 May, 2010 Glen Cowan.
G. Cowan Statistical Methods in Particle Physics1 Statistical Methods in Particle Physics Day 5: Bayesian Methods 清华大学高能物理研究中心 2010 年 4 月 12—16 日 Glen.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1 Lecture 5 1 Probability Definition, Bayes’ theorem, probability densities and their properties,
G. Cowan RHUL Physics Statistical Issues for Higgs Search page 1 Statistical Issues for Higgs Search ATLAS Statistics Forum CERN, 16 April, 2007 Glen Cowan.
1 Nuisance parameters and systematic uncertainties Glen Cowan Royal Holloway, University of London IoP Half.
G. Cowan CERN Academic Training 2010 / Statistics for the LHC / Lecture 21 Statistics for the LHC Lecture 2: Discovery Academic Training Lectures CERN,
G. Cowan SLAC Statistics Meeting / 4-6 June 2012 / Two Developments 1 Two developments in discovery tests: use of weighted Monte Carlo events and an improved.
G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1 Lecture 5 1 Probability Definition, Bayes’ theorem, probability densities and their properties,
Discussion on significance
Computing and Statistical Data Analysis / Stat 11
Statistics for the LHC Lecture 3: Setting limits
Lecture 4 1 Probability (90 min.)
Computing and Statistical Data Analysis / Stat 8
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Statistics for Particle Physics Lecture 3: Parameter Estimation
Statistical Tests and Limits Lecture 2: Limits
Computing and Statistical Data Analysis / Stat 7
Introduction to Statistics − Day 4
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Computing and Statistical Data Analysis / Stat 10
Presentation transcript:

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 31 Statistical Tests and Limits Lecture 3 IN2P3 School of Statistics Autrans, France 17—21 May, 2010 Glen Cowan Physics Department Royal Holloway, University of London

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3 1 page 2 Outline Lecture 1: General formalism Definition and properties of a statistical test Significance tests (and goodness-of-fit), p-values Lecture 2: Setting limits Confidence intervals Bayesian Credible intervals Lecture 3: Further topics for tests and limits More on intervals/limits, CLs Systematics / nuisance parameters Bayesian model selection

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 33 Likelihood ratio limits (Feldman-Cousins) Define likelihood ratio for hypothesized parameter, e.g., for expected number of signal events s: Here is the ML estimator, note Critical region defined by low values of likelihood ratio. Resulting intervals can be one- or two-sided (depending on n). (Re)discovered for HEP by Feldman and Cousins, Phys. Rev. D 57 (1998) 3873.

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 34 Upper limit versus b If n = 0 observed, should upper limit depend on b? Classical: yes FC: yes, but less so Feldman & Cousins, PRD 57 (1998) 3873 Feldman-Cousins “Classical”

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 35 More on intervals from LR test (Feldman-Cousins) Caveat with coverage: suppose we find n >> b. Usually one then quotes a measurement: If, however, n isn’t large enough to claim discovery, one sets a limit on s. FC pointed out that if this decision is made based on n, then the actual coverage probability of the interval can be less than the stated confidence level (‘flip-flopping’). FC intervals remove this, providing a smooth transition from 1- to 2-sided intervals, depending on n. But, suppose FC gives e.g. 0.1 < s < 5 at 90% CL, p-value of s=0 still substantial. Part of upper-limit ‘wasted’? For this reason, one-sided intervals for limits still popular.

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 36 The Bayesian approach to limits In Bayesian statistics need to start with ‘prior pdf’  (  ), this reflects degree of belief about  before doing the experiment. Bayes’ theorem tells how our beliefs should be updated in light of the data x: Integrate posterior pdf p(  | x) to give interval with any desired probability content. For e.g. Poisson parameter 95% CL upper limit from

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 37 Bayesian prior for Poisson parameter Include knowledge that s ≥0 by setting prior  (s) = 0 for s<0. Often try to reflect ‘prior ignorance’ with e.g. Not normalized but this is OK as long as L(s) dies off for large s. Not invariant under change of parameter — if we had used instead a flat prior for, say, the mass of the Higgs boson, this would imply a non-flat prior for the expected number of Higgs events. Doesn’t really reflect a reasonable degree of belief, but often used as a point of reference; or viewed as a recipe for producing an interval whose frequentist properties can be studied (coverage will depend on true s).

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 38 Bayesian interval with flat prior for s Solve numerically to find limit s up. For special case b = 0, Bayesian upper limit with flat prior numerically same as classical case (‘coincidence’). Otherwise Bayesian limit is everywhere greater than classical (‘conservative’). Never goes negative. Doesn’t depend on b if n = 0.

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 39 Priors from formal rules Because of difficulties in encoding a vague degree of belief in a prior, one often attempts to derive the prior from formal rules, e.g., to satisfy certain invariance principles or to provide maximum information gain for a certain set of measurements. Often called “objective priors” Form basis of Objective Bayesian Statistics The priors do not reflect a degree of belief (but might represent possible extreme cases). In a Subjective Bayesian analysis, using objective priors can be an important part of the sensitivity analysis.

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 310 Priors from formal rules (cont.) In Objective Bayesian analysis, can use the intervals in a frequentist way, i.e., regard Bayes’ theorem as a recipe to produce an interval with certain coverage properties. For a review see: Formal priors have not been widely used in HEP, but there is recent interest in this direction; see e.g. L. Demortier, S. Jain and H. Prosper, Reference priors for high energy physics, arxiv: (Feb 2010)

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 311 Jeffreys’ prior According to Jeffreys’ rule, take prior according to where is the Fisher information matrix. One can show that this leads to inference that is invariant under a transformation of parameters. For a Gaussian mean, the Jeffreys’ prior is constant; for a Poisson mean  it is proportional to 1/√ .

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 312 Jeffreys’ prior for Poisson mean Suppose n ~ Poisson(  ). To find the Jeffreys’ prior for , So e.g. for  = s + b, this means the prior  (s) ~ 1/√(s + b), which depends on b. But this is not designed as a degree of belief about s.

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 313 Properties of upper limits Upper limit s up vs. nMean upper limit vs. s Example: take b = 5.0, 1   = 0.95

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 314 Coverage probability of intervals Because of discreteness of Poisson data, probability for interval to include true value in general > confidence level (‘over-coverage’)

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 15 The “CL s ” issue When the cross section for the signal process becomes small (e.g., large Higgs mass), the distribution of the test variable used in a search becomes the same under both the b and s+b hypotheses: In such a case we will reject the signal hypothesis with a probability approaching  = 1 – CL (i.e. 5%) assuming no signal. f (q| b) f (q| s+b)

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 16 The CL s solution The CL s solution (A. Read et al.) is to base the test not on the usual p-value (CL s+b ), but rather to divide this by CL b (one minus the background of the b-only hypothesis, i.e., Define: Reject signal hypothesis if: Reduces “effective” p-value when the two distributions become close (prevents exclusion if sensitivity is low). f (q| b) q f (q| s+b)

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 17 CL s discussion In the CLs method the p-value is reduced according to the recipe Statistics community does not smile upon ratio of p-values; would prefer to regard parameter  as excluded if: (a) p-value of  < 0.05 (b) power of test of  with respect to background-only > some threshold (0.5?) Needs study. In any case should produce CLs result for purposes of comparison with other experiments.

G. Cowan Statistical Methods in Particle Physics18 Systematic errors and nuisance parameters Model prediction (including e.g. detector effects) never same as "true prediction" of the theory: x y model: truth: Model can be made to approximate better the truth by including more free parameters. systematic uncertainty ↔ nuisance parameters

19 Nuisance parameters and limits In general we don’t know the background b perfectly. Suppose we have a measurement of b, e.g., b meas ~ N (b,  b ) So the data are really: n events and the value b meas. In principle the exact confidence interval recipe can be generalized to multiple parameters, minimum coverage guaranteed. Difficult because of overcoverage; see e.g. talks by K. Cranmer at PHYSTAT03 and by G. Punzi at PHYSTAT05. G. Punzi, PHYSTAT05 G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 20 Nuisance parameters in limits (2) Connect systematic to nuisance parameters. Then form e.g. Profile likelihood: Marginal likelihood: and use these to construct e.g. likelihood ratios for tests. Coverage not guaranteed for all values of the nuisance params. Results of both approaches above often similar, but some care is needed in writing down prior; this should truly reflect one’s degree of belief about the parameters.

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 21 Nuisance parameters and profile likelihood Suppose model has likelihood function Parameters of interest Nuisance parameters Define the profile likelihood ratio as Maximizes L for given value of  Maximizes L (  ) reflects level of agreement between data and  (0 ≤ (  ) ≤ 1) Equivalently use q  =  2 ln (  )

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 22 p-value from profile likelihood ratio Large q  means worse agreement between data and  p-value = Prob(data with ≤ compatibility with  when compared to the data we got |  ) rapidly approaches chi-square pdf (Wilks’ theorem) chi-square cumulative distribution, degrees of freedom = dimension of  Reject  if p  <  = 1 – CL (Approx.) confidence interval for  = set of  values not rejected. Coverage not exact for all but very good if

23 Cousins-Highland method Regard b as ‘random’, characterized by pdf  (b). Makes sense in Bayesian approach, but in frequentist model b is constant (although unknown). A measurement b meas is random but this is not the mean number of background events, rather, b is. Compute anyway This would be the probability for n if Nature were to generate a new value of b upon repetition of the experiment with  b (b). Now e.g. use this P(n;s) in the classical recipe for upper limit at CL = 1  : Result has hybrid Bayesian/frequentist character. G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3

24 Marginal likelihood in LR tests Consider again signal s and background b, suppose we have uncertainty in b characterized by a prior pdf  b (b). Define marginal likelihood as also called modified profile likelihood, in any case not a real likelihood. Now use this to construct likelihood ratio test and invert to obtain confidence intervals. Feldman-Cousins & Cousins-Highland (FHC 2 ), see e.g. J. Conrad et al., Phys. Rev. D67 (2003) and Conrad/Tegenfeldt PHYSTAT05 talk. Calculators available (Conrad, Tegenfeldt, Barlow). G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 25 Comment on profile likelihood Suppose originally we measure x, likelihood is L(x|  ). To cover a systematic, we enlarge model to include a nuisance parameter, new model is L(x|  ). To use profile likelihood, data must constrain the nuisance parameters, otherwise suffer loss of accuracy in parameters of interest. Can e.g. use a separate measurement to constrain, e.g., with likelihood L(y| ). This becomes part of the full likelihood, i.e.,

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 26 Comment on marginal likelihood When using a prior to reflect knowledge of, often one treats this as coming from the measurement y, i.e., original prior, Then the marginal likelihood is So here L in the integrand does not include the information from the measurement y; this is included in the prior.

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 27 Bayesian limits with uncertainty on b Uncertainty on b goes into the prior, e.g., Put this into Bayes’ theorem, Marginalize over b, then use p(s|n) to find intervals for s with any desired probability content. Framework for treatment of nuisance parameters well defined; choice of prior can still be problematic, but often less so than finding a “non-informative” prior for a parameter of interest.

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 28

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 29 Bayesian model selection (‘discovery’) no Higgs Higgs The probability of hypothesis H 0 relative to an alternative H 1 is often given by the posterior odds: Bayes factor B 01 prior odds The Bayes factor is regarded as measuring the weight of evidence of the data in support of H 0 over H 1. Interchangeably use B 10 = 1/B 01

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 30 Assessing Bayes factors One can use the Bayes factor much like a p-value (or Z value). There is an “established” scale, analogous to our 5  rule: B 10 Evidence against H to 3Not worth more than a bare mention 3 to 20Positive 20 to 150Strong > 150Very strong Kass and Raftery, Bayes Factors, J. Am Stat. Assoc 90 (1995) 773. Will this be adopted in HEP?

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 31 Rewriting the Bayes factor Suppose we have models H i, i = 0, 1,..., each with a likelihood and a prior pdf for its internal parameters so that the full prior is where is the overall prior probability for H i. The Bayes factor comparing H i and H j can be written

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 32 Bayes factors independent of P(H i ) For B ij we need the posterior probabilities marginalized over all of the internal parameters of the models: Use Bayes theorem So therefore the Bayes factor is The prior probabilities p i = P(H i ) cancel. Ratio of marginal likelihoods

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 33 Numerical determination of Bayes factors Both numerator and denominator of B ij are of the form ‘marginal likelihood’ Various ways to compute these, e.g., using sampling of the posterior pdf (which we can do with MCMC). Harmonic Mean (and improvements) Importance sampling Parallel tempering (~thermodynamic integration) Nested sampling... See e.g.

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 34 Harmonic mean estimator E.g., consider only one model and write Bayes theorem as:  (  ) is normalized to unity so integrate both sides, Therefore sample  from the posterior via MCMC and estimate m with one over the average of 1/L (the harmonic mean of L). posterior expectation

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 35 Improvements to harmonic mean estimator The harmonic mean estimator is numerically very unstable; formally infinite variance (!). Gelfand & Dey propose variant: Rearrange Bayes thm; multiply both sides by arbitrary pdf f(  ): Integrate over  : Improved convergence if tails of f(  ) fall off faster than L(x|  )  (  ) Note harmonic mean estimator is special case f(  ) =  (  )..

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 36 Importance sampling Need pdf f(  ) which we can evaluate at arbitrary  and also sample with MC. The marginal likelihood can be written Best convergence when f(  ) approximates shape of L(x|  )  (  ). Use for f(  ) e.g. multivariate Gaussian with mean and covariance estimated from posterior (e.g. with MINUIT).

G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3page 37 p-values versus Bayes factors Current convention: p-value of background-only < 2.9 × 10  7 (5  ) This should really depend also on other factors: Plausibility of signal Confidence in modeling of background Can also use Bayes factor Should hopefully point to same conclusion as p-value. If not, need to understand why! As yet not widely used in HEP, numerical issues not easy.

page 38 Summary G. Cowan SOS 2010 / Statistical Tests and Limits -- lecture 3 Bayesian approach to setting limits is straightfoward; all information about the parameter is in the posterior probability, integrate this to get intervals with given probability. Difficult to find appropriate “non-informative” prior. Often use Bayesian approach as a recipe for producing interval, then study it in a frequentist way (e.g. coverage) The key to treating systematic uncertainties is to include in the model enough parameters so that it is correct (or very close). But too many parameters degrades information on parameters of interest Bayesian model selection Bayes factor = posterior odds if prior odds = 1. Only requires priors for internal parameters of models. Can be very difficult to compute numerically.