Multidimensional Integration Part I

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Bayesian Estimation in MARK
Ch 11. Sampling Models Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by I.-H. Lee Biointelligence Laboratory, Seoul National.
Gibbs Sampling Qianji Zheng Oct. 5th, 2010.
Markov-Chain Monte Carlo
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Bayesian statistics – MCMC techniques
BAYESIAN INFERENCE Sampling techniques
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
Bayesian Analysis of X-ray Luminosity Functions A. Ptak (JHU) Abstract Often only a relatively small number of sources of a given class are detected in.
Statistical Tools PhyStat Workshop 2004 Harrison B. Prosper1 Statistical Tools A Few Comments Harrison B. Prosper Florida State University PHYSTAT Workshop.
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
Bayesian Neural Networks Pushpa Bhat Fermilab Harrison Prosper Florida State University.
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Bayes Factor Based on Han and Carlin (2001, JASA).
Harrison B. Prosper Workshop on Top Physics, Grenoble Bayesian Statistics in Analysis Harrison B. Prosper Florida State University Workshop on Top Physics:
1 Physical Fluctuomatics 5th and 6th Probabilistic information processing by Gaussian graphical model Kazuyuki Tanaka Graduate School of Information Sciences,
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Bayesian Inversion of Stokes Profiles A.Asensio Ramos (IAC) M. J. Martínez González (LERMA) J. A. Rubiño Martín (IAC) Beaulieu Workshop ( Beaulieu sur.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
G. Cowan RHUL Physics Bayesian Higgs combination page 1 Bayesian Higgs combination based on event counts (follow-up from 11 May 07) ATLAS Statistics Forum.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
MCMC reconstruction of the 2 HE cascade events Dmitry Chirkin, UW Madison.
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
Reducing MCMC Computational Cost With a Two Layered Bayesian Approach
The Uniform Prior and the Laplace Correction Supplemental Material not on exam.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Bayesian Modelling Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
8 Sept 2006, DEMA2006Slide 1 An Introduction to Computer Experiments and their Design Problems Tony O’Hagan University of Sheffield.
Canadian Bioinformatics Workshops
CS498-EA Reasoning in AI Lecture #19 Professor: Eyal Amir Fall Semester 2011.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Markov Chain Monte Carlo in R
(joint work with Ai-ru Cheng, Ron Gallant, Beom Lee)
MCMC Output & Metropolis-Hastings Algorithm Part I
Advanced Statistical Computing Fall 2016
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 12
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Basic simulation methodology
Intro to Sampling Methods
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Jun Liu Department of Statistics Stanford University
Course: Autonomous Machine Learning
Bayesian inference Presented by Amir Hadadi
Markov chain monte carlo
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
Predictive distributions
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 12
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
10701 / Machine Learning Today: - Cross validation,
Markov Chain Monte Carlo
Statistics for Particle Physics Lecture 3: Parameter Estimation
LECTURE 07: BAYESIAN ESTIMATION
Parametric Methods Berlin Chen, 2005 References:
Opinionated Lessons #39 MCMC and Gibbs Sampling in Statistics
Computing and Statistical Data Analysis / Stat 10
Yalchin Efendiev Texas A&M University
Maximum Likelihood Estimation (MLE)
Uncertainty Propagation
Outline Texture modeling - continued Markov Random Field models
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 12
Presentation transcript:

Multidimensional Integration Part I Harrison B. Prosper Florida State University Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper Outline Do we need it? Markov Chain Monte Carlo Adaptive Methods Summary Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper Do we need it? Most analyses in high energy physics are done using frequentist methods. The more sophisticated ones typically involve the minimization of log likelihoods using programs such as the celebrated MINUIT. These methods in general do not need multidimensional integration. Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper But we may need it if… We wish to do analyses using Bayesian methods. Here are a few examples: Limit-setting DØ Single Top Analysis Luminosity estimation SUSY/Higgs Workshop Jet energy scale corrections Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Jet Energy Scale Corrections 1. Assume we have a pure sample of 2. Assume Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper A single event… Likelihood Prior Posterior Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

But need lots of events, in practice! Posterior Number of dimensions = 2N+m, where N is the number of events used and m is the number of ‘a’ parameters. If N ~ 1000 and m ~ 3, we have Ndim ~ 2000!! Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Multidimensional Integration Low dimensions, that is, < than about 20 Adaptive Numerical Integration Recursively partition space while working to reduce the integration error on the partition, which at a given step, has the largest error. High dimensions, that is, > than about 20 Markov Chain Monte Carlo Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper The Basic Idea Generate a sequence of parameter values ai from the posterior distribution Post(a|D) and compute averages: In general, it its very difficult to sample directly from a complicated distribution. Gaussians, of course are easy! Must use indirect method to generate sequence. Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper The Basic Idea, cont. If the sequence of ai are statistically independent the uncertainty in the estimate of the integral is just the error on the mean: Important: The error reduces slowly, but it does so in a manner that is independent of the dimensionality of the space. Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Markov Chain Monte Carlo State, x: a vector of real-valued quantities Transition probability, T: probability to get state x(t+1) given state x(t) Proposal probability, q: probability to propose a new state y(t+1) given state x(t) Acceptance probability, A: probability to accept the proposed state. Markov chain: random sequence of states x(t) with the property that the probability to get state x(t+1) depends only on the previous state x(t). Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper MCMC 1 Let pt+1(x) be the probability of state x at time step t+1 and pt(x) be the probability of state x at time step t. Then The goal is to produce the following condition: as the time step t goes to infinity. That is, to arrive at a stationary (or invariant, or equilibrium) distribution π(x) Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper

Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper MCMC 2 Next time we shall see how that condition can be achieved…! Fermilab RAAG Meeting, 25 May, 2001, Harrison B. Prosper