Reducing MCMC Computational Cost With a Two Layered Bayesian Approach

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

Contrastive Divergence Learning
An Adaptive Learning Method for Target Tracking across Multiple Cameras Kuan-Wen Chen, Chih-Chuan Lai, Yi-Ping Hung, Chu-Song Chen National Taiwan University.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
METHODS FOR HAPLOTYPE RECONSTRUCTION
Bayesian Estimation in MARK
Introduction of Markov Chain Monte Carlo Jeongkyun Lee.
Gibbs Sampling Qianji Zheng Oct. 5th, 2010.
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Bayesian Reasoning: Markov Chain Monte Carlo
BAYESIAN INFERENCE Sampling techniques
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
USE OF LAPLACE APPROXIMATIONS TO SIGNIFICANTLY IMPROVE THE EFFICIENCY
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
CENTER FOR BIOLOGICAL SEQUENCE ANALYSIS Bayesian Inference Anders Gorm Pedersen Molecular Evolution Group Center for Biological Sequence Analysis Technical.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Introduction to Monte Carlo Methods D.J.C. Mackay.
Overview G. Jogesh Babu. Probability theory Probability is all about flip of a coin Conditional probability & Bayes theorem (Bayesian analysis) Expectation,
Phylogeny Estimation: Traditional and Bayesian Approaches Molecular Evolution, 2003
WSEAS AIKED, Cambridge, Feature Importance in Bayesian Assessment of Newborn Brain Maturity from EEG Livia Jakaite, Vitaly Schetinin and Carsten.
Bayesian parameter estimation in cosmology with Population Monte Carlo By Darell Moodley (UKZN) Supervisor: Prof. K Moodley (UKZN) SKA Postgraduate conference,
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Probabilistic Mechanism Analysis. Outline Uncertainty in mechanisms Why consider uncertainty Basics of uncertainty Probabilistic mechanism analysis Examples.
Exam I review Understanding the meaning of the terminology we use. Quick calculations that indicate understanding of the basis of methods. Many of the.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Overview Particle filtering is a sequential Monte Carlo methodology in which the relevant probability distributions are iteratively estimated using the.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
Perceptual Multistability as Markov Chain Monte Carlo Inference.
Fast Simulators for Assessment and Propagation of Model Uncertainty* Jim Berger, M.J. Bayarri, German Molina June 20, 2001 SAMO 2001, Madrid *Project of.
An Efficient Sequential Design for Sensitivity Experiments Yubin Tian School of Science, Beijing Institute of Technology.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
Lecture #9: Introduction to Markov Chain Monte Carlo, part 3
Inference of Non-Overlapping Camera Network Topology by Measuring Statistical Dependence Date :
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Stochastic Frontier Models
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Bayesian Modelling Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
Gaussian Process Networks Nir Friedman and Iftach Nachman UAI-2K.
Markov-Chain-Monte-Carlo (MCMC) & The Metropolis-Hastings Algorithm P548: Intro Bayesian Stats with Psych Applications Instructor: John Miyamoto 01/19/2016:
Kevin Stevenson AST 4762/5765. What is MCMC?  Random sampling algorithm  Estimates model parameters and their uncertainty  Only samples regions of.
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
Introduction to emulators Tony O’Hagan University of Sheffield.
Efficiency Measurement William Greene Stern School of Business New York University.
Monte Carlo Sampling to Inverse Problems Wojciech Dębski Inst. Geophys. Polish Acad. Sci. 1 st Orfeus workshop: Waveform inversion.
10 October, 2007 University of Glasgow 1 EM Algorithm with Markov Chain Monte Carlo Method for Bayesian Image Analysis Kazuyuki Tanaka Graduate School.
CS498-EA Reasoning in AI Lecture #19 Professor: Eyal Amir Fall Semester 2011.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Markov Chain Monte Carlo in R
Reducing Photometric Redshift Uncertainties Through Galaxy Clustering
MCMC Output & Metropolis-Hastings Algorithm Part I
T. E. Dyhoum1, D. Lesnic 1 and R. G. Aykroyd 2
Advanced Statistical Computing Fall 2016
Ex1: Event Generation (Binomial Distribution)
Ch3: Model Building through Regression
Introduction to the bayes Prefix in Stata 15
Dynamical Statistical Shape Priors for Level Set Based Tracking
Bayesian inference Presented by Amir Hadadi
Markov chain monte carlo
Bayesian Refinement of Protein Functional Site Matching
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
Multidimensional Integration Part I
Robust Full Bayesian Learning for Neural Networks
Yalchin Efendiev Texas A&M University
Uncertainty Propagation
Presentation transcript:

Reducing MCMC Computational Cost With a Two Layered Bayesian Approach Ramin Madarshahian, Doctoral Candidate, mdrshhn@email.sc.edu Juan M. Caicedo, Associate Professor http://sdii.ce.sc.edu/

Outline Introduction Methodology Example Future work

Introduction

Markov Chain Monte Carlo (MCMC) A very powerful algorithm for getting sample from high dimensional and complicated probability distribution function. It gets more sample from high probability regions, and with enough number of samples, histogram of samples take similar shape as probability distribution of interest.

Markov Chain Monte Carlo (MCMC) Von Neumann he contributed to the development of the Monte Carlo method, which allowed solutions to complicated problems to be approximated using random numbers. Metropolis Ulam Paper 1949: Using Markov chain for Monte Carlo approximation Rossenburg Teller Paper 1953: They applied MCMC for a chemical problem

Bayesian modeling Bayesian modeling is good method to deal with uncertainty. Bayesian modeling update our belief about the model and its parameters by considering evidences. Evidence comes from inputs and outputs.

Problem? With a Bayesian model we would like to make an inference about a model and its parameters. MCMC can be used to sample the posterior, but for each sample we need to run the model. What if our model is very computational expensive?

Metamodeling Metamodel: Simply an approximation of the computationally expensive model. Also known as : Response surface, emulators, auxiliary models, etc. Computationally expensive models: Models of multi-scale problems like shear band, models of complicated structures like airplane, modelling of physical and biological phenomena like protein folding, etc.

General approach Few numbers of input samples. Using the expensive model to obtain corresponding output samples. Using these I/O relationships to fit the metamodel. For some methods, using obtained metamodel to select next input sample to better fit the metamodel. Replacing the expensive model with obtained metamodel, and using this new model in the Bayesian process and etc.

Proposed method In our proposed method, the surface of interest is obtained from the posterior of the Bayesian model, instead of direct approximation of the expensive model by the surrogate.

Motivation A posterior is a probability distribution function with all common characteristics of that. It is usually more well-defined in comparison to the expensive model itself. Depending upon the type of study, a researcher can focus on high probability regions (like model updating problems) or focus on tails (Reliability problems). This makes sampling more efficient.

Methodology

Bayesian modelling formulation Model’s parameters Data Posterior Likelihood

Proposed method formulation

Example

SDOF Simulated by assuming normal distribution with 𝜇=1000 𝑁 𝑚 and 𝜎=10 𝑁 𝑚 for K 𝑀=100 𝑘𝑔 then what is 𝐾 using the data in the table?

Inference without using the metamodel Assuming uniform prior for K from Zero to 2000:

Inference without using the metamodel MCMC: Selecting the prior: Assuming uniform prior for K from Zero to 2000 𝑁 𝑚 : A total of 10000 samples are generated and the first 2000 are discarded.

Inference without using the metamodel 𝜇 𝑘 =1034.84 𝑁 𝑚 95% 𝐻𝑖𝑔ℎ𝑒𝑠𝑡 𝑃𝑜𝑠𝑡𝑒𝑟𝑖𝑜𝑟 𝐷𝑒𝑛𝑠𝑖𝑡𝑦 𝐻𝑃𝐷 = 642 ,1428 𝑁 𝑚 𝜎=203 𝑁 𝑚

Proposed method

Proposed method Priors: For 𝑐 1 : a uniform distribution with the lower bound of 0.001 and the upper bound of 1.0. For 𝜇 1 : a uniform distribution with the lower bound of zero and the upper bound of 2000. For 𝜎 1 : a normal distribution with 𝜇=300 𝑁 𝑚 , and 𝜎=80 𝑁 𝑚 A total of 12000 samples were obtained, 4000 of them were considered burning samples and were discarded.

Proposed method

Proposed method Parameter Mean 95% HPD 𝜇 1 1021 𝑁 𝑚 [1011,1031] 𝑁 𝑚 [1011,1031] 𝑁 𝑚 𝜎 1 201 𝑁 𝑚 [190,210] 𝑁 𝑚 𝑐 1 0.051 [0.049,0.053]

Comparison Using obtained mean and standard deviation, 95% HPD will be [627, 1414] 𝑁 𝑚 which is comparable with results without the metamodel, i.e.[642, 1428] 𝑁 𝑚 .

Future work

Future work Considering different types of metamodels like polynomial, kriging, etc. Development of sampling strategies. Considering expensive models and study of computational power of method.

Reducing MCMC Computational Cost With a Two Layered Bayesian Approach Thank you! Reducing MCMC Computational Cost With a Two Layered Bayesian Approach Ramin Madarshahian, Doctoral Candidate, mdrshhn@email.sc.edu Juan M. Caicedo, Associate Professor http://sdii.ce.sc.edu/