MCMC Stopping and Variance Estimation: Idea here is to first use multiple Chains from different initial conditions to determine a burn-in period so the.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Introduction to Monte Carlo Markov chain (MCMC) methods
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
Bayesian Estimation in MARK
Chapter 4: Linear Models for Classification
AP Statistics – Chapter 9 Test Review
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Probability & Statistical Inference Lecture 7 MSc in Computing (Data Analytics)
G. Cowan RHUL Physics Profile likelihood for systematic uncertainties page 1 Use of profile likelihood to determine systematic uncertainties ATLAS Top.
Chapter Seventeen HYPOTHESIS TESTING
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Mixture Modeling Chongming Yang Research Support Center FHSS College.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Bayesian Hypothesis Testing and Bayes Factors 1)Bayesian p-values 2)Bayes Factors for model comparison 3)Easy to implement alternatives for model comparison.
Bayesian Model Selection in Factorial Designs Seminal work is by Box and Meyer Seminal work is by Box and Meyer Intuitive formulation and analytical approach,
AP STATISTICS LESSON 10 – 1 (DAY 2)
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Bayesian Hierarchical Clustering Paper by K. Heller and Z. Ghahramani ICML 2005 Presented by HAO-WEI, YEH.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Applied Bayesian Inference, KSU, April 29, 2012 § ❷ / §❷ An Introduction to Bayesian inference Robert J. Tempelman 1.
Bayesian Inversion of Stokes Profiles A.Asensio Ramos (IAC) M. J. Martínez González (LERMA) J. A. Rubiño Martín (IAC) Beaulieu Workshop ( Beaulieu sur.
Lab3: Bayesian phylogenetic Inference and MCMC Department of Bioinformatics & Biostatistics, SJTU.
Likelihood function and Bayes Theorem In simplest case P(B|A) = P(A|B) P(B)/P(A) and we consider the likelihood function in which we view the conditional.
Week 41 Estimation – Posterior mean An alternative estimate to the posterior mode is the posterior mean. It is given by E(θ | s), whenever it exists. This.
Latent Class Regression Model Graphical Diagnostics Using an MCMC Estimation Procedure Elizabeth S. Garrett Scott L. Zeger Johns Hopkins University
Bayesian Statistics Lecture 8 Likelihood Methods in Forest Ecology October 9 th – 20 th, 2006.
On Predictive Modeling for Claim Severity Paper in Spring 2005 CAS Forum Glenn Meyers ISO Innovative Analytics Predictive Modeling Seminar September 19,
Confidence Interval & Unbiased Estimator Review and Foreword.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
1 G Lect 7a G Lecture 7a Comparing proportions from independent samples Analysis of matched samples Small samples and 2  2 Tables Strength.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Chapter 2: Bayesian hierarchical models in geographical genetics Manda Sayler.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
Bayesian Hierarchical Clustering Paper by K. Heller and Z. Ghahramani ICML 2005 Presented by David Williams Paper Discussion Group ( )
Markov-Chain-Monte-Carlo (MCMC) & The Metropolis-Hastings Algorithm P548: Intro Bayesian Stats with Psych Applications Instructor: John Miyamoto 01/19/2016:
ESTIMATION OF THE MEAN. 2 INTRO :: ESTIMATION Definition The assignment of plausible value(s) to a population parameter based on a value of a sample statistic.
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part Four ANALYSIS AND PRESENTATION OF DATA.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Model Comparison.
Bayesian Estimation and Confidence Intervals Lecture XXII.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Step 1: Specify a null hypothesis
Bayesian Estimation and Confidence Intervals
Part Four ANALYSIS AND PRESENTATION OF DATA
Central Tendency and Variability
Special Topics In Scientific Computing
Bayesian inference Presented by Amir Hadadi
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
CONCEPTS OF ESTIMATION
More about Posterior Distributions
Limitations of Hierarchical and Mixture Model Comparisons
Problems: Q&A chapter 6, problems Chapter 6:
Chapter Nine Part 1 (Sections 9.1 & 9.2) Hypothesis Testing
Using Ensembles of Cognitive Models to Answer Substantive Questions
Ch13 Empirical Methods.
Warmup To check the accuracy of a scale, a weight is weighed repeatedly. The scale readings are normally distributed with a standard deviation of
Appetizer: Statistical methods in XSPEC
PSY 626: Bayesian Statistics for Psychological Science
LESSON 18: CONFIDENCE INTERVAL ESTIMATION
Opinionated Lessons #39 MCMC and Gibbs Sampling in Statistics
Determining Which Method to use
CS639: Data Management for Data Science
Mathematical Foundations of BME Reza Shadmehr
Bayesian Networks in Educational Assessment
Classical regression review
Presentation transcript:

MCMC Stopping and Variance Estimation: Idea here is to first use multiple Chains from different initial conditions to determine a burn-in period so the variance is similar within a single Chain (e.g. a single trajectory of the MC) to that across several Chains. If we want to do estimation based on the posterior, then we need some variance estimates to test for example whether two MCMC estimates are statistically the same. One option is to partition the Chain (after burn-in) and use these “batches” to estimate the variance. Alternatively, you can use an estimator of the autocorrelation lag to either “thin” or throw out some of the data on the Chain, or else increase the variance estimate based upon the autcorrelation lag.

Bayesian Estimation: Once we have a posterior (or a sample of it from MCMC) then point estimation for measures of central tendency are simply calculated from the posterior. Interval estimates for parameters such as quartiles are found the same as usual: Then the interval [L ,U] gives the Bayesian confidence interval or  credible set (equal tail case).

Model Choice and Bayes Factors: The idea is to determine a procedure to chose between two models M1 and M2 which represent two choices of parameters 1 and 2 (vectors) and to obtain a posterior way to estimate how much confidence you might place in one model over the other. The Bayes factor is a formal method to determine how much the data y support one model over the other. We wish to find the posterior probabilities of the models P( Mi | y ) but summarize this by comparing the posterior to the prior odds of each model: Which simplifies to the likelihood ratio under appropriate assumptions.

The marginal density of the data given the models is And after using Bayes Theorem the Bayes factor is But obviously if the models differ in how many parameters they have (for example if one is nested inside another but has more parameters) then we need to correct for this when saying one has more support, which is what the Bayesian Information Criterion (BIC), Akaike Information Criterion (AIC) and Deviance Information Criterion (DIC) do.

Bayesian Hierarchical Models: The idea of hierarchical modeling is simply to break a complex system into pieces, construct a model for each piece, and then link them as necessary (e.g. model a plant by having models for root, stem and leaves and then link together via carbon/water transport). Bayesian hierarchical models are similar except that the links between models can be thought of as obtained via Bayesian methods (e.g. updating one component with data leads to iterative updating across the model components). In one view, the joint distribution of the full model depends upon submodels for the data, the process and the parameters and one uses Bayes to update the conditional distributions of these.