Introduction Paleontologists often need to model complex systems with many variables and complex relationships. In such models, information is often characterized.

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

Markov Chain Monte Carlo Convergence Diagnostics: A Comparative Review By Mary Kathryn Cowles and Bradley P. Carlin Presented by Yuting Qi 12/01/2006.
Monte Carlo Methods and Statistical Physics
Bayesian Estimation in MARK
Efficient Cosmological Parameter Estimation with Hamiltonian Monte Carlo Amir Hajian Amir Hajian Cosmo06 – September 25, 2006 Astro-ph/
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
By Addison Euhus, Guidance by Edward Phillips An Introduction To Uncertainty Quantification.
Gibbs Sampling Qianji Zheng Oct. 5th, 2010.
Markov-Chain Monte Carlo
Markov Chains 1.
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Bayesian statistics – MCMC techniques
Stochastic approximate inference Kay H. Brodersen Computational Neuroeconomics Group Department of Economics University of Zurich Machine Learning and.
BAYESIAN INFERENCE Sampling techniques
Computing the Posterior Probability The posterior probability distribution contains the complete information concerning the parameters, but need often.
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
Lecture 5: Learning models using EM
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
An Optimal Learning Approach to Finding an Outbreak of a Disease Warren Scott Warren Powell
Today Introduction to MCMC Particle filters and MCMC
1 Transforming the efficiency of Partial EVSI computation Alan Brennan Health Economics and Decision Science (HEDS) Samer Kharroubi Centre for Bayesian.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Lecture II-2: Probability Review
Robin McDougall, Ed Waller and Scott Nokleby Faculties of Engineering & Applied Science and Energy Systems & Nuclear Science 1.
Introduction to Monte Carlo Methods D.J.C. Mackay.
Bayes Factor Based on Han and Carlin (2001, JASA).
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Priors, Normal Models, Computing Posteriors
Review of Chapters 1- 5 We review some important themes from the first 5 chapters 1.Introduction Statistics- Set of methods for collecting/analyzing data.
Harrison B. Prosper Workshop on Top Physics, Grenoble Bayesian Statistics in Analysis Harrison B. Prosper Florida State University Workshop on Top Physics:
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Module 1: Statistical Issues in Micro simulation Paul Sousa.
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
Bayesian Inversion of Stokes Profiles A.Asensio Ramos (IAC) M. J. Martínez González (LERMA) J. A. Rubiño Martín (IAC) Beaulieu Workshop ( Beaulieu sur.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Lab3: Bayesian phylogenetic Inference and MCMC Department of Bioinformatics & Biostatistics, SJTU.
Fast Simulators for Assessment and Propagation of Model Uncertainty* Jim Berger, M.J. Bayarri, German Molina June 20, 2001 SAMO 2001, Madrid *Project of.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
Bayesian Reasoning: Tempering & Sampling A/Prof Geraint F. Lewis Rm 560:
Sampling distributions rule of thumb…. Some important points about sample distributions… If we obtain a sample that meets the rules of thumb, then…
G. Cowan RHUL Physics Bayesian Higgs combination page 1 Bayesian Higgs combination based on event counts (follow-up from 11 May 07) ATLAS Statistics Forum.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
An Introduction to Markov Chain Monte Carlo Teg Grenager July 1, 2004.
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
1 1 Slide Simulation Professor Ahmadi. 2 2 Slide Simulation Chapter Outline n Computer Simulation n Simulation Modeling n Random Variables and Pseudo-Random.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
STA347 - week 91 Random Vectors and Matrices A random vector is a vector whose elements are random variables. The collective behavior of a p x 1 random.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Bayesian Modelling Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Markov-Chain-Monte-Carlo (MCMC) & The Metropolis-Hastings Algorithm P548: Intro Bayesian Stats with Psych Applications Instructor: John Miyamoto 01/19/2016:
Kevin Stevenson AST 4762/5765. What is MCMC?  Random sampling algorithm  Estimates model parameters and their uncertainty  Only samples regions of.
How many iterations in the Gibbs sampler? Adrian E. Raftery and Steven Lewis (September, 1991) Duke University Machine Learning Group Presented by Iulian.
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
CS498-EA Reasoning in AI Lecture #19 Professor: Eyal Amir Fall Semester 2011.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Markov Chain Monte Carlo in R
MCMC Output & Metropolis-Hastings Algorithm Part I
Optimization of Monte Carlo Integration
Bayesian data analysis
Ch13 Empirical Methods.
Statistical Data Analysis
Parametric Methods Berlin Chen, 2005 References:
Presentation transcript:

Introduction Paleontologists often need to model complex systems with many variables and complex relationships. In such models, information is often characterized by high-dimensional statistical distributions that are difficult to analyze mathematically. In this poster, we describe the use of Markov Chain Monte Carlo (MCMC) in generating an approximate sample from any desired distribution. In so doing, the sample provides important insights into the nature of the distribution at hand. MCMC is an iterative approach towards simulating a sample, creating a sequence of values (or vector of values) whose distribution more closely approximates the desired distribution the longer the chain is allowed to run. More specifically, we present the Metropolis-Hastings Algorithm, a particular implementation of MCMC, which easily adapts to high-dimensional problems. In this poster, we use MCMC to gain insight into the primary causes of the Permian mass extinction. We use a computer model to simulate secondary extinctions in a Late Permian food web. Then, using MCMC methods and Bayesian modeling, we infer the level of primary producer shutdown needed to cause observed levels of secondary extinction in Late Permian Karoo basin fauna. Mysterious Mass Extinction The end-Permian or Permian-Triassic extinction event occurred approximately 251 million years ago and nearly wiped out all life on Earth. Possibly as few as ten percent of all species then extant survived the event. In contrast, approximately half of the species present 65 million years ago survived the end-Cretaceous extinction. However, the differences do not end there. The causes of the end- Cretaceous event are well-explained by clear physical evidence. Conversely, less direct physical evidence remains of the causes (primary extinctions) leading up to the end- Permian event. There are numerous theories to account for the mass extinction, including massive volcanism (Figure 1) and a meteor impact scenario similar to the end-Cretaceous event (Figure 2), Markov Chain Monte Carlo: New Tools for Bayesian Modeling in Paleontology Michael D. Karcher ’07 (1), Steve C. Wang (1), Peter Roopnarine (2), Kenneth Angielczyk (3) (1) Department of Mathematics and Statistics, Swarthmore College (2) Department of Invertebrate Zoology & Geology, California Academy of Sciences (3) Department of Earth Sciences, University of Bristol For further information: or Food Web Collapse During any ecological disturbance, repercussions propagate throughout the entire food web. It is possible to simulate these effects computationally. The necessary framework to create these simulations has already been explored (Roopnarine 2006, Paleobiology 32:1, 1-19). Using a network graph to describe the food web in terms of “guilds” (for example, “Large Herbivores”, “Small Amphibians”, etc.), the function ceg(x) (Cascading Extinction on Graphs) takes a vector of primary extinctions—values in [0,1] representing extinctions as a direct result of an extinction event—and outputs a vector of secondary extinctions—extinctions as a result of food web collapse. For example, in a system with three guilds, ceg(0.45, 0.05, 0.0) = (0.85, 0.75, 0.8), shows that moderate extinction in one guild can cause major effects in other guilds, mirroring reality. Metropolis-Hastings Algorithm One particular MCMC technique that adapts well to high- dimensional problems is the Metropolis-Hastings Algorithm. It takes a Random-Walk (Figure 3) approach towards exploring the given distribution. Suppose we wish to sample from a given distribution with density function f(x). First, we choose a “jumping” distribution g(x|θ) that is easy to sample from and depends only on one parameter (e.g. Normal with mean θ). We choose x 1 from any point where f(x) is positive, and at each step x t of the chain, we sample a new point x * from the jumping distribution and calculate: Example: MCMC in Action The plots in Figure 4 were created using the Metropolis- Hastings Algorithm in one dimension. It was run on the food web networks from the Karoo basin dataset and its objective was to find the true value of the primary extinction. Paleontological Uses Using the data from the simulations of food web collapse, it is possible to treat the potential primary extinction scenarios in the end-Permian event as a statistical distribution and analyze them using MCMC. In order to do so, we set up a Metropolis- Hastings chain with points from the possible primary extinctions. We apply the CEG function (transforming primary extinction into secondary extinction) to each point many times to generate a likelihood function by counting how many of those outputs fall within a predetermined n-cube around the observed secondary extinction. Once the chain has sufficiently explored the distribution, we can apply Bayes’ Theorem with a flat prior to calculate the posterior distribution for the primary extinction scenarios. Using simple statistical techniques, we can then construct estimates and confidence intervals about which primary extinction scenario caused the end-Permian event. Markov Chain Monte Carlo Statisticians often need to explore statistical distributions that are difficult to analyze directly. Inconvenient distributions often have pathological or even no mathematical formulae, domain spaces that are too large to fully exhaust, or other traits that make direct analysis nearly impossible. One class of tools statisticians possess to work through this is Markov Chain Monte Carlo (MCMC). What binds the tools of MCMC together is that they are all iterative and asymptotic processes; they all proceed step-by-step and get more accurate the longer they are run. The Markov property plays a large role in the utility of the techniques. As a given MCMC process iterates, each step depends only on the last step and no other information. It can be shown that as the length of a chain with the Markov property increases, it converges to some distribution, and methods exist to ensure that it converges to any desired distribution, even inconvenient distributions, allowing analysis to continue. Future The simulations produce a great deal of data, and much work still remains in its interpretation. Running the Algorithm on the real world data produces 24-dimensional results. Figure 5 is a collection of histograms from one of our runs summarizing the data marginally. More useful inferences can be gathered from analyzing the variables and their correlations. However, this is no trivial task in 24 dimensions. A great deal of time could be spent in future projects analyzing and determining how to analyze the data that is produced. Fig. 1Fig. 2 Fig. 3: Metropolis-Hastings Random-Walk Markov Chain Fig. 4 And then we set: There are several techniques to improve the sample, including discarding the first part of the chain to reduce the effect of choosing x 1 arbitrarily and discarding all but every k th step to reduce correlation. Fig. 5