Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

MCMC estimation in MlwiN
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Markov Chain Monte Carlo Convergence Diagnostics: A Comparative Review By Mary Kathryn Cowles and Bradley P. Carlin Presented by Yuting Qi 12/01/2006.
METHODS FOR HAPLOTYPE RECONSTRUCTION
Bayesian Estimation in MARK
1 Parametric Sensitivity Analysis For Cancer Survival Models Using Large- Sample Normal Approximations To The Bayesian Posterior Distribution Gordon B.
Gibbs Sampling Qianji Zheng Oct. 5th, 2010.
Markov-Chain Monte Carlo
2005 Hopkins Epi-Biostat Summer Institute1 Module 2: Bayesian Hierarchical Models Francesca Dominici Michael Griswold The Johns Hopkins University Bloomberg.
Reporter: Hsu Hsiang-Jung Modelling stochastic fish stock dynamics using Markov Chain Monte Carlo.
Bayesian statistics – MCMC techniques
Stochastic approximate inference Kay H. Brodersen Computational Neuroeconomics Group Department of Economics University of Zurich Machine Learning and.
Computing the Posterior Probability The posterior probability distribution contains the complete information concerning the parameters, but need often.
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
Bayesian Analysis of X-ray Luminosity Functions A. Ptak (JHU) Abstract Often only a relatively small number of sources of a given class are detected in.
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
Monte Carlo methods for estimating population genetic parameters Rasmus Nielsen University of Copenhagen.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
1 Spatial and Spatio-temporal modeling of the abundance of spawning coho salmon on the Oregon coast R Ruben Smith Don L. Stevens Jr. September.
Bayes Factor Based on Han and Carlin (2001, JASA).
Overview G. Jogesh Babu. Probability theory Probability is all about flip of a coin Conditional probability & Bayes theorem (Bayesian analysis) Expectation,
Model Inference and Averaging
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
A Beginner’s Guide to Bayesian Modelling Peter England, PhD EMB GIRO 2002.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
R2WinBUGS: Using R for Bayesian Analysis Matthew Russell Rongxia Li 2 November Northeastern Mensurationists Meeting.
Bioscience – Aarhus University Measuring plant abundance and ecological processes with the pin-point method Christian Damgaard Department of Bioscience.
A Comparison of Two MCMC Algorithms for Hierarchical Mixture Models Russell Almond Florida State University College of Education Educational Psychology.
Three Frameworks for Statistical Analysis. Sample Design Forest, N=6 Field, N=4 Count ant nests per quadrat.
Latent Class Regression Model Graphical Diagnostics Using an MCMC Estimation Procedure Elizabeth S. Garrett Scott L. Zeger Johns Hopkins University
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
MCMC reconstruction of the 2 HE cascade events Dmitry Chirkin, UW Madison.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
Bayesian Travel Time Reliability
Expectation-Maximization (EM) Algorithm & Monte Carlo Sampling for Inference and Approximation.
Item Parameter Estimation: Does WinBUGS Do Better Than BILOG-MG?
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Bayesian Modelling Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
 Occupancy Model Extensions. Number of Patches or Sample Units Unknown, Single Season So far have assumed the number of sampling units in the population.
A Study on Speaker Adaptation of Continuous Density HMM Parameters By Chin-Hui Lee, Chih-Heng Lin, and Biing-Hwang Juang Presented by: 陳亮宇 1990 ICASSP/IEEE.
Overview G. Jogesh Babu. R Programming environment Introduction to R programming language R is an integrated suite of software facilities for data manipulation,
Bayesian analysis of a conceptual transpiration model with a comparison of canopy conductance sub-models Sudeep Samanta Department of Forest Ecology and.
Stochasticity and Probability. A new approach to insight Pose question and think of the answer needed to answer it. Ask: How do the data arise? What is.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Markov Chain Monte Carlo in R
Hierarchical Models.
MCMC Output & Metropolis-Hastings Algorithm Part I
Module 2: Bayesian Hierarchical Models
Introducing Bayesian Approaches to Twin Data Analysis
Model Inference and Averaging
Ch3: Model Building through Regression
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
Predictive distributions
Sampling Distribution
Sampling Distribution
Multidimensional Integration Part I
Filtering and State Estimation: Basic Concepts
Ch13 Empirical Methods.
CS639: Data Management for Data Science
Bayesian Statistics on a Shoestring Assaf Oron, May 2008
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Yalchin Efendiev Texas A&M University
Uncertainty Propagation
Presentation transcript:

Hierarchical Models

Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis? Estimators for hierarchical models – Maximum Likelihood Estimators – Bayesian estimation (MCMC)

Hierarchical Models Hierarchical model, multi-level model, random effects model, repeated measures, randomized blocks, sub-sampling Characterized by random processes at one-level being a function of random processes at another level Typically involve a nested/hierarchical structure

Simple representation  β  Non-hierarchical Hierarchical Single random process = process generating the data (stat. distribution) True state of Nature (fixed) Y: Data (variable) Additional random process = process generating Z Z is variable (stat. distribution) Y (variable) Z (variable) higher level (fixed) parameter Y is only function of a fixed parameter:  Y is function of , but also conditional on a R.V (Z)

Typical Hierarchical Model for Population Analysis We can often divide data generating mechanisms hierarchically Ecological Process: survival, occupancy, abundance Observation Process: detection, mis-classification, sampling error Data: counts, recapture history, # of detections for a patch  ECOLOGICAL PROCESS  OBSERVATION PROCESS Z Y

Example of a Random Effects Model Interest in counts (Y) of migrating shorebirds: several sources of variation (at different levels) – Refuges differ in ‘mean abundance’ ψ (process variation) – Ponds within refuges differ in abundance (f[ N|ψ]) (process variation) – Sampling error in counts (Y), due to imperfect detection (  ): f[Y| N  ]  Counts REFUGES PONDS Y Abundance  

Statistical model Joint probability distribution:  Data     Data

Statistical model Joint probability distribution:  Data     Data Likelihood: In simple cases we can write out the joint distribution and calculate MLE by optimization Bayesian: (1) Integrate to obtain  |Data] (2) Marginalize posterior to obtain  |Data] (likelihood fct) (3) Simulate using Markov Chain Monte Carlo

Simple Likelihood Example We go to a site 5 times and record whether a species is present Ecological Process: Occupancy f(z| ψ) ~ Bern(ψ) Observation Process: Probability of detecting the species y times is conditional on whether the species was present f(y|z) ~ Bin(p*z,5) p  DATA (times detected in 5 visits) OCCUPANCY DETECTION Y

Statistical model Joint probability distribution: f  y  p  f  p  f  p  DATA OCCUPANCY DETECTION Y  bin(y|1*p,5) + (1-  bin(y|0*p,5)

More complicated example Let’s say p varies around a mean value for a set of sites and that ψ varies annually around a mean value The full model is the joint distribution of all the parts: f(y|p, ψ, a,b, c,d) Y ~ bin(z*p,5) p ~ beta(a,b) z ~ bern( ψ ) ψ ~ beta(c,d) MLE methods rapidly become difficult so we typically turn to a Bayesian approach (MCMC estimation)

Markov Chain Monte Carlo MCMC is simulation based approximation of posterior distributions using dependent sequences of random variables. Bayesian (need to specify priors) Simulation (sample from the posterior distribution) Prior Knowledge + Insights from Data= Basis of Inference

For the sake of estimation, the model can be specified by its components Y ~ bin(z*p,5) p ~ beta(a,b) z ~ bern( ψ ) ψ ~ beta(c,d) And priors a ~ f( θ a ) b ~ f( θ b ) c ~ f( θ c ) d ~ f( θ d )

Summary of MCMC Metropolis-Hastings is one approach to MCMC (omnibus and easily described, but not terribly efficient) Best method: Gibbs sampling (many flavors) BUGS (WinBUGS, OpenBUGS, JAGS) automatically selects appropriate algorithm Issues: Starting values Convergence of MC markov chains Autocorrelation MCMC is simulation based evaluation of posterior distributions using dependent sequences of random variables.

Why consider MCMC Flexible approach for fitting models typical of ecological systems (hierarchical) Is intuitive Random-effects, spatial correlation, partially observed covariates can all readily be incorporated Calculating derived parameters and their uncertainty When priors are “non-informative” or data is abundant answers similar to MLE methods Naturally incorporate prior knowledge/belief