Bayesian calibration and comparison of process-based forest models Marcel van Oijen & Ron Smith (CEH-Edinburgh) Jonathan Rougier (Durham Univ.)

Slides:



Advertisements
Similar presentations
Bayesian methods for calibrating and comparing process-based vegetation models Marcel van Oijen (CEH-Edinburgh)
Advertisements

Process-based modelling of vegetations and uncertainty quantification Marcel van Oijen (CEH-Edinburgh) Course Statistics for Environmental Evaluation Glasgow,
Process-based modelling of vegetations and uncertainty quantification Marcel van Oijen (CEH-Edinburgh) Statistics for Environmental Evaluation Glasgow,
Bayesian calibration and uncertainty analysis of dynamic forest models Marcel van Oijen CEH-Edinburgh.
1 Uncertainty in rainfall-runoff simulations An introduction and review of different techniques M. Shafii, Dept. Of Hydrology, Feb
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
1 Bayesian methods for parameter estimation and data assimilation with crop models David Makowski and Daniel Wallach INRA, France September 2006.
Bayesian Estimation in MARK
Inspiral Parameter Estimation via Markov Chain Monte Carlo (MCMC) Methods Nelson Christensen Carleton College LIGO-G Z.
By Addison Euhus, Guidance by Edward Phillips An Introduction To Uncertainty Quantification.
Lectures on Stochastic System Analysis and Bayesian Updating June 29-July James L. Beck, California Institute of Technology Jianye Ching, National.
Part 24: Bayesian Estimation 24-1/35 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Markov-Chain Monte Carlo
Bayesian statistics – MCMC techniques
BAYESIAN INFERENCE Sampling techniques
Computing the Posterior Probability The posterior probability distribution contains the complete information concerning the parameters, but need often.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
CENTER FOR BIOLOGICAL SEQUENCE ANALYSIS Bayesian Inference Anders Gorm Pedersen Molecular Evolution Group Center for Biological Sequence Analysis Technical.
USE OF LAPLACE APPROXIMATIONS TO SIGNIFICANTLY IMPROVE THE EFFICIENCY
1 Transforming the efficiency of Partial EVSI computation Alan Brennan Health Economics and Decision Science (HEDS) Samer Kharroubi Centre for Bayesian.
CENTER FOR BIOLOGICAL SEQUENCE ANALYSIS Bayesian Inference Anders Gorm Pedersen Molecular Evolution Group Center for Biological Sequence Analysis Technical.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
G. Cowan Lectures on Statistical Data Analysis Lecture 14 page 1 Statistical Data Analysis: Lecture 14 1Probability, Bayes’ theorem 2Random variables and.
Bayesian Analysis for Extreme Events Pao-Shin Chu and Xin Zhao Department of Meteorology School of Ocean & Earth Science & Technology University of Hawaii-
Lecture II-2: Probability Review
Robin McDougall, Ed Waller and Scott Nokleby Faculties of Engineering & Applied Science and Energy Systems & Nuclear Science 1.
Bayes Factor Based on Han and Carlin (2001, JASA).
WSEAS AIKED, Cambridge, Feature Importance in Bayesian Assessment of Newborn Brain Maturity from EEG Livia Jakaite, Vitaly Schetinin and Carsten.
Correlation With Errors-In-Variables3/28/20021 Correlation with Errors-In-Variables and an Application to Galaxies William H. Jefferys University of Texas.
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Priors, Normal Models, Computing Posteriors
Exam I review Understanding the meaning of the terminology we use. Quick calculations that indicate understanding of the basis of methods. Many of the.
EFFICIENT CHARACTERIZATION OF UNCERTAINTY IN CONTROL STRATEGY IMPACT PREDICTIONS EFFICIENT CHARACTERIZATION OF UNCERTAINTY IN CONTROL STRATEGY IMPACT PREDICTIONS.
Computer Science, Software Engineering & Robotics Workshop, FGCU, April 27-28, 2012 Fault Prediction with Particle Filters by David Hatfield mentors: Dr.
Modelling protocols Marcel van Oijen (CEH-Edinburgh) With input from participants Garmisch Mar & May 2006, Protocol-GMP team, Protocol-UQ/UA team.
Why it is good to be uncertain ? Martin Wattenbach, Pia Gottschalk, Markus Reichstein, Dario Papale, Jagadeesh Yeluripati, Astley Hastings, Marcel van.
17 May 2007RSS Kent Local Group1 Quantifying uncertainty in the UK carbon flux Tony O’Hagan CTCD, Sheffield.
Bayesian Reasoning: Tempering & Sampling A/Prof Geraint F. Lewis Rm 560:
CAMELS CCDAS A Bayesian approach and Metropolis Monte Carlo method to estimate parameters and uncertainties in ecosystem models from eddy-covariance data.
G. Cowan RHUL Physics Bayesian Higgs combination page 1 Bayesian Higgs combination based on event counts (follow-up from 11 May 07) ATLAS Statistics Forum.
1 Bayesian Essentials Slides by Peter Rossi and David Madigan.
Application of the MCMC Method for the Calibration of DSMC Parameters James S. Strand and David B. Goldstein The University of Texas at Austin Sponsored.
MCMC reconstruction of the 2 HE cascade events Dmitry Chirkin, UW Madison.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
Reducing MCMC Computational Cost With a Two Layered Bayesian Approach
Adaptive Spatial Resampling as a McMC Method for Uncertainty Quantification in Seismic Reservoir Modeling Cheolkyun Jeong*, Tapan Mukerji, and Gregoire.
Bayesian Modelling Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
Uncertainty analysis of carbon turnover time and sequestration potential in terrestrial ecosystems of the Conterminous USA Xuhui Zhou 1, Tao Zhou 1, Yiqi.
Introduction: Metropolis-Hasting Sampler Purpose--To draw samples from a probability distribution There are three steps 1Propose a move from x to y 2Accept.
Markov-Chain-Monte-Carlo (MCMC) & The Metropolis-Hastings Algorithm P548: Intro Bayesian Stats with Psych Applications Instructor: John Miyamoto 01/19/2016:
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
SIR method continued. SIR: sample-importance resampling Find maximum likelihood (best likelihood × prior), Y Randomly sample pairs of r and N 1973 For.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Artificial Intelligence as a Technology of Approximation Derek Partridge, University of Exeter, UK
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
CS498-EA Reasoning in AI Lecture #19 Professor: Eyal Amir Fall Semester 2011.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Markov Chain Monte Carlo in R
MCMC Output & Metropolis-Hastings Algorithm Part I
Probability Theory and Parameter Estimation I
ICS 280 Learning in Graphical Models
Special Topics In Scientific Computing
Markov Chain Monte Carlo
CSCI 5822 Probabilistic Models of Human and Machine Learning
CS 188: Artificial Intelligence
Yalchin Efendiev Texas A&M University
Uncertainty Propagation
Presentation transcript:

Bayesian calibration and comparison of process-based forest models Marcel van Oijen & Ron Smith (CEH-Edinburgh) Jonathan Rougier (Durham Univ.)

ContentsContents 1.Bayesian calibration of a forest model 2.… what measurements to take? 3.Bayesian comparison of forest models

1. Bayesian calibration of a forest model

Process-based forest models Soil C NPP Height Environmental scenarios Initial values Parameters Model

Using data We need a method that: 1.Quantifies how uncertainty about inputs and model structure causes output uncertainty 2.Efficiently uses data, on inputs & outputs, to reduce uncertainties

BASic FORest model (BASFOR) BASFOR 12 output variables 39 parameters

BASFOR: Inputs BASFOR 12 output variables

BASFOR: Prior pdf for parameters

Example: Simulating growth of Norway spruce Skogaby

BASFOR: Prior predictive uncertainty Height Biomass Prior uncertainty for Skogaby

BASFOR: Predictive uncertainty BASFOR 12 output variables High output uncertainty 39 parameters High input uncertainty Data: measurements of output variables Calibration of parameters

Bayes’ Theorem P( |D) = P( ) P(D| ) / P(D)  P( ) P(D|f( )) “Posterior distribution of parameters” “Prior distribution of parameters” “Likelihood” of data, given mismatch with model output f = the model, e.g. BASFOR

Finding the posterior: MCMC MCMC: walk through parameter-space →set of visited points approaches the posterior parameter distribution P( |D) [e.g. using Metropolis-Hastings random walk] Sample of parameter vectors from the posterior distribution P( |D) for the parameters P( |D)  P( ) P(D|f( ))

MCMC: Metropolis-Hastings random walk Metropolis (1953) algorithm 1.Start anywhere in parameter-space: p (i=0) 2.Randomly choose p(i+1) = p(i) + δ 3.IF:[ P(p(i+1)) P(D|f(p(i+1))) ] / [ P(p(i)) P(D|f(p(i))) ] > Random[0,1] THEN: accept p(i+1) ELSE: reject p(i+1) i=i+1 4.IF i < 10 4 GOTO 2 Sample of parameter vectors from the posterior distribution P( |D) for the parameters

Forest data from Skogaby (Sweden) Planted: 1966, (2300 trees ha -1 ) Weather data: Soil data: C, N, Mineralisation rate Tree data: Biomass, NPP, Height, [N], LAI Skogaby

BASFOR: Prior predictive uncertainty Height Biomass Prior pred. uncertainty Data Skogaby Data: Göran Ågren

MCMC parameter trace plots: steps Steps in MCMC Param. value

Posterior marginal distributions for parameters

Parameter correlations 39 parameters

Bayesian calibration: overview Data Bayesian calibration

Prior & posterior predictive uncertainty Height Biomass Prior pred. uncertainty Posterior uncertainty (using data Skogaby)

Partial corr. coefficients (PCC) parameters – outputs 12 output variables 39 parameters

2. What kind of measurements would have reduced uncertainty the most?

Prior predictive uncertainty & height-data Height Biomass Prior pred. uncertainty Height data Skogaby

Prior & posterior uncertainty: use of height data Height Biomass Prior pred. uncertainty Posterior uncertainty (using height data) Height data Skogaby

Prior & posterior uncertainty: use of height data Height Biomass Prior pred. uncertainty Posterior uncertainty (using height data) Height data (hypothet.)

Prior & posterior uncertainty: use of height data Height Biomass Prior pred. uncertainty Posterior uncertainty (using height data) Posterior uncertainty (using precision height data)

Summary of procedure Data D ± σModel f Prior P( ) Calibrated parameters, with covariances Uncertainty of model output Sensitivity analysis of model parameters “Error function” e.g. N(0, σ) MCMC Samples of (10 4 – 10 5 ) Samples of f( ) (10 4 – 10 5 ) P(D|f( )) Posterior P( |D) PCC

3. Bayesian comparison of forest models

Uncertainty regarding model structure Soil C NPP Height Environmental scenarios Initial values Parameters Model Imperfect understanding

Bayesian comparison of two models Bayes Theorem for model probab.: P(M|D) = P(M) P(D|M) / P(D) The “Integrated likelihood” P(D|M i ) can be approximated from the MCMC sample of outputs for model M i ( * ) Model 1 Model 2 P(M 2 |D) / P(M 1 |D) = P(D|M 2 ) / P(D|M 1 ) The “Bayes Factor” P(D|M 2 ) / P(D|M 1 ) quantifies how the data D change the odds of M 2 over M 1 P(M 1 ) = P(M 2 ) = ½ (*)(*) harmonic mean of likelihoods in MCMC-sample (Kass & Raftery, 1995)

Bayes Factor for two big forest models MCMC 5000 steps Calculation of P(D|BASFOR) Calculation of P(D|BASFOR+) Data Rajec: Emil Klimo

Bayes Factor for two big forest models MCMC 5000 steps Calculation of P(D|BASFOR) Calculation of P(D|BASFOR+) Data Rajec: Emil Klimo P(D|M 1 ) = 7.2e-016 P(D|M 2 ) = 5.8e-15 Bayes Factor = 7.8, so BASFOR+ supported by the data

Summary of procedure Data D Prior P( 1 ) Updated parameters MCMC Samples of 1 (10 4 – 10 5 ) Posterior P( 1 |D) Model 1 MCMC Prior P( 2 ) Model 2 Samples of 2 (10 4 – 10 5 ) Posterior P( 2 |D) Updated parameters P(D|M 1 )P(D|M 2 ) Bayes factor Updated model odds

ConclusionsConclusions Bayesian calibration using MCMC: Improves model predictive capacity, by updating parameters Quantifies uncertainty in parameters and output Forest model calibration: Benefits from high-precision tree height measurement Bayesian model comparison: Same probabilistic approach as Bayesian calibration Bayes Factor shows how new data change the odds of models Aid in model development

AppendicesAppendices

Bayesian calibration of big models P( |D)  P( ) P(D|f( )) Calculating P( |D) costs much time: 1.Sample parameter-space representatively 2.For each sampled set of parameter-values: a.Calculate P( ) b.Run the model to calculate likelihood P(D|f( )) Sampling problem: Markov Chain Monte Carlo (MCMC) methods Computing problem: increased processor speed Solutions

Bayes Factor for two big forest models BASFOR 39 params BASFOR + 41 params (Penman eq., corrections) MCMC steps Calibration

Bayesian methods Bayes, T. (1763) Metropolis, N. (1953) Green, E.J. / MacFarlane, D.W. / Valentine, H.T., Strawderman, W.E. (1996, 1998, 1999, 2000) Jansen, M. (1997) Jaynes, E.T. (2003) Bayes’ Theorem MCMC Forest models Crop models Probability theory