Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.

Slides:



Advertisements
Similar presentations
Case studies in Gaussian process modelling of computer codes for carbon accounting Marc Kennedy, Clive Anderson, Stefano Conti, Tony OHagan.
Advertisements

Bayesian tools for analysing and reducing uncertainty Tony OHagan University of Sheffield.
Quantifying and managing uncertainty with Gaussian process emulators Tony O’Hagan University of Sheffield.
Insert Date HereSlide 1 Using Derivative and Integral Information in the Statistical Analysis of Computer Models Gemma Stephenson March 2007.
Getting started with GEM-SA Marc Kennedy. This talk  Starting GEM-SA program  Creating input and output files  Explanation of the menus, toolbars,
Pattern Recognition and Machine Learning
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
Integration of sensory modalities
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Model Assessment, Selection and Averaging
Validating uncertain predictions Tony O’Hagan, Leo Bastos, Jeremy Oakley, University of Sheffield.
Gaussian Processes I have known
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Gaussian process emulation of multiple outputs Tony O’Hagan, MUCM, Sheffield.
Computational statistics 2009 Sensitivity analysis (SA) - outline 1.Objectives and examples 2.Local and global SA 3.Importance measures involving derivatives.
Constraining Astronomical Populations with Truncated Data Sets Brandon C. Kelly (CfA, Hubble Fellow, 6/11/2015Brandon C. Kelly,
A Two Level Monte Carlo Approach To Calculating
Prediction and model selection
End of Chapter 8 Neil Weisenfeld March 28, 2005.
1 Transforming the efficiency of Partial EVSI computation Alan Brennan Health Economics and Decision Science (HEDS) Samer Kharroubi Centre for Bayesian.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Value of Information for Complex Economic Models Jeremy Oakley Department of Probability and Statistics, University of Sheffield. Paper available from.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Lecture II-2: Probability Review
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Modeling Menstrual Cycle Length in Pre- and Peri-Menopausal Women Michael Elliott Xiaobi Huang Sioban Harlow University of Michigan School of Public Health.
Gaussian process modelling
Calibration and Model Discrepancy Tony O’Hagan, MUCM, Sheffield.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Model Inference and Averaging
1 Institute of Engineering Mechanics Leopold-Franzens University Innsbruck, Austria, EU H.J. Pradlwarter and G.I. Schuëller Confidence.
29 May 2008IMA Scottish Branch1 Quantifying uncertainty in the UK carbon flux Tony O’Hagan University of Sheffield.
Using an emulator. Outline So we’ve built an emulator – what can we use it for? Prediction What would the simulator output y be at an untried input x?
Learning Lateral Connections between Hidden Units Geoffrey Hinton University of Toronto in collaboration with Kejie Bao University of Toronto.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Slide 1 Marc Kennedy, Clive Anderson, Anthony O’Hagan, Mark Lomas, Ian Woodward, Andreas Heinemayer and John Paul Gosling Uncertainty in environmental.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Why it is good to be uncertain ? Martin Wattenbach, Pia Gottschalk, Markus Reichstein, Dario Papale, Jagadeesh Yeluripati, Astley Hastings, Marcel van.
17 May 2007RSS Kent Local Group1 Quantifying uncertainty in the UK carbon flux Tony O’Hagan CTCD, Sheffield.
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
Slide 1 Marc Kennedy, Clive Anderson, Anthony O’Hagan, Mark Lomas, Ian Woodward, Andreas Heinemayer and John Paul Gosling Quantifying uncertainty in the.
Additional Topics in Prediction Methodology. Introduction Predictive distribution for random variable Y 0 is meant to capture all the information about.
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
- 1 - Calibration with discrepancy Major references –Calibration lecture is not in the book. –Kennedy, Marc C., and Anthony O'Hagan. "Bayesian calibration.
Options and generalisations. Outline Dimensionality Many inputs and/or many outputs GP structure Mean and variance functions Prior information Multi-output,
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Introduction to emulators Tony O’Hagan University of Sheffield.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
CSC321: Lecture 8: The Bayesian way to fit models Geoffrey Hinton.
8 Sept 2006, DEMA2006Slide 1 An Introduction to Computer Experiments and their Design Problems Tony O’Hagan University of Sheffield.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Marc Kennedy, Tony O’Hagan, Clive Anderson,
Ch3: Model Building through Regression
Ecosystem Demography model version 2 (ED2)
Overview of Supervised Learning
Filtering and State Estimation: Basic Concepts
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
10701 / Machine Learning Today: - Cross validation,
Integration of sensory modalities
LECTURE 09: BAYESIAN LEARNING
LECTURE 07: BAYESIAN ESTIMATION
Parametric Methods Berlin Chen, 2005 References:
Multivariate Methods Berlin Chen
Mathematical Foundations of BME Reza Shadmehr
Uncertainty Propagation
Presentation transcript:

Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive Anderson, Stefano Conti, Tony O’Hagan Probability & Statistics, University of Sheffield

Outline Uncertainties in computer simulators Uncertainties in computer simulators Bayesian inference about simulator outputs Bayesian inference about simulator outputs –Creating an emulator for the simulator –Deriving uncertainty and sensitivity measures Example application Example application Some recent extensions Some recent extensions

Uncertainties in computer simulators Consider a complex deterministic code with a vector of inputs and single output Consider a complex deterministic code with a vector of inputs and single output Use of the code is subject to: Use of the code is subject to: –Input uncertainty –Code uncertainty

Input uncertainty The inputs to the simulator are unknown for a given real world scenario The inputs to the simulator are unknown for a given real world scenario Therefore the true value of the output is uncertain Therefore the true value of the output is uncertain A Monte Carlo approach is often used to take this uncertainty into account A Monte Carlo approach is often used to take this uncertainty into account –Sample from the probability distribution of X –Run the simulator for each point in the sample to give a sample from the distribution of Y –Very inefficient…not practical for complex codes

Code uncertainty The code output at a given input point is unknown until we run it at that point The code output at a given input point is unknown until we run it at that point –In practice codes can take hours or days to run, so we have a limited number of runs We have some prior beliefs about the output We have some prior beliefs about the output –Smooth function of the inputs

Bayesian inference about simulator outputs Bayesian solution involves building an emulator Bayesian solution involves building an emulator Highly efficient Highly efficient –Makes maximum use of all available information –A single set of simulator runs is required to train the emulator. All sensitivity and uncertainty information is derived directly from this –The inputs for these runs can be chosen to give good information about the simulator output A natural way to treat the different uncertainties within a coherent framework A natural way to treat the different uncertainties within a coherent framework

Inference about functions using Gaussian processes We model as an unknown function having a Gaussian process prior distribution We model as an unknown function having a Gaussian process prior distribution h(.) is a vector of regression functions and are unknown coefficients h(.) is a vector of regression functions and are unknown coefficients Prior expectation of the model output as a function of the inputs

Inference about functions using Gaussian processes We model as an unknown function having a Gaussian process prior distribution We model as an unknown function having a Gaussian process prior distribution c(.,.) is a correlation function, which defines our beliefs about smoothness of the output and is the GP variance c(.,.) is a correlation function, which defines our beliefs about smoothness of the output and is the GP variance Prior beliefs about covariance between model outputs

Choice of correlation function We use the product of univariate Gaussian functions: We use the product of univariate Gaussian functions: Where is a measure of the roughness of the function in the kth input Where is a measure of the roughness of the function in the kth input

roughness = 0.5

roughness = 0.2

roughness = 0.1

roughness = 0.01

Conditioning on code runs Conditional on the observed set of training runs, Conditional on the observed set of training runs, is still a Gaussian process, with simple analytical forms for the posterior mean and covariance functions is still a Gaussian process, with simple analytical forms for the posterior mean and covariance functions

2 code runs

Large b Small b

3 code runs

5 code runs

More about the emulator The emulator mean is an estimate of the model output and can be used as a surrogate The emulator mean is an estimate of the model output and can be used as a surrogate The emulator is much more… The emulator is much more… –It is a probability distribution for the whole function –This allows us to derive inferences for many output related quantities, particularly integrals

Inference for integrals For particular forms of input distribution (Gaussian or uniform), analytical forms have been derived for integration-based sensitivity measures For particular forms of input distribution (Gaussian or uniform), analytical forms have been derived for integration-based sensitivity measures –Main effects of individual inputs –Joint effects of pairs of inputs –Sensitivity indices

Example Application

Sheffield Dynamic Global Vegetation Model (SDGVM) Developed within the Centre for Terrestrial Carbon Dynamics Developed within the Centre for Terrestrial Carbon Dynamics Our job with SDGVM is to: Our job with SDGVM is to: –Apply sensitivity analysis for model testing –Identify the greatest sources of uncertainty –Correctly reflect the uncertainty in predictions

Plant respiration Photosynthesis Loss Soil respiration Loss – Terrestrial carbon source if NEP is negative – Terrestrial carbon sink if NEP is positive Net Ecosystem Production (CARBON FLUX)

Some Inputs Parameters Leaf life span Leaf life span Leaf area Leaf area Budburst temperature Budburst temperature Senescence temperature Senescence temperature Wood density Wood density Maximum carbon storage Maximum carbon storage Xylem conductivity Xylem conductivity Soil clay % Soil sand % Soil depth Soil bulk density

Main Effect: Leaf life span If leaves die young, NEP is predicted to be higher, on average. Why?

Main Effect: Leaf life span (updated) If leaves die young, SDGVM allowed a second growing season, resulting in increased carbon uptake. This problem was fixed by the modellers

Main Effect: Senescence Temperature Small values mean the leaves stay until the temperature is very low Large values mean the leaves drop earlier, so reduce the growing season

When soil bulk density was added to the active parameter set, the Gaussian Process model did not fit the training data properly

Error discovered in the soil module NEP Before…After… Bulk density

Our GP model depends on the output being a smooth function of the inputs. The problem was again fixed by the modellers

SDGVM: new sensitivity analysis Extended sensitivity analysis to 14 input parameters (using a more stable version) Extended sensitivity analysis to 14 input parameters (using a more stable version) Assumed uniform probability distributions for each of the parameters Assumed uniform probability distributions for each of the parameters The aim here is to identify the greatest potential sources of uncertainty The aim here is to identify the greatest potential sources of uncertainty

NEP (g/m 2 /y)

by investing effort to learn more about this parameter, output uncertainty could be significantly reduced Leaf life span 69.1% by investing effort to learn more about this parameter, output uncertainty could be significantly reduced Minimum growth rate 14.2% Water potential 3.4% Maximum age 1.0% Percentage of total output variance

Extensions to the theory

Multiple outputs So far we have created independent emulators for each output So far we have created independent emulators for each output –Ignores information about the correlation between outputs We are experimenting with simple models linking the outputs together We are experimenting with simple models linking the outputs together This is an important first step in treating dynamic emulators and in aggregating code outputs This is an important first step in treating dynamic emulators and in aggregating code outputs

Dynamic emulators Physical systems typically evolve over time Physical systems typically evolve over time Their behaviour is modelled via dynamic codes Their behaviour is modelled via dynamic codes –where x are tuning constants and z t are context- specific drivers –Recursive emulation of y t over the appropriate time span shows promising results

CENTURY output ( ) and dynamic emulator ( )

Aggregating outputs Motivated by the UK carbon budget problem Motivated by the UK carbon budget problem –The total UK carbon absorbed by vegetation is a sum of individual pixels/sites –Each site has a different set of input parameters (e.g. vegetation/soil properties), but some of these are correlated This is a multiple output code This is a multiple output code –Each site represents a different output Bayesian uncertainty analysis is being extended, to make inference about the sum Bayesian uncertainty analysis is being extended, to make inference about the sum

References For Bayesian analysis of computer models: For Bayesian analysis of computer models: –Kennedy, M. C. and O’Hagan, A. (2001). Bayesian calibration of computer models (with discussion) J. Roy. Statist. Soc. B, 63: For Bayesian Sensitivity analysis: For Bayesian Sensitivity analysis: –Oakley, J. E. and O’Hagan, A. (2004). Probabilistic sensitivity analysis of complex models: A Bayesian approach. J. Roy. Statist. Soc. B, 66: