Calibration and Model Discrepancy Tony O’Hagan, MUCM, Sheffield.

Slides:



Advertisements
Similar presentations
Bayesian tools for analysing and reducing uncertainty Tony OHagan University of Sheffield.
Advertisements

The Maximum Likelihood Method
Interfacing physical experiments and computer models Preliminary remarks Tony O’Hagan.
Welcome to PHYS 225a Lab Introduction, class rules, error analysis Julia Velkovska.
Copyright © 2010 Pearson Education, Inc. Chapter 27 Inferences for Regression.
Copyright © 2009 Pearson Education, Inc. Chapter 29 Multiple Regression.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Simple Linear Regression. Start by exploring the data Construct a scatterplot  Does a linear relationship between variables exist?  Is the relationship.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Chapter 8 Linear Regression © 2010 Pearson Education 1.
Validating uncertain predictions Tony O’Hagan, Leo Bastos, Jeremy Oakley, University of Sheffield.
Gaussian process emulation of multiple outputs Tony O’Hagan, MUCM, Sheffield.
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Multiple regression analysis
Regression Analysis. Unscheduled Maintenance Issue: l 36 flight squadrons l Each experiences unscheduled maintenance actions (UMAs) l UMAs costs $1000.
4. Multiple Regression Analysis: Estimation -Most econometric regressions are motivated by a question -ie: Do Canadian Heritage commercials have a positive.
The Simple Regression Model
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
Bootstrapping applied to t-tests
Simple Linear Regression
Correlation & Regression
Session 3: Calibration Using observations of the real process.
Inference for regression - Simple linear regression
Engineering subprogramme, 7 November 2006 Tony O’Hagan.
Gaussian process modelling
Inferences for Regression
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
Model Inference and Averaging
Basic linear regression and multiple regression Psych Fraley.
Review of Statistical Models and Linear Regression Concepts STAT E-150 Statistical Methods.
R. Kass/W03P416/Lecture 7 1 Lecture 7 Some Advanced Topics using Propagation of Errors and Least Squares Fitting Error on the mean (review from Lecture.
Ch4 Describing Relationships Between Variables. Pressure.
Correlation and Linear Regression. Evaluating Relations Between Interval Level Variables Up to now you have learned to evaluate differences between the.
Using an emulator. Outline So we’ve built an emulator – what can we use it for? Prediction What would the simulator output y be at an untried input x?
Ch4 Describing Relationships Between Variables. Section 4.1: Fitting a Line by Least Squares Often we want to fit a straight line to data. For example.
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
17 May 2007RSS Kent Local Group1 Quantifying uncertainty in the UK carbon flux Tony O’Hagan CTCD, Sheffield.
Stat 112: Notes 2 Today’s class: Section 3.3. –Full description of simple linear regression model. –Checking the assumptions of the simple linear regression.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Statistics : Statistical Inference Krishna.V.Palem Kenneth and Audrey Kennedy Professor of Computing Department of Computer Science, Rice University 1.
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
Correlation & Regression Analysis
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 8- 1.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
Machine Learning 5. Parametric Methods.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
More on regression Petter Mostad More on indicator variables If an independent variable is an indicator variable, cases where it is 1 will.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Chapter 26 Inferences for Regression. An Example: Body Fat and Waist Size Our chapter example revolves around the relationship between % body fat and.
Copyright © Cengage Learning. All rights reserved. 8 9 Correlation and Regression.
4-1 MGMG 522 : Session #4 Choosing the Independent Variables and a Functional Form (Ch. 6 & 7)
11-1 Copyright © 2014, 2011, and 2008 Pearson Education, Inc.
CORRELATION-REGULATION ANALYSIS Томский политехнический университет.
Introduction to emulators Tony O’Hagan University of Sheffield.
Describing Relationships. Least-Squares Regression  A method for finding a line that summarizes the relationship between two variables Only in a specific.
Remember the equation of a line: Basic Linear Regression As scientists, we find it an irresistible temptation to put a straight line though something that.
8 Sept 2006, DEMA2006Slide 1 An Introduction to Computer Experiments and their Design Problems Tony O’Hagan University of Sheffield.
The simple linear regression model and parameter estimation
Regression Analysis AGEC 784.
Chapter 11: Simple Linear Regression
Statistical Methods For Engineers
Introduction to Instrumentation Engineering
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Simple Linear Regression
Simple Linear Regression
Presentation transcript:

Calibration and Model Discrepancy Tony O’Hagan, MUCM, Sheffield

Outline  Why model discrepancy  The meaning of parameters  Modelling discrepancy  Conclusions 9/9/2011SAMSI UQ Program: Methodology Workshop -2

Why model discrepancy Is calibration even possible? 9/9/2011SAMSI UQ Program: Methodology Workshop - 3

The calibration problem 9/9/2011SAMSI UQ Program: Methodology Workshop - 4  The problem  We have a simulator that predicts a real world phenomenon  We have some observations of the real world  We want to use those to learn about some unknown parameters  Formally, the simulator takes two kinds of inputs  The calibration parameters θ  Control inputs x  Simulator is written y = f(x, θ )  Observation z i is obtained at control input values x i

Traditional formulation 9/9/2011SAMSI UQ Program: Methodology Workshop - 5  Write z i = f(x i, θ ) + ε i  where ε i are independent observation errors  Estimate θ, e.g. by minimising sum of squared residuals  Call estimate t and predict real world process at a new x value by f(x, t)  Two things wrong with this  The formulation is wrong because the simulation model is inevitably wrong  So errors are not independent  Traditional calibration ignores uncertainty about θ  Treats it as now known to equal t

Little example 9/9/2011SAMSI UQ Program: Methodology Workshop - 6  One control input, one calibration parameter  Three observations marked with crosses  Red lines are possible simulator outputs  Calibration parameter just changes slope of line  No value of the calibration parameter gets close to all the observations  And yet they seem to lie on a straight line X X X x

Is calibration even possible? 9/9/2011SAMSI UQ Program: Methodology Workshop - 7  Green line is best fit  Minimises sum of squared residuals  Red line seems better  But with constant bias  Green seems to be over-fitting  Errors don’t look independent  Can we learn the true value of the calibration parameter?  With more data  Keeping close to a straight line  Over a different range of x X X X x

Model discrepancy 9/9/2011SAMSI UQ Program: Methodology Workshop - 8  The little example suggests that we need to allow that the model does not correctly represent reality  For any values of the calibration parameters  The simulator outputs deviate systematically from reality  Call it model bias or model discrepancy  It is claimed that acknowledging model discrepancy may allow us to achieve more realistic and appropriate estimates of calibration parameters  But to evaluate that claim, look at a simpler problem

The meaning of parameters What can we learn from simple models? 9/9/2011SAMSI UQ Program: Methodology Workshop - 9

What do parameters mean?  All models are wrong  “All models are wrong but some are useful”  George E P Box, 1979  So, what does a parameter (in an admittedly wrong model) mean?  How do we specify prior information about a parameter  when we know the model is wrong?  What have we learnt when we make inference about a parameter  in a model we know to be wrong? 9/9/2011SAMSI UQ Program: Methodology Workshop - 10

Example: Poisson sample  Suppose we model data as a sample from a Poisson distribution with parameter λ  We need a prior distribution  Do we ask for beliefs about the population mean or variance?  Or about the proportion p of zeros in the population?  And infer a prior for λ = – log p  Given that the Poisson assumption is wrong these are all asking about different things  When we derive a posterior distribution for λ  Is it a belief statement about the population mean or variance?  Or even about anything real? 9/9/2011SAMSI UQ Program: Methodology Workshop - 11

Example: Linear regression  Suppose we assume a simple linear regression model with slope β and intercept α  We are interested in the strength of the relationship as represented by β  But the model is wrong; the relationship is not truly linear  We know if we switch to a quadratic model the coefficient of x will change  If we assume a linear relationship when the truth is different, e.g. quadratic, the slope will depend on the range of x over which we fit  How can we elicit prior beliefs about such a parameter?  What do inferences about it mean? 9/9/2011SAMSI UQ Program: Methodology Workshop - 12

Example: A simple machine  A machine produces an amount of work y which depends on the amount of effort t put into it  Model is y = β t + ε  Where β is the rate at which effort is converted to work  And ε is observation error, assumed iid  True value of β is 0.65  Graph shows observed data  All points lie below y = 0.65t  Because the model is wrong  Losses due to friction etc.  Fitted slope is /9/2011SAMSI UQ Program: Methodology Workshop - 13

 The true model is shown as the dashed line here  In this example, the efficiency parameter β is physically meaningful  Theoretical value is 0.65  This value is of interest to experimenters  They have genuine prior information  They want the experiment to help them identify this true value  But because of model error the estimate is biased  And given enough data it will over-rule any prior information Simple machine – true model 9/9/2011SAMSI UQ Program: Methodology Workshop - 14

Calibration is just nonlinear regression 9/9/2011SAMSI UQ Program: Methodology Workshop - 15  Returning to the context of computer models y = f(x, θ ) + ε  where f is a computer simulator of some phenomenon  We can view this as just a nonlinear regression model  The regression function f(x, θ ) is complex and we can’t try alternatives (as we would do in regression modelling)  But we have all the same problems as in simple regression  Given that the model is wrong:  What do the calibration parameters θ mean?  We can’t expect to learn their ‘true’ values from observations  Even with unlimited data

Tuning and physical parameters  Simulator parameters may be physical or just for tuning  We adjust tuning parameters so the model fits reality better  We are not really interested in their ‘true’ values  We calibrate tuning parameters for prediction  Physical parameters are different  We are often really interested in true physical values  And we like to think that calibration can help us learn about them  But the model is inevitably wrong, so estimates are distorted  And getting more data does not take us closer to their true values  Calibration to learn about physical parameters is a delusion  Unless … ? 9/9/2011SAMSI UQ Program: Methodology Workshop - 16

Modelling discrepancy Is model discrepancy the answer? 9/9/2011SAMSI UQ Program: Methodology Workshop - 17

Model discrepancy  In the context of computer models, it is necessary to acknowledge model discrepancy  There is a difference between the model with best/true parameter values and reality y = f(x, θ ) + δ (x) + ε  where δ (x) accounts for this discrepancy  Will typically itself have uncertain parameters  Kennedy and O’Hagan (JRSSB, 2001) introduced this model discrepancy  Modelled it as a zero-mean Gaussian process  They claimed it acknowledges additional uncertainty  And mitigates against over-fitting of θ 9/9/2011SAMSI UQ Program: Methodology Workshop - 18

Simple machine revisited  So add this model discrepancy term to the linear model of the simple machine y = β t + δ (t) + ε  With δ (t) modelled as a zero-mean GP  As in Kennedy and O’Hagan  Now the estimate of β is  It’s even further from the true value of 0.65!  Without model discrepancy we got  What’s going on? 9/9/2011SAMSI UQ Program: Methodology Workshop - 19

9/9/2011SAMSI UQ Program: Methodology Workshop - 20  In order to get the right answer we have to infer:  The true value of the efficiency parameter is 0.65  The solid line  Model discrepancy is negative for all t  And more so for larger t  Like the black curve below  But the GP says:  It’s much more likely to be the green curve

Nonidentifiability 9/9/2011SAMSI UQ Program: Methodology Workshop - 21  Formulation with model discrepancy is not identifiable  For any θ, there is a δ (x)  Reality is some function ζ (x) = f(x, θ ) + δ (x)  Given θ, model discrepancy is δ (x) = ζ (x) - f(x, θ )  As in the three curves in the previous example  Suppose we had an unlimited number of observations  We would learn reality’s true function ζ (x) exactly  But we would still not learn θ  It could in principle be anything  In a Bayesian analysis, the prior distribution is used to resolve nonidentifiability

The joint posterior 9/9/2011SAMSI UQ Program: Methodology Workshop - 22  Calibration leads to a joint posterior distribution for θ and δ (x)  But nonidentifiability means there are many equally good fits ( θ, δ (x)) to the data  Induces strong correlation between θ and δ (x)  This may be compounded by the fact that simulators often have large numbers of parameters  (Near-)redundancy means that different θ values produce (almost) identical predictions  Sometimes called equifinality  Within this set, the prior distributions for θ and δ (x) count

Modelling the discrepancy  The nonparametric GP term allows the model to fit and predict reality accurately given enough data  But it doesn’t mean physical parameters are correctly estimated  The separation between original model and discrepancy is unidentified  Estimates depend on prior information  Unless the real model discrepancy is just the kind expected a priori the physical parameter estimates will still be biased  It is necessary to think about the model discrepancy  In the machine example, a prior distribution saying δ (t) is likely to be negative and decreasing will produce a better answer 9/9/2011SAMSI UQ Program: Methodology Workshop - 23

What do I mean by ‘better’? 9/9/2011SAMSI UQ Program: Methodology Workshop - 24  Posterior distribution for θ will typically be quite wide  And won’t become degenerate with infinite data  Values of θ with very low posterior probability will typically either  Have very low prior probability or  Imply a δ (x) with very low prior probability  Assuming we don’t get prior information about θ wrong, it’s important to model δ (x) carefully  Flexibly but informatively  So ‘better’ means getting a posterior distribution that covers the true θ

Reification  In the computer models context, Goldstein and Rougier (JSPI, 2009) introduced a formal mechanism for modelling model discrepancy  Called reification  Based on imagining hypothetical improved model(s)  The reified model is such that we have no knowledge of how it might differ from reality,  So homogeneous zero-mean discrepancy is appropriate  We may also be able to consider the next-generation model in which specific improvements have been made  Their framework may be over-kill, but  if we want parameters to have meaning we have to think seriously about model discrepancy 9/9/2011SAMSI UQ Program: Methodology Workshop - 25

Extrapolation  Here’s an example of how important discrepancy modelling is  Consider prediction using a regression model but for x values far from the data  We often hear how extrapolation is a bad idea, because of model discrepancy  But if δ (x) has finite variance the impact is bounded, no matter how far we extrapolate  For unbounded impact in large extrapolations we could model discrepancy using something like a random walk  Or a localised regression model (O’Hagan, JRSSB, 1978) 9/9/2011SAMSI UQ Program: Methodology Workshop - 26

Conclusions And challenges! 9/9/2011SAMSI UQ Program: Methodology Workshop - 27

Key messages 9/9/2011SAMSI UQ Program: Methodology Workshop - 28  If you don’t include some representation of model discrepancy, you can expect to get nonsense  Posterior distribution of θ will converge to wrong value  Overfitting and spurious accuracy  Even if you do include model discrepancy, it’s essential to think about it and model it carefully  Use knowledge about aspects of reality that are not adequately represented in the model  Even if you do think about it and model it carefully,  You will not be able to learn the true physical values of calibration parameters  Not even with an unlimited number of physical observations  Posterior distribution of θ will not converge  But should cover the true value

Challenges 9/9/2011SAMSI UQ Program: Methodology Workshop - 29  We need to gain much more practical experience of how calibration works when we incorporate model discrepancy  A major challenge is how to model the discrepancy  ‘Flexibly but informatively’

Thanks to...  Colleagues in the MUCM project  Managing Uncertainty in Complex Models  9/9/2011SAMSI UQ Program: Methodology Workshop - 30