Presentation is loading. Please wait.

Presentation is loading. Please wait.

Validating uncertain predictions Tony O’Hagan, Leo Bastos, Jeremy Oakley, University of Sheffield.

Similar presentations


Presentation on theme: "Validating uncertain predictions Tony O’Hagan, Leo Bastos, Jeremy Oakley, University of Sheffield."— Presentation transcript:

1 Validating uncertain predictions Tony O’Hagan, Leo Bastos, Jeremy Oakley, University of Sheffield

2 Why am I here?  I probably know less about finite elements modelling than anyone else at this meeting  But I have been working with mechanistic models of all kinds for almost 20 years  Models of climate, oil reservoirs, rainfall runoff, aero-engines, sewer systems, vegetation growth, disease progression,...  What I do know about is uncertainty  I’m a statistician  My field is Bayesian statistics  One of my principal research areas is to understand, quantify and reduce uncertainty in the predictions made by models  I bring a different perspective on model validation 6/9/20092mucm.group.shef.ac.uk

3 Some background  Models are often highly computer intensive  Long run times  FE models on fine grid  Oil reservoir simulator runs can take days  Things we want to do with them may require many runs  Uncertainty analysis  Exploring output uncertainty induced by uncertainty in model inputs  Calibration  Searching for parameter values to match observational data  Optimisation  Searching for input settings to optimise output  We need efficient methods requiring minimal run sets 6/9/20093mucm.group.shef.ac.uk

4 Emulation  We use Bayesian statistics  Based on a training sample of model runs, we estimate what the model output would be at all untried input configurations  The result is a statistical representation of the model  In the form of a stochastic process over input space  The process mean is our best estimate of what the output would be at any input configuration  Uncertainty is captured by variances and covariances  It correctly returns what we know  At any training sample point, the mean is the observed value  With zero variance 6/9/20094mucm.group.shef.ac.uk

5 2 code runs  Consider one input and one output  Emulator estimate interpolates data  Emulator uncertainty grows between data points mucm.group.shef.ac.uk6/9/20095

6 3 code runs  Adding another point changes estimate and reduces uncertainty mucm.group.shef.ac.uk6/9/20096

7 5 code runs  And so on mucm.group.shef.ac.uk6/9/20097

8 MUCM  The emulator is a fast meta-model but with a full statistical representation of uncertainty  We can build the emulator and use it for tasks such as calibration with far fewer model runs than other methods  Typically 10 or 100 times fewer  The RCUK Basic Technology grant Managing Uncertainty in Complex Models is developing this approach  http://mucm.group.shef.ac.uk http://mucm.group.shef.ac.uk  See in particular the MUCM toolkit 6/9/20098mucm.group.shef.ac.uk

9 Validation  What does it mean to validate a simulation model?  Compare model predictions with reality  But the model is always wrong  How can something which is always wrong ever be called valid?  Conventionally, a model is said to be valid if its predictions are close enough to reality  How close is close enough?  Depends on purpose  Conventional approaches to validation confuse the absolute (valid) with the relative (fit for this purpose)  Let’s look at an analogous validation problem 6/9/20099mucm.group.shef.ac.uk

10 Validating an emulator 6/9/2009mucm.group.shef.ac.uk10  What does it mean to validate an emulator?  Compare the emulator’s predictions with the reality of model output  Make a validation sample of runs at new input configurations  The emulator mean is the best prediction and is always wrong  But the emulator predicts uncertainty around that mean  The emulator is valid if its expressions of uncertainty are correct  Actual outputs should fall in 95% intervals 95% of the time  No less and no more than 95% of the time  Standardised residuals should have zero mean and unit variance  See Bastos and O’Hagan preprint on MUCM website

11 Validation diagnostics 6/9/2009mucm.group.shef.ac.uk11

12 Validating the model 6/9/2009mucm.group.shef.ac.uk12  Let’s accept that there is uncertainty around model predictions  We need to be able to make statistical predictions  Then if we compare with observations we can see whether reality falls within the prediction bounds correctly  The difference between model output and reality is called model discrepancy  It’s also a function of the inputs  Like the model output, it’s typically a smooth function  Like the model output, we can emulate this function  We can validate this

13 Model discrepancy 6/9/2009mucm.group.shef.ac.uk13  Model discrepancy was first introduced within the MUCM framework in the context of model calibration  Ignoring discrepancy leads to over-fitting and over-confidence in the calibrated parameters  Understanding that it is a smooth error term rather than just noise is also crucial  To learn about discrepancy we need a training sample of observations of the real process  Then we can validate our emulation of reality using further observations  This is one ongoing strand of the MUCM project

14 Beyond validation 6/9/2009mucm.group.shef.ac.uk14  An emulator (of a model or of reality) can be valid and yet useless in practice  Given a sample of real-process observations, we can predict the output at any input to be the sample mean plus or minus two sample standard deviations  This will validate OK  Assuming the sample is representative  But it ignores the model and makes poor use of the sample!  Two valid emulators can be compared on the basis of the variance of their predictions  And declared fit for purpose if the variance is small enough

15 In conclusion 6/9/2009mucm.group.shef.ac.uk15  I think it is useful to separate the absolute property of validity from the relative property of fitness for purpose  Model predictions alone are useless without some idea of how accurate they are  Quantifying uncertainty in the predictions by building an emulator allows us to talk about validity  Only valid statistical predictions of reality should be accepted  Model predictions with a false measure of their accuracy are also useless!  We can choose between valid predictions on the basis of how accurate they are  And ask if they are sufficiently accurate for purpose

16 Advertisement 6/9/2009mucm.group.shef.ac.uk16  Workshop on emulators and MUCM methods  “Uncertainty in Simulation Models”  Friday 10th July 2009  10.30am - 4pm  National Oceanography Centre Southampton  http://mucm.group.shef.ac.uk/Pages/Project_News.htm  Please register with Katherine Jeays-Ward  (k.jeays-ward@sheffield.ac.uk) by 3rd July 2009  Registration is free, and lunch/refreshments will be provided


Download ppt "Validating uncertain predictions Tony O’Hagan, Leo Bastos, Jeremy Oakley, University of Sheffield."

Similar presentations


Ads by Google