Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Rachel T. Johnson Douglas C. Montgomery Bradley Jones
MCMC estimation in MlwiN
1 -Classification: Internal Uncertainty in petroleum reservoirs.
Assessing Uncertainties in Radiative Shock Modeling James Paul Holloway University of Michegan Joslin Goh, Mike Grosskopf, Bruce Fryxell, Derek Bingham.
Insert Date HereSlide 1 Using Derivative and Integral Information in the Statistical Analysis of Computer Models Gemma Stephenson March 2007.
Running a model's adjoint to obtain derivatives, while more efficient and accurate than other methods, such as the finite difference method, is a computationally.
Designing Ensembles for Climate Prediction
Pattern Recognition and Machine Learning
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Variance reduction techniques. 2 Introduction Simulation models should be coded such that they are efficient. Efficiency in terms of programming ensures.
1 Parametric Sensitivity Analysis For Cancer Survival Models Using Large- Sample Normal Approximations To The Bayesian Posterior Distribution Gordon B.
Markov-Chain Monte Carlo
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Cox Model With Intermitten and Error-Prone Covariate Observation Yury Gubman PhD thesis in Statistics Supervisors: Prof. David Zucker, Prof. Orly Manor.
Computing the Posterior Probability The posterior probability distribution contains the complete information concerning the parameters, but need often.
Quantifying Uncertainties in Radiative Shock Experiments Carolyn C. Kuranz CRASH Annual Review Fall 2010.
1 PERFORMANCE EVALUATION H Often one needs to design and conduct an experiment in order to: – demonstrate that a new technique or concept is feasible –demonstrate.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
CRASH UQ Program: Overview & Results James Paul Holloway CRASH Annual Review Fall 2010.
An Optimal Learning Approach to Finding an Outbreak of a Disease Warren Scott Warren Powell
Results 2 (cont’d) c) Long term observational data on the duration of effective response Observational data on n=50 has EVSI = £867 d) Collect data on.
Assessment of Predictive Capability James Paul Holloway CRASH Review Meeting October
Stat Stat Introduction to the Design of Experiments Instructor: Derek Bingham, Office: West Hall 451 Contact Information:
Statistics, data, and deterministic models NRCSE.
Preliminary Sensitivity Studies With CRASH 3D Bruce Fryxell CRASH Review October 20, 2009.
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Lecture 10 Comparison and Evaluation of Alternative System Designs.
The Calibration Process
Ken Powell and Ryan McClarren CRASH Review, October 2010 CRASH Students and Courses.
Lecture II-2: Probability Review
Hydrologic Modeling: Verification, Validation, Calibration, and Sensitivity Analysis Fritz R. Fiedler, P.E., Ph.D.
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
Bayesian parameter estimation in cosmology with Population Monte Carlo By Darell Moodley (UKZN) Supervisor: Prof. K Moodley (UKZN) SKA Postgraduate conference,
Specification of a CRM model Ken Cheung Department of Biostatistics, Columbia University (joint work with Shing Columbia)
Crystal Linkletter and Derek Bingham Department of Statistics and Actuarial Science Simon Fraser University Acknowledgements This research was initiated.
Brian Macpherson Ph.D, Professor of Statistics, University of Manitoba Tom Bingham Statistician, The Boeing Company.
17 May 2007RSS Kent Local Group1 Quantifying uncertainty in the UK carbon flux Tony O’Hagan CTCD, Sheffield.
Center for Radiative Shock Hydrodynamics Fall 2011 Review PDT and radiation transport Marvin L. Adams.
CHAPTER 17 O PTIMAL D ESIGN FOR E XPERIMENTAL I NPUTS Organization of chapter in ISSO –Background Motivation Finite sample and asymptotic (continuous)
Additional Topics in Prediction Methodology. Introduction Predictive distribution for random variable Y 0 is meant to capture all the information about.
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
Designing Factorial Experiments with Binary Response Tel-Aviv University Faculty of Exact Sciences Department of Statistics and Operations Research Hovav.
- 1 - Calibration with discrepancy Major references –Calibration lecture is not in the book. –Kennedy, Marc C., and Anthony O'Hagan. "Bayesian calibration.
Options and generalisations. Outline Dimensionality Many inputs and/or many outputs GP structure Mean and variance functions Prior information Multi-output,
Reducing the risk of volcanic ash to aviation Natalie Harvey, Helen Dacre (Reading) Helen Webster, David Thomson, Mike Cooke (Met Office) Nathan Huntley.
Javad Azimi, Ali Jalali, Xiaoli Fern Oregon State University University of Texas at Austin In NIPS 2011, Workshop in Bayesian optimization, experimental.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
1 Design of experiment for computer simulations Let X = (X 1,…,X p )  R p denote the vector of input values chosen for the computer program Each X j is.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
Statistics for Business and Economics 8 th Edition Chapter 7 Estimation: Single Population Copyright © 2013 Pearson Education, Inc. Publishing as Prentice.
Jump to first page Bayesian Approach FOR MIXED MODEL Bioep740 Final Paper Presentation By Qiang Ling.
Introduction to emulators Tony O’Hagan University of Sheffield.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
8 Sept 2006, DEMA2006Slide 1 An Introduction to Computer Experiments and their Design Problems Tony O’Hagan University of Sheffield.
A Primer on Running Deterministic Experiments
Multiple Imputation using SOLAS for Missing Data Analysis
The Calibration Process
Chapter 6 Calibration and Application Process
Variable Selection for Gaussian Process Models in Computer Experiments
More about Posterior Distributions
CS639: Data Management for Data Science
DESIGN OF EXPERIMENTS by R. C. Baker
Using Clustering to Make Prediction Intervals For Neural Networks
Applied Statistics and Probability for Engineers
Probabilistic Surrogate Models
Presentation transcript:

Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1

Experiment design Screening (identifying most important inputs) Emulator construction Prediction Calibration/tuning (solving inverse problems) Confidence/prediction interval estimation Analysis of multiple simulators 2 Will focus the framework where we can quantify uncertainties in predictions and the impact of the sources of variability CRASH has required innovations to most UQ activities

Page where: o model or system inputs o system response o simulator response o calibration parameters o observational error *Kennedy and O’Hagan (2001); Higdon et al. (2004) 3 3 The predictive modeling approach is often called model calibration *

Page where: o model or system inputs o system response o simulator response o calibration parameters o observational error 4 4 The predictive modeling approach is often called model calibration

Page where: o model or system inputs o system response o simulator response o calibration parameters o observational error 5 Gaussian Process Models (looking at other models) 5 The predictive modeling approach is often called model calibration

Page where: o model or system inputs o system response o simulator response o calibration parameters o observational error 6 Goal is to estimate unknown calibration parameters and also make predictions of the physical system 6 The predictive modeling approach is often called model calibration

Page Vector of observations and simulations denoted as 7 The Gaussian process model specifications links simulations and observations through the covariance

We have used 2-D CRASH simulations and observations to build and explore the predictive model for shock location and breakout time Experiment data: o 2008 and 2009 experiments o Experiment variables: Be thickness, Laser energy, Xe fill pressure, Observation time o Response: Shock location (2008) and shock breakout time (2009) 2-D CRASH Simulations o 104 simulations, varied over 5 inputs o Experiment variables: Be thickness, Laser energy, Observation time o Calibration parameters: Electron flux limiter, Be gamma, Wall opacity 8

Can sample from joint posterior distribution of the calibration parameters 9 Breakout time calibration Shock location calibration Joint calibration

10 A look at the posterior marginal distributions of the calibration parameters

Statistical model can be used to evaluate sensitivity of codes or system to inputs 11 2-D CRASH shock breakout time sensitivity plots

12 The statistical model is used to predict shock breakout time incorporating sources of uncertainty

13 (μs) The statistical model is used to predict shock location incorporating sources of uncertainty

Have simulations from 1-D and 2-D models 2-D models runs come at a higher computational cost Would like to use all simulations, and experiments, to make predictions 14 We developed a new statistical model for combining outputs from multi-fidelity simulators

Have simulations from 1-D and 2-D models 2-D models runs come at a higher computational cost Would like to use all simulations, and experiments, to make predictions 1-D CRASH Simulations o 1024 simulations o Experiment variables: Be thickness, Laser energy, Xe fill pressure, Observation time o Calibration parameters: Electron flux limiter, Laser energy scale factor 2-D CRASH Simulations o 104 simulations o Experiment variables: Be thickness, Laser energy, Xe fill pressure, Observation time o Calibration parameters: Electron flux limiter, Wall opacity, Be gamma 15 We developed a new statistical model for combining outputs from multi-fidelity simulators

The available shock information comes from models and experiments where: o model or system inputs o system response o simulator response o vectors of calibration parameters Modeling approach in the spirit of Kennedy and O’Hagan (2000); Kennedy and O’Hagan (2001); Higdon et al. (2004) 1-D simulator …calibration parameters are adjusted 2-D simulator …calibration parameters are adjusted Experiments … calibration parameters are fixed and unknown 16

Idea is that the 1-D code does not match the 2-D code for two reasons 17 Calibrate lower fidelity code to higher fidelity code

Link the simulator responses and observations through joint model and discrepancies 18

Link the simulator responses and observations through joint model and discrepancies 19

Link the simulator responses and observations through joint model and discrepancies 20

Link the simulator responses and observations through joint model and discrepancies 21 Comments: o For deciding what variables belong in the discrepancy, one can ask “what is fixed at this level” o The interpretation of the calibration parameters changes somewhat o Discrepancies are almost guaranteed for this specification

Link the simulator responses and observations through joint model and discrepancies 22 Gaussian Process Models

Need to specify prior distributions Approach is Bayesian Inverted-gamma priors for variance components Beta priors for the correlation parameters Log-normal priors for the calibration parameters 23

Can illustrate using a simple example 24 Low fidelity model

Can illustrate using a simple example 25 Low fidelity model High fidelity model

Can illustrate using a simple example 26 Low fidelity model High fidelity model True model + replication error

How would this work in practice? Evaluate each computer model at at different input settings We evaluated the low fidelity (LF) model 20 times with inputs (x, t 1, t f ) chosen according to a Latin hypercube design The high fidelity (HF) model was evaluated 5 times with inputs (x, t 2, t f ) chosen according to a Latin hypercube design The experimental data was generated by evaluating the true model 3 times and adding replication error from a N(0,0.2) 27

Observations and response functions at the true value of the calibration parameters 28

We can construct 95% posterior prediction intervals at the observations 29

Comparison of predicted response surfaces 30

New methodology applied to CRASH for breakout time 31

Observations Able to build a statistical model that appears to predict the observations well Prediction error is in the order of the experimental uncertainty Care must be taken choosing priors for the variances of GP’s 32

Approach to combine outputs from experiments and several different computer models Experiments: The mean function is just one of many possible response functions View computer model evaluations as biased versions of this “super-reality” 33 Developing new statistical model for combining simulations and experiments

Experiments: Computer model: Each computer model will be calibrated directly to the observations Information for estimating individual unknown calibration parameters comes from observations and models with that parameter as on input 34 Super-reality model for prediction and calibration

Use the model calibration framework to perform a variety of tasks such as explore the simulation response surfaces, making predictions for experiments and sensitivity analysis Developed new statistical model for calibration of multi- fidelity computer models with field data Can make predictions with associated uncertainty informed by multi-fidelity models Developing model to combine several codes (not necessarily ranked by fidelity) and observations 35 Have deployed state of the art UQ techniques to leverage CRASH codes and experiments

Allocation of computational budget The goal is to use available simulations and experiments to evaluate the allocation of the computational budget to computational models Since prediction is our goal, will use the reduction in the integrated mean square error (IMSE) This measures the prediction variance, averaged across the input space The optimal set of simulations is the one that maximized the expected reduction in the IMSE 36

Criterion can be evaluated in the current statistical framework Can compute an estimate of the mean square error at any potential input, conditional on the model parameters Would like a new trial to improve the prediction everyone in the input region This criterion is difficult to optimize 37

A quick illustration – CRASH 1-D using shock location Can use the 1-D predictive calibration model to evaluate the value of adding new trials Suppose wish to conduct 10 new field trials Which 10? What do we expect to gain? 38

Expected reduction in IMSE for up to 10 new experiments 39 Expected reduction in IMSE Number of follow-up experiments

Can compare the value of new experiments to simulations One new field trial yields an expected reduction in the IMSE of about 5% The optimal IMSE design with D new computer trials yields an expected reduction of of about 3% The value of an experiment is substantially more than that of a computer trial Can do the same exercise when there are multiple codes 40

Fin 41