1 Physical Experiments & Computer Experiments, Preliminaries Chapters 1&2 “Design and Analysis of Experiments“ by Thomas J. Santner, Brian J. Williams.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Pattern Recognition and Machine Learning
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Pattern Recognition and Machine Learning
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Sensitivity Analysis In deterministic analysis, single fixed values (typically, mean values) of representative samples or strength parameters or slope.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Chapter 4: Linear Models for Classification
All Hands Meeting, 2006 Title: Grid Workflow Scheduling in WOSE (Workflow Optimisation Services for e- Science Applications) Authors: Yash Patel, Andrew.
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
1 A Prediction Interval for the Misclassification Rate E.B. Laber & S.A. Murphy.
A gentle introduction to Gaussian distribution. Review Random variable Coin flip experiment X = 0X = 1 X: Random variable.
. PGM: Tirgul 10 Parameter Learning and Priors. 2 Why learning? Knowledge acquisition bottleneck u Knowledge acquisition is an expensive process u Often.
Lecture 4 Measurement Accuracy and Statistical Variation.
Thanks to Nir Friedman, HU
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
CSC2535: 2013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton.
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Review of Lecture Two Linear Regression Normal Equation
Gaussian process modelling
1 Chapter 1: Introduction to Design of Experiments 1.1 Review of Basic Statistical Concepts (Optional) 1.2 Introduction to Experimental Design 1.3 Completely.
PATTERN RECOGNITION AND MACHINE LEARNING
沈致远. Test error(generalization error): the expected prediction error over an independent test sample Training error: the average loss over the training.
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
Random Sampling, Point Estimation and Maximum Likelihood.
LECTURER PROF.Dr. DEMIR BAYKA AUTOMOTIVE ENGINEERING LABORATORY I.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Learning Lateral Connections between Hidden Units Geoffrey Hinton University of Toronto in collaboration with Kejie Bao University of Toronto.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Conjugate Priors Multinomial Gaussian MAP Variance Estimation Example.
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
17 May 2007RSS Kent Local Group1 Quantifying uncertainty in the UK carbon flux Tony O’Hagan CTCD, Sheffield.
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
MSE-415: B. Hawrylo Chapter 13 – Robust Design What is robust design/process/product?: A robust product (process) is one that performs as intended even.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
5-1 ANSYS, Inc. Proprietary © 2009 ANSYS, Inc. All rights reserved. May 28, 2009 Inventory # Chapter 5 Six Sigma.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
Simulated Annealing G.Anuradha.
Bias and Variance of the Estimator PRML 3.2 Ethem Chp. 4.
BCS547 Neural Decoding.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
Additional Topics in Prediction Methodology. Introduction Predictive distribution for random variable Y 0 is meant to capture all the information about.
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Bias and Variance of the Estimator PRML 3.2 Ethem Chp. 4.
Chapter 11: Linear Regression and Correlation Regression analysis is a statistical tool that utilizes the relation between two or more quantitative variables.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
Javad Azimi, Ali Jalali, Xiaoli Fern Oregon State University University of Texas at Austin In NIPS 2011, Workshop in Bayesian optimization, experimental.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Maximum likelihood estimators Example: Random data X i drawn from a Poisson distribution with unknown  We want to determine  For any assumed value of.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
6/11/20161 Process Optimisation For Micro Laser Welding in Fibre Optics Asif Malik Supervisors: Prof. Chris Bailey & Dr. Stoyan Stoyanov 14 May 2008.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Bayesian Estimation and Confidence Intervals Lecture XXII.
Ch3: Model Building through Regression
G Lecture 10b Example: Recognition Memory
Topics Randomized complete block design (RCBD) Latin square designs
Special Topics In Scientific Computing
Data Mining Lecture 11.
Overview of Supervised Learning
Bias and Variance of the Estimator
Propagating Uncertainty In POMDP Value Iteration with Gaussian Process
Parametric Methods Berlin Chen, 2005 References:
Probabilistic Surrogate Models
Optimization under Uncertainty
Presentation transcript:

1 Physical Experiments & Computer Experiments, Preliminaries Chapters 1&2 “Design and Analysis of Experiments“ by Thomas J. Santner, Brian J. Williams and William I. Notz EARG presentation Oct 3, 2005 by Frank Hutter

2 Preface What they mean by Computer Experiments Code that serves as a proxy for physical process Code that serves as a proxy for physical process Can modify inputs and observe how process output is affedcted Can modify inputs and observe how process output is affedctedMath Should be understandable with a Masters level training of Statistics Should be understandable with a Masters level training of Statistics

3 Overview of the book Chapter 1: intro, application domains Chapter 2: Research goals for various types of inputs (random, controlled, model parameters) Research goals for various types of inputs (random, controlled, model parameters) Lots of definitions, Gaussian processes Lots of definitions, Gaussian processes Chapters 3-4: Predicting Output from Computer Experiments Chapter 5: Basic Experimental Design (similar to last term) Chapter 6: Active Learning (sequential experimental design) Chapter 7: Sensitivity analysis Appendix C: code

4 Physical experiments: a few concepts we saw last term (2f) Randomization In order to prevent unrecognized nuisance variables from systematically affecting response In order to prevent unrecognized nuisance variables from systematically affecting responseBlocking Deals with recognized nuisance variables Deals with recognized nuisance variables Group experimental units into homogeneous groups Group experimental units into homogeneous groupsReplication Reduce unavoidable measurement variation Reduce unavoidable measurement variation Computer experiments Deterministic outputs Deterministic outputs “None of the traditional principles [...] are of use“ “None of the traditional principles [...] are of use“

5 Types of input variables (15f) Control variables x c Can be set by experimenter / engineer Can be set by experimenter / engineer Engineering variables, manufactoring variables Engineering variables, manufactoring variables Environmental variables X e Depend on the environment/user Depend on the environment/user Random variables with known or unknown distribution Random variables with known or unknown distribution When known for a particular problem: x e When known for a particular problem: x e Noise variables Noise variables Model variables x m Parameters of the computer model that need to be set to get the best approximation of the physical process Parameters of the computer model that need to be set to get the best approximation of the physical process Model parameters, tuning parameters Model parameters, tuning parameters

6 Examples of Computer Models (6ff) ASET (Available Safe Egress Time) 5 inputs, 2 outputs 5 inputs, 2 outputs Design of Prosthesis Devices 3 environment variables, 2 control variables 3 environment variables, 2 control variables 2 competing outputs 2 competing outputs Formation of Pockets in Sheet Metal 6 control variables, 1 output 6 control variables, 1 output Other examples Optimally shaping helicopter blade – 31 control variables Optimally shaping helicopter blade – 31 control variables Public policy making: greenhouse gases – 30 input variables, some of them modifiable (control variables) Public policy making: greenhouse gases – 30 input variables, some of them modifiable (control variables)

7 ASET (Available Safe Egress Time) (4f) Evolution of fires in enclosed areas Inputs Room ceiling height and room area Room ceiling height and room area Height of burning object Height of burning object Heat loss fraction for the room (depends on insulation) Heat loss fraction for the room (depends on insulation) Material-specific heat release rate Material-specific heat release rate Maximum time for simulation (!) Maximum time for simulation (!)Outputs Temperature of the hot smoke layer Temperature of the hot smoke layer Distance of hot smoke layer from fire source Distance of hot smoke layer from fire source

8 Design of Prosthesis Devices (6f) 2 control variables b, the length of the bullet tip d, the midstem parameter 3 environment variables , the joint angle E, the elastic modulus of the surrounding cancellous bone Implant-bone interface friction 2 conflicting outputs Femoral stress shielding Implant toggling (flexible prostheses minimize stress, but toggle more  loosen)

9 Formation of Pockets in Steel (8ff) 6 control variables Length l Width w Fillet radius f Clearance c Punch plan view radius p Lock bead distance dOutput Failure depth (depth at which the metal tears)

10 Research goals for homogeneous-input codes (17f) Homogeneous-input: only one of the three possible variable types present All control variables: x=x c Predict y(x) well for all x in some domain X Predict y(x) well for all x in some domain X Global perspective Global perspective Integrated squared error s X [y‘(x) - y(x)] 2 w(x) dx Can‘t be computed since y(x) unknown, but in Chapter 6 we‘ll replace [y‘(x)-y(x)] 2 by a computable posterior mean squared value Local perspective Local perspective Level set: Find x such that y(x) = t 0 t 0 = maximum value

11 All environmental variables: x=X e (18) How does the variability in X e transmit through the computer code ? Find the distribution of y(X e ) When the problem is to find the mean: Latin hypercube designs for choosing the training sites Latin hypercube designs for choosing the training sites

12 All model variables x=x m (18f) Mathematical modelling contains unknown parameters (unknown rates or physical constants) Calibration (parameter fitting) Choose the model variables x m so that the computer output best matches the output from the physical experiment Choose the model variables x m so that the computer output best matches the output from the physical experiment

13 Research goals for mixed inputs (19ff) Focus on case with control and environmental variables x=(x c,X e ) where X e has a known distribution Example: hip prosthesis y(x c,X e ) is a random variable whose distribution is induced by X e Mean  x c ) = E{y(x c,X e )} Mean  x c ) = E{y(x c,X e )} Upper alpha quantile   =   (x c ): Upper alpha quantile   =   (x c ): P{y(x c,X e ) >=   } = 

14 Research goals for mixed inputs: simple adaption of previous goals (20ff) Predict y well over its domain: Minimize s X [  ‘(x) -  (x)] 2 w(x) dx Minimize s X [  ‘(x) -  (x)] 2 w(x) dx (again, there is a Bayesian analog with computable mean) (again, there is a Bayesian analog with computable mean) Maximize the mean output: max x c  (x c )

15 When the distribution of X e is unknown Various flavours of robustness G-robust: minimax Want to minimize your maximal loss Want to minimize your maximal loss Pessimistic Pessimistic  (.)-robust  (.)-robust Minimize weighted loss (weighted by prior density on distribution over X e ) Minimize weighted loss (weighted by prior density on distribution over X e )M-robust Suppose for a given x c, y(x c,x e ) is fairly flat Suppose for a given x c, y(x c,x e ) is fairly flat Then value of X e doesn‘t matter so much for that x c Then value of X e doesn‘t matter so much for that x c Maximize  (x c ) subject to constraints on variance w.r.t. X e Maximize  (x c ) subject to constraints on variance w.r.t. X e