Connie M. Borror, Arizona State University West Christine M. Anderson-Cook, Los Alamos National Laboratory Bradley Jones, JMP SAS Institute Construction.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Outline Background Model Development Variance Dispersion Graphs
Discussion ofAspects of Bias, Prediction Variance and Mean Square Error Geoff Vining Virginia Tech.
Introduction to parameter optimization
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Sampling plans for linear regression
Introduction to Design of Experiments by Dr Brad Morantz
3.3 Hypothesis Testing in Multiple Linear Regression
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Notes Sample vs distribution “m” vs “µ” and “s” vs “σ” Bias/Variance Bias: Measures how much the learnt model is wrong disregarding noise Variance: Measures.
Incomplete Block Designs. Randomized Block Design We want to compare t treatments Group the N = bt experimental units into b homogeneous blocks of size.
Statistical Analysis Professor Lynne Stokes Department of Statistical Science Lecture 15 Analysis of Data from Fractional Factorials and Other Unbalanced.
CHAPTER 13 M ODELING C ONSIDERATIONS AND S TATISTICAL I NFORMATION “All models are wrong; some are useful.”  George E. P. Box Organization of chapter.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Data Modeling and Parameter Estimation Nov 9, 2005 PSCI 702.
Experimental Design, Response Surface Analysis, and Optimization
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Sampling plans Given a domain, we can reduce the prediction error by good choice of the sampling points The choice of sampling locations is called “design.
Simple Linear Regression. Start by exploring the data Construct a scatterplot  Does a linear relationship between variables exist?  Is the relationship.
Linear regression models
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression.
11.1 Introduction to Response Surface Methodology
Section 4.2 Fitting Curves and Surfaces by Least Squares.
1 Chapter 3 Multiple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Minimaxity & Admissibility Presenting: Slava Chernoi Lehman and Casella, chapter 5 sections 1-2,7.
Industrial Applications of Response Surface Methodolgy John Borkowski Montana State University Pattaya Conference on Statistics Pattaya, Thailand.
Topic 2: Statistical Concepts and Market Returns
1 Chapter 9 Variable Selection and Model building Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
CSC2535: 2013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton.
Correlation & Regression
1 14 Design of Experiments with Several Factors 14-1 Introduction 14-2 Factorial Experiments 14-3 Two-Factor Factorial Experiments Statistical analysis.
Objectives of Multiple Regression
Simple Linear Regression Models
Partition Experimental Designs for Sequential Process Steps: Application to Product Development Leonard Perry, Ph.D., MBB, CSSBB, CQE Associate Professor.
Statistical Design of Experiments
Multiple Regression Selecting the Best Equation. Techniques for Selecting the "Best" Regression Equation The best Regression equation is not necessarily.
Engineering Statistics ENGR 592 Prepared by: Mariam El-Maghraby Date: 26/05/04 Design of Experiments Plackett-Burman Box-Behnken.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Jeroen Pannekoek - Statistics Netherlands Work Session on Statistical Data Editing Oslo, Norway, 24 September 2012 Topic (I) Selective and macro editing.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
V. Nonlinear Regression Objective-Function Surfaces Thus far, we have: Parameterized the forward model Obtained head and flow observations and their weights.
6-1 Introduction To Empirical Models Based on the scatter diagram, it is probably reasonable to assume that the mean of the random variable Y is.
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
International Conference on Design of Experiments and Its Applications July 9-13, 2006, Tianjin, P.R. China Sung Hyun Park, Hyuk Joo Kim and Jae-Il.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Model Building and Model Diagnostics Chapter 15.
The American University in Cairo Interdisciplinary Engineering Program ENGR 592: Probability & Statistics 2 k Factorial & Central Composite Designs Presented.
BUILDING THE REGRESSION MODEL Data preparation Variable reduction Model Selection Model validation Procedures for variable reduction 1 Building the Regression.
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
1 Prof. Indrajit Mukherjee, School of Management, IIT Bombay High (2 pounds ) Low (1pound) B=60 ( ) Ab=90 ( ) (1)=80 ( ) A=100.
Robust Estimation With Sampling and Approximate Pre-Aggregation Author: Christopher Jermaine Presented by: Bill Eberle.
Bias and Variance of the Estimator PRML 3.2 Ethem Chp. 4.
Statistical Data Analysis 2010/2011 M. de Gunst Lecture 9.
Linear Regression. Regression Consider the following 10 data pairs comparing the yield of an experiment to the temperature at which the experiment was.
CMS SAS Users Group Conference Learn more about THE POWER TO KNOW ® October 17, 2011 Medicare Payment Standardization Modeling using SAS Enterprise Miner.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
11 Designs for the First Order Model “Orthogonal first order designs” minimize the variance of If the off diagonal elements of X´X are all zero then the.
Lecture 6 Feb. 2, 2015 ANNOUNCEMENT: Lab session will go from 4:20-5:20 based on the poll. (The majority indicated that it would not be a problem to chance,
Analysis of Definitive Screening Designs
Probability Theory and Parameter Estimation I
ASEN 5070: Statistical Orbit Determination I Fall 2014
CH 5: Multivariate Methods
Bias and Variance of the Estimator
Design and Analysis of Nonlinear Models for the Mars 2020 Rover
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Special Topics In Design
ENM 310 Design of Experiments and Regression Analysis Chapter 3
14 Design of Experiments with Several Factors CHAPTER OUTLINE
Presentation transcript:

Connie M. Borror, Arizona State University West Christine M. Anderson-Cook, Los Alamos National Laboratory Bradley Jones, JMP SAS Institute Construction and Evaluation of Response Surface Designs Incorporating Bias from Model Misspecification

Motivation Response surface design evaluation (and creation) assuming a particular model Single number efficiencies Prediction variance performance Mean-squared error Model misspecification? What effect does this have on prediction and optimization?

Motivation Examine effect of model misspecification Expected squared bias Prediction variance Expected mean squared error Using fraction of design space (FDS) plots and box plots Evaluate designs based on the contribution of ESB relative to PV.

Scenario Cuboidal regions True form of the model is of higher order than the model being fit. Examine Response surface models when the true form is cubic Screening experiment when the true form is full second order.

Model Specifications The model to be fit is Y = X ε X 1 = n × p design matrix for the assumed form of the model The true form of the model is Y = X X ε X 2 = n × q design matrix pertaining to those parameters ( 2 ) not present in the model to be fit (assumed model).

Model Specifications 2 in general, are not fully estimable Assume 2 ~ N(0, )

Criteria Mean-squared error Expected squared bias (ESB): Expected MSE sum of PV and ESB

Fraction of Design Space (FDS) Plots Zahran, Anderson-Cook, and Myers (2003) scaled prediction variance values are plotted versus the fraction of the design space that has SPV at or below the given value Adapt this to plot ESB and EMSE as well as PV. We use FDS plots and box plots to assess the designs

Cases I. Two-factor response surface design Assume a second-order model: True form of the model is cubic:

Case I Designs Central Composite Design (CCD) Quadratic I-optimal (Q I-opt) Quadratic D-optimal (Q D-opt) Cubic I-optimal (C I-opt) Cubic D-optimal (C D-opt) Cubic Bayes I-optimal (C Bayes I-opt) Cubic Bayes D-optimal (C Bayes D-opt)

Case I CCD

Case I Designs CCD (ESB and EMSE performance as bias increases)

Case I Designs PV for all designs

Case I Designs ESB for all designs

Case I Designs EMSE for all designs

Case I Designs FDS for EMSE for all designs

Case II Four factor response surface design Assume a second-order model: True form of the model is cubic: 20 additional terms as we move from the second- order model to cubic.

Case II Designs Six possible designs, with n = 27 runs Central Composite Design (CCD) Box Behnken Design (BBD) Quadratic I-optimal (Q I-Opt) Quadratic D-optimal (Q D-Opt) Cubic Bayes I-optimal (C Bayes I-Opt) Cubic Bayes D-optimal (C Bayes D-Opt) Note: Cubic I- and D-Optimal not possible with available size of design

Case II PV for all designs

Case II EMSE for all designs

Case II FDS plot of EMSE for Four Factors

Case III Eight-factor Screening Design Assume a first-order model: True form of the model is full second-order:

Case III Designs fractional factorial design with 4 center runs D-optimal (for first order) Bayes I-optimal (for second order) Bayes D-optimal (for second order)

Case III Designs The difference in the number of terms from the assumed to the true form of the models increases from 8 to 44. We would expect bias to quickly dominate EMSE.

Case III PV for all designs

Case III ESB for all designs

Case III EMSE for all designs

Design Notes For the two-factor case: The I-optimal and CCD were equivalent. They performed the best based on minimizing the maximum EMSE They performed the best based on prediction variance

Design Notes For the four factor case, the BBD was best based on EMSE criteria (in particular, the 95 th percentile, median, mean) when size of the coefficients of missing terms are moderate to large The I-optimal design was competitive for this case only if small amounts of bias were present. As the number of missing cubic terms increases, the BBD was best for EMSE.

Design Notes I-optimal designs were highly competitive over 95% of the design region; not with respect to the maximum PV, ESB, and EMSE. Cubic Bayesian designs did not perform well.

Design Notes In the screening design example: The D-optimal designs best if the assumed model is correct, but break down quickly if quadratic terms are in the model Much more pronounced than in the response surface design cases. Quadratic Bayesian I-optimal design was best based on mean, median, and 95 th percentile of EMSE The fractional factorial design was best with respect to the maximum EMSE. The design was best for both PV and ESB when the PV and ESB contribution to the model were balanced.

Conclusions Appropriate design can strongly depend on the assumption that we know the true form of the underlying model If we select designs carefully it is often possible to select a model that predicts well in the design space, and provide some protection against missing model terms. The ESB approach to assessing the effect of missing terms provides is advantageous: do not have to specify coefficient values for the true underlying model, Instead, the relative size of the missing terms can be calibrated relative to the variance of the observations.

Conclusions Size of the bias variance relative to observational error needed to balance contributions from PV and ESB is highly dependent on the number of missing terms from the assumed model. As the number of missing terms increases, the ability of designs to cope with the bias decreases substantially different designs are able to handle this increasing bias differently.