Designing Factorial Experiments with Binary Response Tel-Aviv University Faculty of Exact Sciences Department of Statistics and Operations Research Hovav.

Slides:



Advertisements
Similar presentations
Statistics for Improving the Efficiency of Public Administration Daniel Peña Universidad Carlos III Madrid, Spain NTTS 2009 Brussels.
Advertisements

Random Forest Predrag Radenković 3237/10
Object Specific Compressed Sensing by minimizing a weighted L2-norm A. Mahalanobis.
The General Linear Model. The Simple Linear Model Linear Regression.
Computer vision: models, learning and inference
Visual Recognition Tutorial
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
Mutual Information Mathematical Biology Seminar
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
Statistical Methods Chichang Jou Tamkang University.
NORM BASED APPROACHES FOR AUTOMATIC TUNING OF MODEL BASED PREDICTIVE CONTROL Pastora Vega, Mario Francisco, Eladio Sanz University of Salamanca – Spain.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
An Optimal Learning Approach to Finding an Outbreak of a Disease Warren Scott Warren Powell
Collaborative Ordinal Regression Shipeng Yu Joint work with Kai Yu, Volker Tresp and Hans-Peter Kriegel University of Munich, Germany Siemens Corporate.
Planning operation start times for the manufacture of capital products with uncertain processing times and resource constraints D.P. Song, Dr. C.Hicks.
Visual Recognition Tutorial
Modeling Gene Interactions in Disease CS 686 Bioinformatics.
Generalized Linear Models
Review of Lecture Two Linear Regression Normal Equation
Simple Linear Regression
CHAPTER 17 O PTIMAL D ESIGN FOR E XPERIMENTAL I NPUTS Organization of chapter in ISSO* –Background Motivation Finite sample and asymptotic (continuous)
Surface Simplification Using Quadric Error Metrics Michael Garland Paul S. Heckbert.
Generalized Minimum Bias Models
Sequential Experimental Designs For Sensitivity Experiments NIST GLM Conference April 18-20, 2002 Joseph G. Voelkel Center for Quality and Applied Statistics.
CSDA Conference, Limassol, 2005 University of Medicine and Pharmacy “Gr. T. Popa” Iasi Department of Mathematics and Informatics Gabriel Dimitriu University.
Stochastic Linear Programming by Series of Monte-Carlo Estimators Leonidas SAKALAUSKAS Institute of Mathematics&Informatics Vilnius, Lithuania
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Benk Erika Kelemen Zsolt
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Generalized Linear Models All the regression models treated so far have common structure. This structure can be split up into two parts: The random part:
Unsupervised Learning: Clustering Some material adapted from slides by Andrew Moore, CMU. Visit for
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
Xianwu Ling Russell Keanini Harish Cherukuri Department of Mechanical Engineering University of North Carolina at Charlotte Presented at the 2003 IPES.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Vaida Bartkutė, Leonidas Sakalauskas
Multi-Speaker Modeling with Shared Prior Distributions and Model Structures for Bayesian Speech Synthesis Kei Hashimoto, Yoshihiko Nankaku, and Keiichi.
CHAPTER 17 O PTIMAL D ESIGN FOR E XPERIMENTAL I NPUTS Organization of chapter in ISSO –Background Motivation Finite sample and asymptotic (continuous)
Generalized Linear Models (GLMs) and Their Applications.
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Lecture 2: Statistical learning primer for biologists
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
Simulation Study for Longitudinal Data with Nonignorable Missing Data Rong Liu, PhD Candidate Dr. Ramakrishnan, Advisor Department of Biostatistics Virginia.
Lecture 18 Today: More Chapter 9 Next day: Finish Chapter 9.
1 Optimizing Decisions over the Long-term in the Presence of Uncertain Response Edward Kambour.
Bayesian Speech Synthesis Framework Integrating Training and Synthesis Processes Kei Hashimoto, Yoshihiko Nankaku, and Keiichi Tokuda Nagoya Institute.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
NTU & MSRA Ming-Feng Tsai
Maximin D-optimal designs for binary longitudinal responses Fetene B. Tekle, Frans E. S. Tan and Martijn P. F. Berger Department of Methodology and Statistics,
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
D/RS 1013 Discriminant Analysis. Discriminant Analysis Overview n multivariate extension of the one-way ANOVA n looks at differences between 2 or more.
1 Design of experiment for computer simulations Let X = (X 1,…,X p )  R p denote the vector of input values chosen for the computer program Each X j is.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation Yee W. Teh, David Newman and Max Welling Published on NIPS 2006 Discussion.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Introduction We consider the data of ~1800 phenotype measurements Each mouse has a given probability distribution of descending from one of 8 possible.
Data Modeling Patrice Koehl Department of Biological Sciences
Chapter 7. Classification and Prediction
Non-Parametric Models
Robust Multipath Channel Identification
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Generally Discriminant Analysis
Learning From Observed Data
Uncertainty Propagation
Stochastic Methods.
Optimization under Uncertainty
Presentation transcript:

Designing Factorial Experiments with Binary Response Tel-Aviv University Faculty of Exact Sciences Department of Statistics and Operations Research Hovav A. Dror & David M. Steinberg July /40 International Conference on DOE – Nankai University

2/40 Overview  Introduction – Designs for GLM’s  Local D-optimal Designs  Robust Designs  Sequential Designs  Conclusions Robust Experimental Design for multivariate GLM Technical reports and MATLAB macros available at

3/40 D-optimal GLM designs  Theory like that for linear model, but with a crucial, difference. Fisher’s information matrix changes: F T F  F T WF  D-optimality: maximize  (Local D-optimal, and Local D-Efficiency)

4/40 Introduction – Visualization

5/40 Introduction – Main Objectives  Construction of an algorithm to find Local D-optimal Designs  Generalization: From locally optimal designs into robust designs (which take account of the uncertainty in the model parameters)  Further robustness – for different link functions, linear predictors, etc.  Sequential design – use data to estimate the model and improve the design as the experiment runs.

6/40 Overview  Introduction  Local D-optimal Designs  Robust Designs  Sequential Designs  Conclusions Robust Experimental Design for multivariate GLM

7/40 Local D-optimal designs – Algorithm  Mimics algorithms for linear models.  Main element – a row exchange procedure.  Rows are added or deleted, weighting the regression functions in accord with the mean value.  Timing: 1 second for a 16 point Poisson regression with 5 variables + interactions (accuracy 2 decimal places)

8/40  Introduction  Local D-optimal Designs  Robust Designs –Clustering: Motivating Example –Clustering vs. Bayesian Designs –Clustering vs. Compromise Designs –Linear Predictor and Link function Robustness –Ink Production Example  Sequential Designs  Conclusions Overview Robust Experimental Design for multivariate GLM

9/40 Clustering – Motivating Example  Proximity of 25 local D-optimal designs for a logistic model with intercept value uncertainty

10/40 Overview  Introduction  Local D-optimal Designs  Robust Designs –Clustering: Motivating Example –Clustering vs. Bayesian Designs –Clustering vs. Compromise Designs –Linear Predictor and Link function Robustness –Ink Production Example  Conclusions Robust Experimental Design for multivariate GLM

11/40 CLUSTERING vs. BAYESIAN DESIGNS (1)  Chaloner & Larntz (1989) Design Criterion: maximize the mean (over a prior distribution) of the information matrix log determinant Their optimal Bayesian Design:  Uses 7 support points  Reported value of for the criterion

12/40  Both designs (almost) meet sufficient requirements for optimality proof CLUSTERING vs. BAYESIAN DESIGNS (2)  K-means Clustering over 100 Local Designs  Local Designs’ coefficients: Low-Discrepancy sequence (Niederreiter’s)Niederreiter Number of Support Points Average Log Determinant of the Information Matrix Chaloner and Larntz (1989) Reported Value  Evaluated over 10,000 Coefficients vectors

13/40 CLUSTERING vs. BAYESIAN DESIGNS (3)  Expect Bayesian to be generally better  But… If Clustering does not fall much:  Simplicity of creation  Considerably less computational needs  Extension to multivariate problems – almost trivial

14/40 Overview  Introduction  Local D-optimal Designs  Robust Designs –Clustering: Motivating Example –Clustering vs. Bayesian Designs –Clustering vs. Compromise Designs –Linear Predictor and Link function Robustness –Ink Production Example  Conclusions Robust Experimental Design for multivariate GLM

15/40 Clustering vs. Multivariate Compromise Designs (1)  Woods, Lewis, Eccleston and Russell (Technometrics, May 2006): –A method for finding exact designs for experiments in which there are several explanatory variables –Use Simulated Annealing to find a design with the same criterion as Chaloner & Larntz –They note that evaluating the integral is too computationally intensive for incorporation within a search algorithm, and therefore average over a partial set

16/40 Clustering vs. Multivariate Compromise Designs (2)  Crystallography experiment –4 variables (rate of agitation during mixing, volume of composition, temperature and evaporation rate) –Affect the probability that a new product is formed –First order logistic model (with no interactions) –16 (/48) observations –Parameter space: (demonstrating algorithm’s superiority) Performance evaluated using median and minimum Local D-Efficiencies relative to 10,000 random parameter vectors

17/40 Clustering vs. Multivariate Compromise Designs (3) Minimum Efficiency Median Efficiency Design Standard 2 4 factorial Woods’ Compromise design

18/40 Clustering vs. Multivariate Compromise Designs (4)  Clustering procedure (1): –First, created Local Designs for 100 parameter vectors (Neiderreiter sequence)Neiderreiter –1,600 points K-means clustering (K=16) 30 seconds 0.25 seconds Minutes Minimum Efficiency Median Efficiency Design Standard 2 4 factorial Woods’ Compromise Clustering (1) [0.06,0.12][0.38,0.42]

19/40 Clustering vs. Multivariate Compromise Designs (5)  Clustering procedure (2): –Choose the cluster with highest average log determinant of information matrix, over N clustering repetitions: Minutes Minimum Efficiency Median Efficiency Design Standard 2 4 factorial Woods’ Compromise Clustering (1) [0.06,0.13] 0.42 [0.416,0.430] Clustering (2)

20/40 Clustering vs. Multivariate Compromise Designs (6)  Fast procedure Examine effect of # of Support points Number of Support Points Approximate Efficiency Median Efficiency Minimum Efficiency 20 seconds

21/40 Clustering vs. Multivariate Compromise Designs (7) Minutes Minimum Efficiency Median Efficiency Design Standard 2 4 factorial Woods’ Compromise Clustering (1) Clustering (2) [0.141, ] [0.415, 0.432] Clustering (3) Crystallography experiment - summary

22/40 Clustering vs. Multivariate Compromise Designs (6)  Advantageous byproduct of clustering: Number of Support Points Approximate Efficiency Median Efficiency Minimum Efficiency 20 seconds

23/40 Overview  Introduction  Prior Work  Local D-optimal Designs  Robust Designs –Clustering: Motivating Example –Clustering vs. Bayesian Designs –Clustering vs. Compromise Designs –Linear Predictor and Link function Robustness –Ink Production Example  Conclusions Robust Experimental Design for multivariate GLM

24/40 Robustness for Linear Predictors and Link functions  (again from Woods et al.)  2 variables  2 linear predictors: with / without interactions  2 link functions: Probit / CLL  Given (known) coefficients values

25/40 Overview  Introduction  Local D-optimal Designs  Robust Designs –Clustering: Motivating Example –Clustering vs. Bayesian Designs –Clustering vs. Compromise Designs –Linear Predictor and Link function Robustness –Ink Production Example  Conclusions Robust Experimental Design for multivariate GLM

26/40 Ink Production Example (1)  A Poisson Model  5 Variables  Normally Distributed Coefficients values uncertainty  Uncertainty about interaction effects  Centroid design reasonably efficient

27/40 Ink Production Example (2)  5 Tubes, each with different chemical  Each tube: Chosen concentration (fixed volume)  Ink quality classification: # of imperfect marks (on a standard printed test page)  Low concentrations – low quality, unusable  High concentrations – expensive  Model building based on experts opinions

28/40 Ink Production Example (3)  Model building based on experts opinions

29/40 Ink Production Example (4)  Full Factorial D-Efficiency:

30/40 Ink Production Example (5)  Cluster Design D-Efficiency:

31/40 Ink Production Example (6)  Centroid Design D-Efficiency:

32/40 Ink Production Example (6) Centroid Design D-Efficiency Cluster Design D-Efficiency

33/40 Ink Production Example (7) Efficiency Equivalent Sample Size

34/40 Overview  Introduction  Local D-optimal Designs  Robust Designs  Sequential Designs  Conclusions Robust Experimental Design for multivariate GLM

35/40 Sequential Designs  Good design requires knowledge of coefficients.  Use the data thus far to assess the model and the coefficients.  Augment the design accordingly.  Bayesian framework is natural. Robust Experimental Design for multivariate GLM

36/40 Sequential Designs Current methods:  Bruceton (Dixon and Mood 1948)  Langlie (1965)  Neyer (1994)  Wang, Smith & Ye (2006) Robust Experimental Design for multivariate GLM

37/40 Sequential Designs Robust Experimental Design for multivariate GLM Our method can be applied with many factors and in both fully sequential and group-sequential settings. Current methods are limited to:  One-factor experiments.  Fully sequential experiments.

38/40 Efficiency Comparison Efficiency One-stage ROBUST SEQUENTIAL Median: % quantile: 0.30 Median: % quantile: points

39/40 Overview  Introduction  Local D-optimal Designs  Robust Designs  Sequential Designs  Conclusions Robust Experimental Design for multivariate GLM

40/40 Summary & Conclusions  Local D-optimal designs for GLM can be easily found  Clustering a database of local D-optimal designs creates a robust design  Clustering is Robust for many uncertainty types: –parameter space, linear predictors, link functions, …  Simple procedure, minimal computational resources  Speed allows exploration of various designs and investigation of different number of support points  Outperforms more sophisticated and complex design optimization methods  Efficient sequential designs by combining the ideas with a Bayesian updating approach.