1 Efficient experimentation for nanostructure synthesis using Sequential Minimum Energy Designs (SMED) V. Roshan Joseph +, Tirthankar Dasgupta* and C.

Slides:



Advertisements
Similar presentations
Sampling plans for linear regression
Advertisements

Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Partially Observable Markov Decision Process (POMDP)
Designing Ensembles for Climate Prediction
Monte Carlo Methods and Statistical Physics
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Experimental Design, Response Surface Analysis, and Optimization
Introduction to Function Minimization. Motivation example Data on height of a group of people, men and women Data on gender not recorded, not known.
CHAPTER 8 A NNEALING- T YPE A LGORITHMS Organization of chapter in ISSO –Introduction to simulated annealing –Simulated annealing algorithm Basic algorithm.
11.1 Introduction to Response Surface Methodology
Planning under Uncertainty
x – independent variable (input)
Optimization via Search CPSC 315 – Programming Studio Spring 2009 Project 2, Lecture 4 Adapted from slides of Yoonsuck Choe.
MAE 552 – Heuristic Optimization Lecture 6 February 6, 2002.
Detecting Image Region Duplication Using SIFT Features March 16, ICASSP 2010 Dallas, TX Xunyu Pan and Siwei Lyu Computer Science Department University.
An Optimal Learning Approach to Finding an Outbreak of a Disease Warren Scott Warren Powell
MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #16 3/1/02 Taguchi’s Orthogonal Arrays.
Optimization Methods One-Dimensional Unconstrained Optimization
Maximum Likelihood (ML), Expectation Maximization (EM)
Active Appearance Models Suppose we have a statistical appearance model –Trained from sets of examples How do we use it to interpret new images? Use an.
Lecture 17 Today: Start Chapter 9 Next day: More of Chapter 9.
Optimization Methods One-Dimensional Unconstrained Optimization
Introduction to Simulated Annealing 22c:145 Simulated Annealing  Motivated by the physical annealing process  Material is heated and slowly cooled.
Optimization via Search CPSC 315 – Programming Studio Spring 2008 Project 2, Lecture 4 Adapted from slides of Yoonsuck Choe.
10.1 Chapter 10 Optimization Designs Optimization Designs CS RO R Focus: A Few Continuous Factors Output: Best Settings Reference: Box, Hunter &
Radial Basis Function Networks
Elements of the Heuristic Approach
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Gwangju Institute of Science and Technology Intelligent Design and Graphics Laboratory Multi-scale tensor voting for feature extraction from unstructured.
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
Alignment and classification of time series gene expression in clinical studies Tien-ho Lin, Naftali Kaminski and Ziv Bar-Joseph.
Language Models Hongning Wang Two-stage smoothing [Zhai & Lafferty 02] c(w,d) |d| P(w|d) = +  p(w|C) ++ Stage-1 -Explain unseen words -Dirichlet.
Engineering Statistics ENGR 592 Prepared by: Mariam El-Maghraby Date: 26/05/04 Design of Experiments Plackett-Burman Box-Behnken.
Chapter 11Design & Analysis of Experiments 8E 2012 Montgomery 1.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Simulated Annealing.
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
An Efficient Sequential Design for Sensitivity Experiments Yubin Tian School of Science, Beijing Institute of Technology.
Xianwu Ling Russell Keanini Harish Cherukuri Department of Mechanical Engineering University of North Carolina at Charlotte Presented at the 2003 IPES.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Experimental Algorithmics Reading Group, UBC, CS Presented paper: Fine-tuning of Algorithms Using Fractional Experimental Designs and Local Search by Belarmino.
Vaida Bartkutė, Leonidas Sakalauskas
Tetris Agent Optimization Using Harmony Search Algorithm
Probabilistic Algorithms Evolutionary Algorithms Simulated Annealing.
CIAR Summer School Tutorial Lecture 1b Sigmoid Belief Nets Geoffrey Hinton.
Optimization Problems
Designing Factorial Experiments with Binary Response Tel-Aviv University Faculty of Exact Sciences Department of Statistics and Operations Research Hovav.
Machine Learning 5. Parametric Methods.
- 1 - Calibration with discrepancy Major references –Calibration lecture is not in the book. –Kennedy, Marc C., and Anthony O'Hagan. "Bayesian calibration.
Robust Synthesis of Nanostructures C.F.Jeff Wu* Georgia Institute of Technology (joint with Tirthankar Dasgupta*, Christopher Ma +, Roshan Joseph*, Z L.
Javad Azimi, Ali Jalali, Xiaoli Fern Oregon State University University of Texas at Austin In NIPS 2011, Workshop in Bayesian optimization, experimental.
Sporadic model building for efficiency enhancement of the hierarchical BOA Genetic Programming and Evolvable Machines (2008) 9: Martin Pelikan, Kumara.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Chance Constrained Robust Energy Efficiency in Cognitive Radio Networks with Channel Uncertainty Yongjun Xu and Xiaohui Zhao College of Communication Engineering,
Metaheuristics for the New Millennium Bruce L. Golden RH Smith School of Business University of Maryland by Presented at the University of Iowa, March.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
Introduction to Statistical Quality Control, 4th Edition Chapter 13 Process Optimization with Designed Experiments.
Written By: Presented By: Swarup Acharya,Amr Elkhatib Phillip B. Gibbons, Viswanath Poosala, Sridhar Ramaswamy Join Synopses for Approximate Query Answering.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Bayesian Optimization. Problem Formulation Goal  Discover the X that maximizes Y  Global optimization Active experimentation  We can choose which values.
Clustering (2) Center-based algorithms Fuzzy k-means Density-based algorithms ( DBSCAN as an example ) Evaluation of clustering results Figures and equations.
C. F. Jeff Wu+ (joint with Roshan Joseph+ & Tirthankar Dasgupta* )
Heuristic Optimization Methods
Clustering (3) Center-based algorithms Fuzzy k-means
Dynamical Statistical Shape Priors for Level Set Based Tracking
CSE 589 Applied Algorithms Spring 1999
The loss function, the normal equation,
Xin-She Yang, Nature-Inspired Optimization Algorithms, Elsevier, 2014
Mathematical Foundations of BME Reza Shadmehr
Stochastic Methods.
Presentation transcript:

1 Efficient experimentation for nanostructure synthesis using Sequential Minimum Energy Designs (SMED) V. Roshan Joseph +, Tirthankar Dasgupta* and C. F. Jeff Wu + + ISyE, Georgia Tech *Statistics, Harvard

2 Statistical modeling and analysis for robust synthesis of nanostructures Dasgupta, Ma, Joseph, Wang and Wu (2008), Journal of The American Statistical Association, to appear. Robust conditions for synthesis of Cadmium Selenide (CdSe) nanostructures derived New sequential algorithm for fitting multinomial logit models. Internal noise factors considered.

3 Fitted quadratic response surfaces & optimal conditions

4 The need for more efficient experimentation A 9x5 full factorial experiment was too expensive and time consuming. Quadratic response surface did not capture nanowire growth satisfactorily (Generalized R 2 was 50% for CdSe nanowire sub-model).

5 What makes exploration of optimum difficult? Complete disappearance of morphology in certain regions leading to large, disconnected, non-convex yield regions. Multiple optima. Expensive and time-consuming experimentation 36 hours for each run Gold catalyst required

6 “Actual” contour plot of CdSe nanowire yield Obtained by averaging yields over different substrates. Large no-yield (deep green region). Small no-yield region embedded within yield regions. Scattered regions of highest yield.

7 How many trials needed to hit the point of maximum yield ? Pressure Temperature

8 How many trials ? Let’s try one factor at-a-time ! Temperature Pressure  Could not find optimum  Almost 50% trials wasted (no yield)  Too few data for statistical modeling

9 A 5x9 full-factorial experiment Yield = f(temp, pressure) 17 out of 45 trials wasted (no morphology)! Pressure

10 Why are traditional methods inappropriate ? Need a sequential approach to keep run size to a minimum. Fractional factorials / orthogonal arrays Large number of runs as number of levels increase. Several no-morphology scenarios possible. Do not facilitate sequential experimentation. Response Surface Methods Complexity of response surface. Categorical (binary in the extreme case) possible.

11 The Objective To find a design strategy that Is model-independent, Can “carve out’’ regions of no-morphology quickly, Allows for exploration of complex response surfaces, Facilitates sequential experimentation.

12 What if design points are positively charged particles ? q1q1 q2q2 E = Kq 1 q 2 / d Charge inversely proportional to yield, e.g., q = 1-yield Pressure = 0.6 = 1.0 Y = 40% Y = 0

13 What position will a newly introduced particle occupy ? q1q1 q2q2 Pressure = 0.6 = 1.0 Total Potential Energy Minimized !!

14 The key idea Pick a point x. Conduct experiment at x and observe yield y ( x ). Assign charge q ( x ) inversely proportional to y ( x ) How quickly will you reach the optimum ? Once you reach there, how will you know that THIS IS IT ? Use y ( x ) to update your knowledge about yields at various points in the design space (How ?) Pick the next point as the one that minimizes the total potential energy in the design space.

15 The SMED algorithm

16 The next design point

17 Charge at unselected points

18 Choice of tuning constants PROPOSITION : There exists a value of  (inverse of the maximum yield p g ) for which the algorithm will stick to the global optimum, once it reaches there. In practice, p g will not be known. The constant  determines the rate of convergence. Both  and  will be estimated iteratively.

19 Performance with known 

20 Performance with known  (Contd.)

21 Performance with known  (Contd.) Initial point = (0.55,0.50) Initial point = (0.77,0.50)

22 Criteria for estimators of  and 

23 Iterative estimation of  and 

24 Improved SMED for random response Instead of an interpolating function, use a smoothing function to predict yields (and charges) at unobserved points. Update the charges of selected points as well, using the smoothing function. Local polynomial smoothing used. Two parameters: n T (threshold number of iterations after which smoothing is started). (smoothing constant; small  local fitting).

25 Improvement achieved for r = 5 Last row gives the performance of the standard algorithm. Modified algorithm significantly improves the number of times the global optimum is reached, does worse with respect to no-yield points (higher perturbation).

26 Summary A new sequential space-filling design SMED proposed. SMED is model independent, can quickly “carve out” no-morphology regions and allows for exploration of complex surfaces. Origination from laws of electrostatics. Algorithm for deterministic functions. Modified algorithm for random functions. Performance studied using nanowire data, modified Branin (2 dimensional) and Levy-Montalvo (4 dimensional) functions.

27 Predicting the future What the hell! I don’t want to use this stupid strategy for experimentation ! Use my SMED ! Image courtesy : Nano Stat

28

29 Advantages of space filling designs LHD (McCay et al. 1979), Uniform designs (Fang 2002) are primarily used for computer experiments. Can be used to explore complex surfaces with small number of runs. Model free. No problems with categorical/binary data. CAN THEY BE USED FOR SEQUENTIAL EXPERIMENTATION ? CARVE OUT REGIONS OF NO-MORPHOLOGY QUICKLY?

30 Sequential experimentation strategies for global optimization SDO, a grid-search algorithm by Cox and John (1997) Initial space-filling design. Prediction using Gaussian Process Modeling. Lower bounds on predicted values used for sequential selection of evaluation points. Jones, Schonlau and Welch (1998) Similar to SDO. Expected Improvement (EI) Criterion used. Balances the need to exploit the approximating surface with the need to improve the approximation.

31 Why they are not appropriate Most of them good for multiple optima, but do not shrink the experimental region fast. Algorithms that reduce the design space (Henkenjohann et al. 2005) assume connected and convex failure regions. Initial design may contain several points of no-morphology. Current scenario focuses less on convergence and more on quickly shrinking the design space.

32 Some performance measures for n 0 - run designs.

33 Performance with estimated  and  with 30-run designs

34 First 20 iterations (out of 30) with estimated  and 

35 Contour plots of estimated p(x) ( =y/r ) where y ~ binomial(r,p(x))

36 Performance of the algorithm with random response  Result of 100 simulations with  = 1.25, starting point = (0,0).  The last row represents the case of deterministic response and first three random response.  Concern: as r decreases, the number of cases in which the global optimum is identified reduces drastically.