Evolutionary Computational Intelligence Lecture 10a: Surrogate Assisted Ferrante Neri University of Jyväskylä.

Slides:



Advertisements
Similar presentations
Lindsey Bleimes Charlie Garrod Adam Meyerson
Advertisements

CSE 330: Numerical Methods
Partially Observable Markov Decision Process (POMDP)
Kriging.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to form.
1 Evolutionary Computational Inteliigence Lecture 6b: Towards Parameter Control Ferrante Neri University of Jyväskylä.
Variance reduction techniques. 2 Introduction Simulation models should be coded such that they are efficient. Efficiency in terms of programming ensures.
Experimental Design, Response Surface Analysis, and Optimization
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to.
ICS-271:Notes 6: 1 Notes 6: Game-Playing ICS 271 Fall 2008.
1. Algorithms for Inverse Reinforcement Learning 2
Optimal Design Laboratory | University of Michigan, Ann Arbor 2011 Design Preference Elicitation Using Efficient Global Optimization Yi Ren Panos Y. Papalambros.
EARS1160 – Numerical Methods notes by G. Houseman
The loss function, the normal equation,
1 Structure of search space, complexity of stochastic combinatorial optimization algorithms and application to biological motifs discovery Robin Gras INRIA.
A Heuristic Bidding Strategy for Multiple Heterogeneous Auctions Patricia Anthony & Nicholas R. Jennings Dept. of Electronics and Computer Science University.
Complexity Analysis (Part I)
Evolutionary Computational Intelligence
Sample size computations Petter Mostad
MAE 552 – Heuristic Optimization Lecture 6 February 6, 2002.
Evaluating Hypotheses
ICS-271:Notes 6: 1 Notes 6: Game-Playing ICS 271 Fall 2006.
Evolutionary Computational Intelligence Lecture 9: Noisy Fitness Ferrante Neri University of Jyväskylä.
Evolutionary Computational Intelligence Lecture 8: Memetic Algorithms Ferrante Neri University of Jyväskylä.
(hyperlink-induced topic search)
Lecture 10 Comparison and Evaluation of Alternative System Designs.
D Nagesh Kumar, IIScOptimization Methods: M1L4 1 Introduction and Basic Concepts Classical and Advanced Techniques for Optimization.
MNG221- Management Science –
Machine Learning Theory Maria-Florina Balcan Lecture 1, Jan. 12 th 2010.
Algorithm design techniques
3.7. O THER G AME P HYSICS A PPROACHES Overview of other game engine physics approaches.
Chapter 8 Introduction to Hypothesis Testing. Hypothesis Testing Hypothesis testing is a statistical procedure Allows researchers to use sample data to.
Machine Learning Theory Maria-Florina (Nina) Balcan Lecture 1, August 23 rd 2011.
Data Mining Techniques
Statistical Methods For Engineers ChE 477 (UO Lab) Larry Baxter & Stan Harding Brigham Young University.
Genetic Algorithm.
An efficient distributed protocol for collective decision- making in combinatorial domains CMSS Feb , 2012 Minyi Li Intelligent Agent Technology.
Introduction to variable selection I Qi Yu. 2 Problems due to poor variable selection: Input dimension is too large; the curse of dimensionality problem.
FDA- A scalable evolutionary algorithm for the optimization of ADFs By Hossein Momeni.
Stochastic Algorithms Some of the fastest known algorithms for certain tasks rely on chance Stochastic/Randomized Algorithms Two common variations – Monte.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Hierarchical Distributed Genetic Algorithm for Image Segmentation Hanchuan Peng, Fuhui Long*, Zheru Chi, and Wanshi Siu {fhlong, phc,
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
CSCI-256 Data Structures & Algorithm Analysis Lecture Note: Some slides by Kevin Wayne. Copyright © 2005 Pearson-Addison Wesley. All rights reserved. 4.
Experimental Evaluation of Learning Algorithms Part 1.
Ch9. Inferences Concerning Proportions. Outline Estimation of Proportions Hypothesis concerning one Proportion Hypothesis concerning several proportions.
Model Construction: interpolation techniques 1392.
Statistical Methods II&III: Confidence Intervals ChE 477 (UO Lab) Lecture 5 Larry Baxter, William Hecker, & Ron Terry Brigham Young University.
Solving the Maximum Cardinality Bin Packing Problem with a Weight Annealing-Based Algorithm Kok-Hua Loh University of Maryland Bruce Golden University.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
Tetris Agent Optimization Using Harmony Search Algorithm
Patch Based Prediction Techniques University of Houston By: Paul AMALAMAN From: UH-DMML Lab Director: Dr. Eick.
Chapter 12 FUSION OF FUZZY SYSTEM AND GENETIC ALGORITHMS Chi-Yuan Yeh.
1 Mean Analysis. 2 Introduction l If we use sample mean (the mean of the sample) to approximate the population mean (the mean of the population), errors.
NTU & MSRA Ming-Feng Tsai
Optimization in Engineering Design 1 Introduction to Non-Linear Optimization.
Evolving RBF Networks via GP for Estimating Fitness Values using Surrogate Models Ahmed Kattan Edgar Galvan.
Selection and Recombination Temi avanzati di Intelligenza Artificiale - Lecture 4 Prof. Vincenzo Cutello Department of Mathematics and Computer Science.
Chapter 9 Introduction to the t Statistic
CWR 6536 Stochastic Subsurface Hydrology Optimal Estimation of Hydrologic Parameters.
Sampling and Sampling Distribution
Heuristic Optimization Methods
Ana Wu Daniel A. Sabol A Novel Approach for Library Materials Acquisition using Discrete Particle Swarm Optimization.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
“Hard” Optimization Problems
Curve fit metrics When we fit a curve to data we ask:
Curve fit metrics When we fit a curve to data we ask:
Optimization of Real-Time Systems with Deadline Miss Ratio Constraints
Reinforcement Learning Dealing with Partial Observability
Evolutionary Computational Intelligence
Presentation transcript:

Evolutionary Computational Intelligence Lecture 10a: Surrogate Assisted Ferrante Neri University of Jyväskylä

Computationally expensive problems Optimization problems can be computationally expensive because of two reasons: high cardinality decision space (usually combinatorial) computationally expensive fitness function (e.g. design of on-line electric drives)

High Cardinality Decision Space Under such conditions it should be tried, on the basis of the application, to reduce the cardinality by means of an ”a priori” analysis or an heuristic to detect a promising region of the decision space Memetic approach (e.g. intelligent initial sampling) can be beneficial

Computationally expensive fitness It might happen that the fitness function evaluation requires by itself a lot of computational effort (e.g. in online PMSM drive design each fitness evaluation requires 8 s) In such conditions it should be found a way to reduce the numer of fitness evaluations and still reach the optimum

Surrogate Assisted Algorithms Surrogate Assisted Algorithms employ approximated models of the fitness function (cheap) alternatively with the real fitness (expensive) One of the crucial problems is the model to be employed and how to arrange such a combination

Global vs Local Surrogate models There are two complementary and contrasting algorithmic philosophy: – Global Surrogate Models: attempt of finding an approximated model of the landscape over all the decision space – Local Surrogate Models: attempt of approximating locally the landscape over the neighborhood of a certain point

Comparison between the two philosophies Global models assume that a wide knowledge of the decision space allows to build up an accurate model that can be employed as a cheap alternative of the real fitness Local models assume that a huge amount of information does not help in determining an accurate model and thus it is preferable to build up models that approximate only locally the behavior of the landscape Global models employ one very complex model, Local models employ many simple approximated functions

Coordination of models/real fitness The right way to perform the coordination is very problem dependent, both deterministic and stocastic rules have been implemented Models can be ”installed” in both evolutionary framework and local searchers

Surrogate Assisted Hooke-Jeeves Algorithm Surrogate Assisted Hooke Jeeves Algorithm (SAHJA): deterministic scheme for coordinating real fitness and a linear model obtained by least square method Computes N+1 points and generates a local linear model for calculating the remaining N points (Cost of exploratory move is thus kept constant) Check every directional move, by calculating the real fitness if a surrogate was prevously calculated (does not allow search directions by means of surrogate points)

SAHJA

SAHJA Results Very promising algorithmic performance Noise filtering

Evolutionary Computational Intelligence Lecture 10b: Experimentalism Ferrante Neri University of Jyväskylä

Goals To propose a research protocol in order to execute a fair experimental comparison which allows us to check whether the newly proposed algorithm outperforms the methods existing in literature In other words, if I designed a novel algorithm how can I be sure that my work outperform (for a certain problem) the state of the art?

Towards Performance Comparison If I designed a novel algorithm B how can I prove that B outperforms the benchmark algorithm A? How can I thus have a confirmation that the novel algorithmic component is really effective? Performance is an abstract concept not related to a specific machine. It is the capability of an algorithm of reaching a good performing solution in a certain time interval. The time trigger is the number of fitness (functional) evaluations

Experimental Setup For both A and B, a certain number n of runs must be performed The average best fitness values (e.g. at the end of each generation) must be saved N.B. for making the trends comparable an interpolation can be necessary Standard deviation bars can also be included

Two Possible kinds of outperforming Case 1: A and B converge on different final values Case 2: A and B converge on the same final values but with different convergence velocities

…Case 1 The data define two Tolerance Intervals (TIs) It is fixed a desired confidence level δ The proportion γ of a set of data which falls within a given interval with a given confidence level δ are determined by: γ =1 −a/n where n is the number of available samples and a is the positive root of the equation (1 + a) − (1 − δ) · e a = 0

…Case 2 A threshold value f thr is fixed. If during an experiment f best < f thr then the algorithm ”almost converged” For the n experiments, the number of fitness evaluations necessary to verify the inequality f best < f thr defines a TI The probability γ that an algorithm requires no more fitness evaluations than the most unlucky case is given by: γ =1 −d/n where n is the number of the available experiments. d is given by d = −ln(1 − δ)

How to conclude In both the cases, if the tolerance intervals are not separated it is impossible to establish that B outperforms A in all the cases. In this case it is possible only to state that B outperforms A in average