Evolutionary Computational Intelligence

Slides:



Advertisements
Similar presentations
Genetic Algorithms Chapter 3. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Genetic Algorithms GA Quick Overview Developed: USA in.
Advertisements

Evolution strategies Chapter 4.
Genetic Algorithms Chapter 3.
Parameter control Chapter 8. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Parameter Control in EAs 2 Motivation 1 An EA has many.
1 Evolutionary Computational Inteliigence Lecture 6b: Towards Parameter Control Ferrante Neri University of Jyväskylä.
Parameter Control A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Chapter 8.
Evolution strategies Chapter 4. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolution Strategies ES quick overview Developed: Germany.
Tuesday, May 14 Genetic Algorithms Handouts: Lecture Notes Question: when should there be an additional review session?
1 Lecture 8: Genetic Algorithms Contents : Miming nature The steps of the algorithm –Coosing parents –Reproduction –Mutation Deeper in GA –Stochastic Universal.
EvoNet Flying Circus Introduction to Evolutionary Computation Brought to you by (insert your name) The EvoNet Training Committee The EvoNet Flying Circus.
Evolutionary Computational Intelligence
A new crossover technique in Genetic Programming Janet Clegg Intelligent Systems Group Electronics Department.
Doug Downey, adapted from Bryan Pardo, Machine Learning EECS 349 Machine Learning Genetic Programming.
Tutorial 1 Temi avanzati di Intelligenza Artificiale - Lecture 3 Prof. Vincenzo Cutello Department of Mathematics and Computer Science University of Catania.
Evolution strategies (ES) Chapter 4. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolution Strategies Evolution strategies Overview.
CS 447 Advanced Topics in Artificial Intelligence Fall 2002.
Lecture 5: EP and DE 1 Evolutionary Computational Intelligence Lecture 5a: Overview about Evolutionary Programming Ferrante Neri University of Jyväskylä.
Genetic Programming Chapter 6. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Genetic Programming GP quick overview Developed: USA.
Genetic algorithms: an introduction
Prepared by Barış GÖKÇE 1.  Search Methods  Evolutionary Algorithms (EA)  Characteristics of EAs  Genetic Programming (GP)  Evolutionary Programming.
Genetic Algorithm.
SOFT COMPUTING (Optimization Techniques using GA) Dr. N.Uma Maheswari Professor/CSE PSNA CET.
Evolution Strategies Evolutionary Programming Genetic Programming Michael J. Watts
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
What is Genetic Programming? Genetic programming is a model of programming which uses the ideas (and some of the terminology) of biological evolution to.
Genetic Algorithms Genetic algorithms imitate a natural optimization process: natural selection in evolution. Developed by John Holland at the University.
Introduction to Evolutionary Algorithms Session 4 Jim Smith University of the West of England, UK May/June 2012.
Neural and Evolutionary Computing - Lecture 6
Crossovers and Mutation Richard P. Simpson. Genotype vs. Phenotype The genotype of a chromosome is just the basic data structure (it bits in a binary.
Evolution strategies Luis Martí DEE/PUC-Rio. ES quick overview Developed: Germany in the 1970’s Early names: I. Rechenberg, H.-P. Schwefel Typically applied.
Why do GAs work? Symbol alphabet : {0, 1, * } * is a wild card symbol that matches both 0 and 1 A schema is a string with fixed and variable symbols 01*1*
Evolutionary Programming
/ 26 Evolutionary Computing Chapter 8. / 26 Chapter 8: Parameter Control Motivation Parameter setting –Tuning –Control Examples Where to apply parameter.
Evolutionary Computing Dialects Presented by A.E. Eiben Free University Amsterdam with thanks to the EvoNet Training Committee and its “Flying Circus”
Genetic Algorithms. 2 Overview Introduction To Genetic Algorithms (GAs) GA Operators and Parameters Genetic Algorithms To Solve The Traveling Salesman.
MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #12 2/20/02 Evolutionary Algorithms.
Evolution strategies Chapter 4. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolution Strategies ES quick overview Developed: Germany.
Evolutionary Computing Chapter 4. / 62 Chapter 4: Representation, Mutation, and Recombination Role of representation and variation operators Most common.
EE749 I ntroduction to Artificial I ntelligence Genetic Algorithms The Simple GA.
Genetic Programming A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Chapter 6.
Evolution strategies Chapter 4. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolution Strategies ES quick overview Developed: Germany.
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
Genetic Algorithm Dr. Md. Al-amin Bhuiyan Professor, Dept. of CSE Jahangirnagar University.
Selection and Recombination Temi avanzati di Intelligenza Artificiale - Lecture 4 Prof. Vincenzo Cutello Department of Mathematics and Computer Science.
Artificial Intelligence By Mr. Ejaz CIIT Sahiwal Evolutionary Computation.
Evolutionary Programming A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Chapter 5.
A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolutionary Programming.
Genetic Algorithms And other approaches for similar applications Optimization Techniques.
Genetic Algorithm. Outline Motivation Genetic algorithms An illustrative example Hypothesis space search.
Genetic Programming.
Genetic Algorithms Author: A.E. Eiben and J.E. Smith
Evolutionary Programming
Evolution Strategies Evolutionary Programming
Evolution strategies Chapter 4.
Introduction to Operators
Genetic Algorithms Chapter 3.
Parameter control Chapter 8.
Evolutionary Programming
Boltzmann Machine (BM) (§6.4)
Genetic Programming Chapter 6.
Genetic Programming Chapter 6.
A Tutorial (Complete) Yiming
Genetic Programming Chapter 6.
Evolution strategies Chapter 4.
Parameter control Chapter 8.
Evolution Strategies Originally developed in Germany in the early 60s with Rechenberg and Schwefel being the main contributors. Main ideas include: floating.
Evolution strategies Chapter 4.
Evolution Strategies Originally developed in Germany in the early 60s with Rechenberg and Schwefel being the main contributors. Main ideas include: floating.
Parameter control Chapter 8.
Presentation transcript:

Evolutionary Computational Intelligence Lecture 4: Real valued GAs and ES Ferrante Neri University of Jyväskylä Real valued GAs and ES

Real valued problems Many problems occur as real valued problems, e.g. continuous parameter optimisation f :  n   Illustration: Ackley’s function (often used in EC) Real valued GAs and ES

Mapping real values on bit strings z  [x,y]   represented by {a1,…,aL}  {0,1}L [x,y]  {0,1}L must be invertible (one phenotype per genotype) : {0,1}L  [x,y] defines the representation Only 2L values out of infinite are represented L determines possible maximum precision of solution High precision  long chromosomes (slow evolution) Real valued GAs and ES

Floating point mutations 1 General scheme of floating point mutations Uniform mutation: Analogous to bit-flipping (binary) or random resetting (integers) Real valued GAs and ES

Floating point mutations 2 Non-uniform mutations: Many methods proposed,such as time-varying range of change etc. Most schemes are probabilistic but usually only make a small change to value Most common method is to add random deviate to each variable separately, taken from N(0, ) Gaussian distribution and then curtail to range Standard deviation  controls amount of change (2/3 of deviations will lie in range (-  to + ) Real valued GAs and ES

Crossover operators for real valued GAs Discrete: each allele value in offspring z comes from one of its parents (x,y) with equal probability: zi = xi or yi Could use n-point or uniform Intermediate exploits idea of creating children “between” parents (hence a.k.a. arithmetic recombination) zi =  xi + (1 - ) yi where  : 0    1. The parameter  can be: constant: uniform arithmetical crossover variable (e.g. depend on the age of the population) picked at random every time Real valued GAs and ES

Single arithmetic crossover Parents: x1,…,xn  and y1,…,yn Pick a single gene (k) at random, child1 is: reverse for other child. e.g. with  = 0.5 Real valued GAs and ES

Simple arithmetic crossover Parents: x1,…,xn  and y1,…,yn Pick random gene (k) after this point mix values child1 is: reverse for other child. e.g. with  = 0.5 Real valued GAs and ES

Whole arithmetic crossover Parents: x1,…,xn  and y1,…,yn child1 is: reverse for other child. e.g. with  = 0.5 Real valued GAs and ES

Box crossover Parents: x1,…,xn  and y1,…,yn child1 is: Where α is a VECTOR of numers [0,1] Real valued GAs and ES

Comparison of the crossovers Arithmetic crossover works on a line which connects the two parents Box crossover works in a hyper-rectangular where the two parents are located in the vertexes Real valued GAs and ES

Evolution Strategies Ferrante Neri University of Jyväskylä Real valued GAs and ES

ES quick overview Developed: Germany in the 1970’s Early names: I. Rechenberg, H.-P. Schwefel Typically applied to: numerical optimisation Attributed features: fast good optimizer for real-valued optimisation relatively much theory Special: self-adaptation of (mutation) parameters standard Real valued GAs and ES

ES technical summary tableau Representation Real-valued vectors Recombination Discrete or intermediary Mutation Gaussian perturbation Parent selection Uniform random Survivor selection (,) or (+) Specialty Self-adaptation of mutation step sizes Real valued GAs and ES

Representation Chromosomes consist of three parts: Object variables: x1,…,xn Strategy parameters: Mutation step sizes: 1,…,n Rotation angles: 1,…, n Not every component is always present Full size:  x1,…,xn, 1,…,n ,1,…, k  where k = n(n-1)/2 (no. of i,j pairs) Real valued GAs and ES

Parent selection Parents are selected by uniform random distribution whenever an operator needs one/some Thus: ES parent selection is unbiased - every individual has the same probability to be selected Note that in ES “parent” means a population member (in GA’s: a population member selected to undergo variation) Real valued GAs and ES

Mutation Main mechanism: changing value by adding random noise drawn from normal distribution x’i = xi + N(0,) Key idea:  is part of the chromosome  x1,…,xn,    is also mutated into ’ (see later how) Thus: mutation step size  is coevolving with the solution x Real valued GAs and ES

Mutate  first Net mutation effect:  x,     x’, ’  Order is important: first   ’ (see later how) then x  x’ = x + N(0,’) Rationale: new  x’ ,’  is evaluated twice Primary: x’ is good if f(x’) is good Secondary: ’ is good if the x’ it created is good Reversing mutation order this would not work Real valued GAs and ES

Mutation case 0: 1/5 success rule z values drawn from normal distribution N(,) mean  is set to 0 variation  is called mutation step size  is varied on the fly by the “1/5 success rule”: This rule resets  after every k iterations by  =  / c if ps > 1/5  =  • c if ps < 1/5  =  if ps = 1/5 where ps is the % of successful mutations, 0.8  c < 1 Real valued GAs and ES

Mutation case 1: Uncorrelated mutation with one  Chromosomes:  x1,…,xn,   ’ =  • exp( • N(0,1)) x’i = xi + ’ • N(0,1) Typically the “learning rate”   1/ n½ And we have a boundary rule ’ < 0  ’ = 0 Real valued GAs and ES

Mutants with equal likelihood Circle: mutants having the same chance to be created Real valued GAs and ES

Mutation case 2: Uncorrelated mutation with n ’s Chromosomes:  x1,…,xn, 1,…, n  ’i = i • exp(’ • N(0,1) +  • Ni (0,1)) x’i = xi + ’i • Ni (0,1) Two learning rate parameters: ’ overall learning rate  coordinate wise learning rate   1/(2 n)½ and   1/(2 n½) ½ And i’ < 0  i’ = 0 Real valued GAs and ES

Mutants with equal likelihood Ellipse: mutants having the same chance to be created Real valued GAs and ES

Mutation case 3: Correlated mutations Chromosomes:  x1,…,xn, 1,…, n ,1,…, k where k = n • (n-1)/2 and the covariance matrix C is defined as: cii = i2 cij = 0 if i and j are not correlated cij = ½ • ( i2 - j2 ) • tan(2 ij) if i and j are correlated Note the numbering / indices of the ‘s Real valued GAs and ES

Correlated mutations The mutation mechanism is then: ’i = i • exp(’ • N(0,1) +  • Ni (0,1)) ’j = j +  • N (0,1) x ’ = x + N(0,C’) x stands for the vector  x1,…,xn  C’ is the covariance matrix C after mutation of the  values   1/(2 n)½ and   1/(2 n½) ½ and   5° i’ < 0  i’ = 0 and | ’j | >   ’j = ’j - 2  sign(’j) Real valued GAs and ES

Mutants with equal likelihood Ellipse: mutants having the same chance to be created Real valued GAs and ES

Recombination Creates one child Acts per variable / position by either Averaging parental values, or Selecting one of the parental values From two or more parents by either: Using two selected parents to make a child Selecting two parents for each position anew Real valued GAs and ES

Names of recombinations Two fixed parents Two parents selected for each i zi = (xi + yi)/2 Local intermediary Global intermediary zi is xi or yi chosen randomly Local discrete Global Real valued GAs and ES

Survivor selection Applied after creating  children from the  parents by mutation and recombination Deterministically chops off the “bad stuff” Basis of selection is either: The set of children only: (,)-selection The set of parents and children: (+)-selection Real valued GAs and ES

Survivor selection (+)-selection is an elitist strategy (,)-selection can “forget” Often (,)-selection is preferred for: Better in leaving local optima Better in following moving optima Using the + strategy bad  values can survive in x, too long if their host x is very fit Selection pressure in ES is very high Real valued GAs and ES

Survivor Selection On the other hand, (,)-selection can lead to the loss of genotypic information The (+)-selection can be preferred when are looking for “marginal” enhancements Real valued GAs and ES