Schemata Theory Chapter 11. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Why Bother with Theory? Might provide performance.

Slides:



Advertisements
Similar presentations
Genetic Algorithms Chapter 3. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Genetic Algorithms GA Quick Overview Developed: USA in.
Advertisements

Representing Hypothesis Operators Fitness Function Genetic Programming
Genetic Algorithms Representation of Candidate Solutions GAs on primarily two types of representations: –Binary-Coded –Real-Coded Binary-Coded GAs must.
Genetic Algorithms An Example Genetic Algorithm Procedure GA{ t = 0; Initialize P(t); Evaluate P(t); While (Not Done) { Parents(t) = Select_Parents(P(t));
1 Lecture 8: Genetic Algorithms Contents : Miming nature The steps of the algorithm –Coosing parents –Reproduction –Mutation Deeper in GA –Stochastic Universal.
Theory Chapter 11. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Overview (reduced w.r.t. book) Motivations and problems Holland’s.
COMP305. Part II. Genetic Algorithms. Genetic Algorithms.
Multimodal Problems and Spatial Distribution Chapter 9.
COMP305. Part II. Genetic Algorithms. Genetic Algorithms.
COMP305. Part II. Genetic Algorithms. Genetic Algorithms.
Intro to AI Genetic Algorithm Ruth Bergman Fall 2002.
Chapter 14 Genetic Algorithms.
Theory Chapter 11. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Overview Motivations and problems Holland’s Schema Theorem.
Intro to AI Genetic Algorithm Ruth Bergman Fall 2004.
Introduction to Genetic Algorithms
Chapter 6: Transform and Conquer Genetic Algorithms The Design and Analysis of Algorithms.
Genetic Algorithm.
Computer Implementation of Genetic Algorithm
Evolutionary Intelligence
Genetic algorithms. Genetic Algorithms in a slide  Premise Evolution worked once (it produced us!), it might work again  Basics Pool of solutions Mate.
© Negnevitsky, Pearson Education, Lecture 11 Evolutionary Computation: Genetic algorithms Why genetic algorithm work? Why genetic algorithm work?
Theory A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Chapter 11 1.
A Brief Introduction to GA Theory. Principles of adaptation in complex systems John Holland proposed a general principle for adaptation in complex systems:
10/12/20151 V. Evolutionary Computing A. Genetic Algorithms.
CS Machine Learning Genetic Algorithms (II).
CS 484 – Artificial Intelligence1 Announcements Lab 3 due Tuesday, November 6 Homework 6 due Tuesday, November 6 Lab 4 due Thursday, November 8 Current.
By Prafulla S. Kota Raghavan Vangipuram
Dynamical Systems Model of the Simple Genetic Algorithm Introduction to Michael Vose’s Theory Rafal Kicinger Summer Lecture Series 2002.
Theory Chapter 11. A.E. Eiben and J.E. Smith, EC Theory, modified by Ch. Eick Overview Motivations and problems Holland’s Schema Theorem – Derivation,
An Introduction to Genetic Algorithms Lecture 2 November, 2010 Ivan Garibay
1/27 Discrete and Genetic Algorithms in Bioinformatics 許聞廉 中央研究院資訊所.
Genetic Algorithms Introduction Advanced. Simple Genetic Algorithms: Introduction What is it? In a Nutshell References The Pseudo Code Illustrations Applications.
1 Machine Learning: Lecture 12 Genetic Algorithms (Based on Chapter 9 of Mitchell, T., Machine Learning, 1997)
Soft Computing A Gentle introduction Richard P. Simpson.
Genetic Algorithms Siddhartha K. Shakya School of Computing. The Robert Gordon University Aberdeen, UK
1 Chapter 14 Genetic Algorithms. 2 Chapter 14 Contents (1) l Representation l The Algorithm l Fitness l Crossover l Mutation l Termination Criteria l.
GENETIC ALGORITHM A biologically inspired model of intelligence and the principles of biological evolution are applied to find solutions to difficult problems.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Friday, 16 February 2007 William.
GENETIC ALGORITHMS.  Genetic algorithms are a form of local search that use methods based on evolution to make small changes to a popula- tion of chromosomes.
Why do GAs work? Symbol alphabet : {0, 1, * } * is a wild card symbol that matches both 0 and 1 A schema is a string with fixed and variable symbols 01*1*
1 Genetic Algorithms K.Ganesh Introduction GAs and Simulated Annealing The Biology of Genetics The Logic of Genetic Programmes Demo Summary.
Introduction to Genetic Algorithms. Genetic Algorithms We’ve covered enough material that we can write programs that use genetic algorithms! –More advanced.
Edge Assembly Crossover
Genetic Algorithms Genetic algorithms provide an approach to learning that is based loosely on simulated evolution. Hypotheses are often described by bit.
MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #12 2/20/02 Evolutionary Algorithms.
1. Genetic Algorithms: An Overview  Objectives - Studying basic principle of GA - Understanding applications in prisoner’s dilemma & sorting network.
CS621: Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 5: Power of Heuristic; non- conventional search.
1. Genetic Algorithms: An Overview 4 학습목표 GA 의 기본원리를 파악하고, Prisoner’s dilemma 와 sorting network 에의 응용 및 이론적 배경을 이해한 다.
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
GENETIC ALGORITHMS Tanmay, Abhijit, Ameya, Saurabh.
Genetic Algorithms MITM613 (Intelligent Systems).
1 Chapter 3 GAs: Why Do They Work?. 2 Schema Theorem SGA’s features: binary encoding proportional selection one-point crossover strong mutation Schema.
An Introduction to Genetic Algorithms Lecture 2 November, 2010 Ivan Garibay
Why do GAs work? Symbol alphabet : {0, 1, * } * is a wild card symbol that matches both 0 and 1 A schema is a string with fixed and variable symbols 01*1*
Genetic Algorithm Dr. Md. Al-amin Bhuiyan Professor, Dept. of CSE Jahangirnagar University.
CAP6938 Neuroevolution and Artificial Embryogeny Evolutionary Computation Theory Dr. Kenneth Stanley January 25, 2006.
Something about Building Block Hypothesis Ying-Shiuan You Taiwan Evolutionary Intelligence LAB 2009/10/31.
Artificial Intelligence By Mr. Ejaz CIIT Sahiwal Evolutionary Computation.
EVOLUTIONARY SYSTEMS AND GENETIC ALGORITHMS NAME: AKSHITKUMAR PATEL STUDENT ID: GRAD POSITION PAPER.
Genetic Algorithms And other approaches for similar applications Optimization Techniques.
Genetic Algorithm. Outline Motivation Genetic algorithms An illustrative example Hypothesis space search.
Introduction to Genetic Algorithms
Chapter 14 Genetic Algorithms.
Introduction to Genetic Algorithms and Evolutionary Computing
Dr. Kenneth Stanley September 13, 2006
SCHEMATA THEOREM (Holland)
Searching for solutions: Genetic Algorithms
A Gentle introduction Richard P. Simpson
Theory Chapter 11.
Presentation transcript:

Schemata Theory Chapter 11

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Why Bother with Theory? Might provide performance guarantees – Convergence to the global optimum can be guaranteed providing certain conditions hold Might aid better algorithm design – Increased understanding can be gained about operator interplay etc. Mathematical Models of EAs also inform theoretical biologists Because you never know ….

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory What Approach? Define similarity among individuals Study evolution of similarities between candidate solutions Study evolution of groups of similar candidate solutions

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory How Can We Describe Similarity? To describe similarity among individuals in the population we use the concept of schema (schemata) What is a schema? – Similarity template describing a subset of strings – String over the alphabet {0,1,*} – * is the don’t care symbol The don’t care symbol * is a metasymbol: – it is never explicitly process by the genetic algorithm

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Basic Idea The idea of a schema: how can we represent a subset of the population whose individuals all share some characteristics? In case the chromosomes are represented by fixed-length strings over a finite alphabet, a schema is a pattern, of the same length as a chromosome, some of whose positions are filled by elements of the alphabet and some by “don’t care” symbols. If the alphabet is a binary alphabet, a schema could be represented as (* * * 1 * * 0 0 * * 1) indicating any string of 11 binary digits with ones and zeros in the positions indicated, and anything at all (zeros or ones) in the positions containing an asterisk.

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Holland’s Schema Theorem (1975) A schema (pl. schemata) is a string in a ternary alphabet ( 0,1 # = “don’t care”) representing a hyperplane within the solution space. – E.g. 0001# #1# #0#, ##1##0## etc Two values can be used to describe schemata, – the Order (number of defined positions) = 6,2 – the Defining Length - length of sub-string between outmost defined positions = 9, 3

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Examples of Schema Schema ***11 matches, – – – – … – Schema 0***1 matches, – – – – … – 01111

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Schema and Schemata (I) Order of schema H, o(H) the number of specific (non-*) positions in the schema Examples - o(**1*0) = 2 - o(1**11) = 3 Defining length of schema H, δ(H) distance between first and last non-* in a schema Examples - δ(**1*0) = 2 - δ(1**11) = 4

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Schema and Schemata (II) Schema H1 is a competitor of schema H2 if, o H1 and H2 are specified in the same positions. o H1 and H2 differ in at least one specified position. Example o *111* is a competitor of *101* Any string of length l is an instance of 2^l (2 l ) different schemas. Implicit parallelism o One string’s fitness tells us something about the relative fitness of more than one schema.

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Example Schemata 111 1## 000 xzy # # H o(H) d(H) # 2 1 1# ## 1 0

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory How a Simple Genetic Algorithm Works? A simple genetic algorithm – Population size N – Fixed-length binary strings of length l – Fitness-proportionate selection – One-point crossover with probability p c – Bit-flip mutation – Fitness is positive How selection, crossover, and mutation work?

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Notations used Let H be a schema with at least one instance in the population at time t. Let m(H, t) be the number of instances of H at time t. Let f(H, t) be the observed (based on the sample) average fitness of (the strings of) H at time t. Let f(x) be the fitness of a string x; let be the average fitness of the population at time t. We want to calculate E(m(H, t + 1)), the expected number of instances of H at time t + 1.

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Schema Fitnesses (I) The true “fitness” of a schema H is taken by averaging over all possible values in the “don’t care” positions, but this is effectively sampled by the population, giving an estimated fitness f(H, t).

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Schema Fitnesses (I) - Exercise While optimizing a 4-bit problem, you notice that your population, of size 400, consists of 123 copies of genome 0001 (with fitness 0.12), 57 copies of genome 1111 (with fitness 0.23), 201 copies of genome 1010 (with fitness 0.56), and 19 copies of genome 0110 (with fitness 0.43). – What is the estimated fitness of schema H=1***? Answer: The estimated fitness of schema 1*** is: f(H, t) = ( 57* *0.56)/(57+201) =

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Schema Fitnesses (II) With Fitness Proportionate Selection, the expected number of offspring of a string x at time t is equal to f(x) /  proportion of H in next parent pool is: E(m(H,t+1)) = m(H,t) * f(H,t) /  E or, in other words: where x  H denotes “x is an instance of H”.

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Schema Fitnesses (II) - continued Assuming that H is a scheme that at the t generation has a fitness greater than (the population average fitness)  if f(H,t)/ >1 then E(m(H,t+1)) > m(H,t) A schema that has performance above average will have more representatives in the next generation population. With Fitness Proportionate Selection, the number of representatives for a schema that has fitness rate above average, increases exponentially.

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Schema Fitnesses (II) - Exercise While optimizing a 4-bit problem, you notice that your population, of size 400, consists of 123 copies of genome 0001 (with fitness 0.12), 57 copies of genome 1111 (with fitness 0.23), 201 copies of genome 1010 (with fitness 0.56), and 19 copies of genome 0110 (with fitness 0.43). – Assuming fitness-proportionate selection with no crossover or mutation, calculate the estimated fitness of this schema (1***) in the next generation. Answer: Let S 1 = 1111, S 2 = 1010 and – the average fitness of population in time t =( 57* * * *0.43)/400 = E(m(S 1,t+1)) = m(S 1,t) * f(S 1,t) / = 57*0.23 / = E(m(S 2,t+1)) = m(S 2,t) * f(S 2,t) / = 201*0.56 / = E(m(H,t+1)) = m(H,t) * f(H,t) /  E =258* / = Hence, the estimated fitness of schema H=1*** is: f(H, t) = ( * *0.56)/( ) =

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Schema Disruption I One Point Crossover selects a crossover point at random from the l-1 possible points (where l is the length of the schema – number of bits to represent each chromosome) For a schema with defining length d the random point will fall inside the schema with probability = d(H) / (l-1). If recombination is applied with probability P c the survival probability is P c *d(H)/(l-1)

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Schema Disruption II The probability that bit-wise mutation with probability P m will NOT disrupt the schemata is simply the probability that mutation does NOT occur in any of the defining positions, P survive (mutation) = ( 1- P m ) o(H) = 1 – o(H) * P m + terms in P m 2 +… where o(H) is the Order of H schema (number of defined positions) For low mutation rates, this survival probability under mutation approximates to 1 - o(H)* P m

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Schema Survival The probability that the schema survives both crossover and mutation is: P survive (crossover) * P survive (mutation) = = [1 - P c *d(H)/(l-1)] * ( 1- P m ) o(H)

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory The Schema Theorem (I) Put together, the proportion of a schema H in successive generations varies as: Condition for schema to increase its representation is: Inequality is due to convergence affecting crossover disruption, exact versions have been developed

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory The Schema Theorem (II) Two factors influence the number of representatives of H – Over the average schema fitness, leads to higher m(H,t+1) – Greater o(H) and δ(H), lead to a smaller m(H,t+1) Highly fit, short, low order schemata become the partial solutions to a problem (these are called building-blocks) By processing similarities, a genetic algorithm reduces the complexity of an arbitrary problems

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Criticisms of the Schema Theorem It presents an inequality that does not take into account the constructive effects of crossover and mutation – Exact versions have been derived – Have links to Price’s theorem in biology Because the mean population fitness, and the estimated fitness of a schema will vary from generation to generation, it says nothing about gen. t+2 etc. “Royal Road” problems constructed to be GA-easy based on schema theorem turned out to be better solved by random mutation Hill-Climbers BUT it remains a useful conceptual tool and has historical importance

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Building Blocks Low order, low defining-length schemata with above average fitness. Building Block Hypothesis “Short, low-order, and highly fit schemata are sampled, recombined, and resampled to form strings of potentially higher fitness […] we construct better and better strings from the best partial solutions of the past samplings.” David Goldberg, 1989

A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory How Many Schemata are Processed? A population with size n (individuals) and the individual’s length l processes between 2^l (2 l ) to n x 2 l schemata. Consider only schemata that survive with a certain probability. Count how many schemata are effectively processed Effective processing = selection + mixing (recombination). Implicit Parallelism (J. Holland) – A genetic algorithm with a population size n can usefully process on the order of O(n 3 ) schemata in a generation despite the processing of only n solutions