Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-1 Using Localised.

Slides:



Advertisements
Similar presentations
Data points are spread over the space according to two of their component values Using real data sets to simulate evolution within complex environments.
Advertisements

Random Testing Tor Stålhane Jonas G. Brustad. What is random testing The principle of random testing is simple and can be described as follows: 1.For.
A New Algorithm of Fuzzy Clustering for Data with Uncertainties: Fuzzy c-Means for Data with Tolerance Defined as Hyper-rectangles ENDO Yasunori MIYAMOTO.
Using Parallel Genetic Algorithm in a Predictive Job Scheduling
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Using data sets to simulate evolution within complex environments Bruce Edmonds Centre for Policy Modelling Manchester Metropolitan University.
21-May-15 Genetic Algorithms. 2 Evolution Here’s a very oversimplified description of how evolution works in biology Organisms (animals or plants) produce.
Genetic Algorithms as a Tool for General Optimization Angel Kuri 2001.
1 Analysis of Variance This technique is designed to test the null hypothesis that three or more group means are equal.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
A new crossover technique in Genetic Programming Janet Clegg Intelligent Systems Group Electronics Department.
Basic Data Mining Techniques Chapter Decision Trees.
Basic Data Mining Techniques
Evolutionary Algorithms Simon M. Lucas. The basic idea Initialise a random population of individuals repeat { evaluate select vary (e.g. mutate or crossover)
Imagine that I am in a good mood Imagine that I am going to give you some money ! In particular I am going to give you z dollars, after you tell me the.
Special Topic: Missing Values. Missing Values Common in Real Data  Pneumonia: –6.3% of attribute values are missing –one attribute is missing in 61%
Sampling Theory Determining the distribution of Sample statistics.
Chapter 5 Data mining : A Closer Look.
Genetic Algorithm.
Efficient Model Selection for Support Vector Machines
Inductive learning Simplest form: learn a function from examples
by B. Zadrozny and C. Elkan
Adapting Convergent Scheduling Using Machine Learning Diego Puppin*, Mark Stephenson †, Una-May O’Reilly †, Martin Martin †, and Saman Amarasinghe † *
Statistics Chapter 9. Statistics Statistics, the collection, tabulation, analysis, interpretation, and presentation of numerical data, provide a viable.
Stochastic Algorithms Some of the fastest known algorithms for certain tasks rely on chance Stochastic/Randomized Algorithms Two common variations – Monte.
Cristian Urs and Ben Riveira. Introduction The article we chose focuses on improving the performance of Genetic Algorithms by: Use of predictive models.
SOFT COMPUTING (Optimization Techniques using GA) Dr. N.Uma Maheswari Professor/CSE PSNA CET.
Experiments in Machine Learning COMP24111 lecture 5 Accuracy (%) A BC D Learning algorithm.
What is Genetic Programming? Genetic programming is a model of programming which uses the ideas (and some of the terminology) of biological evolution to.
Genetic algorithms Charles Darwin "A man who dares to waste an hour of life has not discovered the value of life"
Biological data mining by Genetic Programming AI Project #2 Biointelligence lab Cho, Dong-Yeon
1 Machine Learning: Lecture 12 Genetic Algorithms (Based on Chapter 9 of Mitchell, T., Machine Learning, 1997)
Measures of Dispersion
Clustering What is clustering? Also called “unsupervised learning”Also called “unsupervised learning”
2005MEE Software Engineering Lecture 11 – Optimisation Techniques.
Biologically Inspired Computing: Optimisation This is mandatory additional material for `Biologically Inspired Computing’ Contents: Optimisation; hard.
Learning by Simulating Evolution Artificial Intelligence CSMC February 21, 2002.
Mining Binary Constraints in Feature Models: A Classification-based Approach Yi Li.
Evolutionary Algorithms for Finding Optimal Gene Sets in Micro array Prediction. J. M. Deutsch Presented by: Shruti Sharma.
Chapter 9 Genetic Algorithms.  Based upon biological evolution  Generate successor hypothesis based upon repeated mutations  Acts as a randomized parallel.
Introduction to Genetic Algorithms. Genetic Algorithms We’ve covered enough material that we can write programs that use genetic algorithms! –More advanced.
Project 2: Classification Using Genetic Programming Kim, MinHyeok Biointelligence laboratory Artificial.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
Coevolutionary Automated Software Correction Josh Wilkerson PhD Candidate in Computer Science Missouri S&T.
Vector Quantization CAP5015 Fall 2005.
Computational Intelligence: Methods and Applications Lecture 33 Decision Tables & Information Theory Włodzisław Duch Dept. of Informatics, UMK Google:
Biologically Inspired Computing: Evolutionary Algorithms: Some Details and Examples This is additional material for lecture 2 of `Biologically Inspired.
Evolving RBF Networks via GP for Estimating Fitness Values using Surrogate Models Ahmed Kattan Edgar Galvan.
Discovering Interesting Patterns for Investment Decision Making with GLOWER-A Genetic Learner Overlaid With Entropy Reduction Advisor : Dr. Hsu Graduate.
Chapter 4 Statistical Inference  Estimation -Confidence interval estimation for mean and proportion -Determining sample size  Hypothesis Testing -Test.
Symbolic Regression via Genetic Programming AI Project #2 Biointelligence lab Cho, Dong-Yeon
Dr. Chen, Data Mining  A/W & Dr. Chen, Data Mining Chapter 3 Basic Data Mining Techniques Jason C. H. Chen, Ph.D. Professor of MIS School of Business.
If we are all the same species (Homo sapien), why don’t we all look the same?
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Data Mining CH6 Implementation: Real machine learning schemes(2) Reporter: H.C. Tsai.
Review Design of experiments, histograms, average and standard deviation, normal approximation, measurement error, and probability.
Stats 242.3(02) Statistical Theory and Methodology.
Clustering CSC 600: Data Mining Class 21.
Data Mining: Concepts and Techniques
Medical Diagnosis via Genetic Programming
Experiments in Machine Learning
Gathering and Organizing Data
`Biologically Inspired Computing’
EE368 Soft Computing Genetic Algorithms.
Searching for solutions: Genetic Algorithms
This is additional material for week 2 of
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Coevolutionary Automated Software Correction
Presentation transcript:

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-1 Using Localised ‘Gossip’ to Structure Distributed Learning Bruce Edmonds Centre for Policy Modelling

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-2 The Problem For many problems/situations universal solutions are unreachable …in such situations one has to seek partial solutions (i.e. solutions that are valid/effective only in a subdomain). Sometimes the relevant subdomains seem obvious (e.g. biology vs. physics) …but in many other situations the best way to subdivide a situation also needs to be discovered (entangled with solution types).

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-3 Fitting data globally and piecewise Data points Problem Domain Graph of global candidate model Graphs of piecewise models

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-4 Solution source Both ecology and human society inhabit situations where universal solutions are not reachable Even closely related species are successful in different regions and niches Human techniques for dealing with the environment have spread over the areas where these techniques work

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-5 Cavalli-Sforza, Menozzi, and Piazza 1994 p. 257 – Cultural Diffusion

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-6 Beef Cows in the USA 2002

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-7 Milk Cows in the USA 2002

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-8 Change in the use of irrigation in USA

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-9 Different ranges of different species Greenstriped grasshopper Striped grasshopper

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-10 Distribution of terms for soft drinks in the USA – Matthew Campbell’s map

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-11 Only occasionally do global parasites arise… …like homo sapiens!

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-12 An Illustration of the Basic Algorithm Some Space of Characteristics D p (Learning Domain & Content)

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-13 The algorithm outline (generic version) Initialise space with a random set of genes Repeat ForEach gene from 1 to popSize Randomly select a locality randomly select from locality a set of sample genes evaluate set in the locality chose two best from set if randomNum < probCrossover then cross two best -> newInd else best -> newInd Next gene New population composed of newInds Until finished

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-14 Two phases of this approach When species successfully propagate over regions they tend to “speciate” into many varieties Information learnt is spread over the population not in a single best individual Thus if you want to understand the results it is helpful to add an “analysis” phase …which does a sort of “cluster analysis” of the locally best solutions in the population I do this by: turning off variation; allowing only one solution per location; and massive but strictly local propagation to nearby locations (in this 2 nd phase)

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-15 An application to the Cleveland Heart Disease Data Set

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-16 Cleveland Heart Disease Data Set – the processed sub-set used In processed sub-set: 281 entries 14 attributes numeric or numerically coded Attribute 14 is the outcome (0, 1, 2, 3, 4) Some attributes: 1 - age, 2 - sex, 4 - resting blood pressure (trestpbs), 5 - cholesterol (chol) Available at the repository of Machine Learning

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-17 Why this particular data set? It is fairly large It is quite complex I know hardly anything about the causes of heart disease Its accessible ML techniques so far have not found a very high performing global solution It seemed a vaguely useful thing to do

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-18 The Solution Form Solutions are a set of 5 numeric functions (one for each outcome), each coded as tree expressions –E.g. Outcome 0 has weight calculated by: [TIMES [MIN [CONST -0.6] [INPUT 8]] [SAFEDIVIDE [INPUT 1] [CONST 0.5]]] –Which simplifies to: 2 * V8 * V1 –Each of 5 functions evaluated (given 13 inputs) –Function with highest value gives prediction Functions: MIN, MAX, IGZ, TIMES, MINUS, PLUS, SAFEDIVIDE Leaves: inputs 1,2,…,13 and constants -1, -.9,.., 1

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-19 The space of characteristics Is essentially the 281 points in the data set …with the distance structure determined by the cartesian distance within the chosen space of characteristics

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-20 The 3 sets of runs (12 runs each) Global: a standard GP approach, evaluation against 10% random sample, population of 281, 90% crossover Local: set of solutions evaluated at a point in the space, taken from point plus some from neighbouring localities, population 800, 20% crossover –Local (1, 2): space defined by age and sex –Local (4, 5): space defined by restbps and chol

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-21 Measuring the success Cost of each approach measured in terms of the number of evaluations of a solution at a point in the space, since this dominates the computational time Effective error is: Global runs: the average error (over all points) of the best solution in the population Local runs: the average of the error of set of the best solution at each point evaluated at that point

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-22 Comparison of global and local runs

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-23 Error and Spread in Local(1, 2)

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-24 Error and Spread in Local(4, 5)

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-25 Spread of solutions using items 1&2 Male Both Female Age

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-26 Spread of solutions using items 4&5 resting blood pressure cholesterol

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-27 Related Work Local Regression (or the slightly more general locally weighted learning) Clustering techniques ‘Demes’ in GP Evolving parts of a problem separately (DCCGA etc.) Decision tree induction (e.g. C4.5) Ecological models

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-28 Conclusion Memetic/ecological processes that combine local propagation and solution development can find and exploit niches in complex problems …but this does not lead to neat global solutions (in cases I have tried) …and can be sensitive to the selection of the space over which propagation occurs (although am investigating systems where this is also discovered, so wish me luck!)

Using localised gossip to structure distributed learning, Bruce Edmonds, Univ. of Herts., April 2005, slide-29 The End Bruce Edmonds bruce.edmonds.name Centre for Policy Modelling cfpm.org