GATree: Genetically Evolved Decision Trees 2012. 05. 02 전자전기컴퓨터공학과 데이터베이스 연구실 G201149006 김태종.

Slides:



Advertisements
Similar presentations
Random Forest Predrag Radenković 3237/10
Advertisements

CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Huffman code and ID3 Prof. Sin-Min Lee Department of Computer Science.
Decision Tree Approach in Data Mining
On the Genetic Evolution of a Perfect Tic-Tac-Toe Strategy
Decision Tree Learning 主講人:虞台文 大同大學資工所 智慧型多媒體研究室.
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
Feature Selection for Regression Problems
A new crossover technique in Genetic Programming Janet Clegg Intelligent Systems Group Electronics Department.
Genetic Algorithm Genetic Algorithms (GA) apply an evolutionary approach to inductive learning. GA has been successfully applied to problems that are difficult.
Genetic Algorithm What is a genetic algorithm? “Genetic Algorithms are defined as global optimization procedures that use an analogy of genetic evolution.
Genetic Algorithms Overview Genetic Algorithms: a gentle introduction –What are GAs –How do they work/ Why? –Critical issues Use in Data Mining –GAs.
Genetic Programming.
Attention Deficit Hyperactivity Disorder (ADHD) Student Classification Using Genetic Algorithm and Artificial Neural Network S. Yenaeng 1, S. Saelee 2.
Game Trees: MiniMax strategy, Tree Evaluation, Pruning, Utility evaluation Adapted from slides of Yoonsuck Choe.
Genetic Algorithm.
Issues with Data Mining
Breeding Decision Trees Using Evolutionary Techniques Written by Athanasios Papagelis Dimitris Kalles Presented by Alexandre Temporel.
Minimax Trees: Utility Evaluation, Tree Evaluation, Pruning CPSC 315 – Programming Studio Spring 2008 Project 2, Lecture 2 Adapted from slides of Yoonsuck.
Machine Learning Chapter 3. Decision Tree Learning
Efficient Model Selection for Support Vector Machines
Breeding Decision Trees Using Evolutionary Techniques Papagelis Athanasios - Kalles Dimitrios Computer Technology Institute & AHEAD RM.
Mohammad Ali Keyvanrad
Genetic Algorithms CS121 Spring 2009 Richard Frankel Stanford University 1.
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
© Negnevitsky, Pearson Education, Lecture 10 Evolutionary Computation: Evolution strategies and genetic programming Evolution strategies Evolution.
CS 484 – Artificial Intelligence1 Announcements Lab 4 due today, November 8 Homework 8 due Tuesday, November 13 ½ to 1 page description of final project.
Chapter 9 – Classification and Regression Trees
CS 484 – Artificial Intelligence1 Announcements Lab 3 due Tuesday, November 6 Homework 6 due Tuesday, November 6 Lab 4 due Thursday, November 8 Current.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
Optimizing Sorting With Genetic Algorithms Xiaoming Li, María Jesús Garzarán, and David Padua University of Illinois at Urbana-Champaign.
Inferring Decision Trees Using the Minimum Description Length Principle J. R. Quinlan and R. L. Rivest Information and Computation 80, , 1989.
Genetic algorithms Charles Darwin "A man who dares to waste an hour of life has not discovered the value of life"
Learning from Observations Chapter 18 Through
Decision Trees DefinitionDefinition MechanismMechanism Splitting FunctionSplitting Function Issues in Decision-Tree LearningIssues in Decision-Tree Learning.
1 Machine Learning: Lecture 12 Genetic Algorithms (Based on Chapter 9 of Mitchell, T., Machine Learning, 1997)
GENETIC ALGORITHM A biologically inspired model of intelligence and the principles of biological evolution are applied to find solutions to difficult problems.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Friday, 16 February 2007 William.
Genetic Algorithms. Evolutionary Methods Methods inspired by the process of biological evolution. Main ideas: Population of solutions Assign a score or.
Indexing and hashing Azita Keshmiri CS 157B. Basic concept An index for a file in a database system works the same way as the index in text book. For.
For Monday No new reading Homework: –Chapter 18, exercises 3 and 4.
Chapter 9 Genetic Algorithms.  Based upon biological evolution  Generate successor hypothesis based upon repeated mutations  Acts as a randomized parallel.
MACHINE LEARNING 10 Decision Trees. Motivation  Parametric Estimation  Assume model for class probability or regression  Estimate parameters from all.
Slides for “Data Mining” by I. H. Witten and E. Frank.
1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 5-Inducción de árboles de decisión (2/2) Eduardo Poggi.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
EE749 I ntroduction to Artificial I ntelligence Genetic Algorithms The Simple GA.
GENETIC ALGORITHM Basic Algorithm begin set time t = 0;
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
Adversarial Search 2 (Game Playing)
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
CS-424 Gregory Dudek Lecture 10 Annealing (final comments) Adversary Search Genetic Algorithms (genetic search)
Dr. Chen, Data Mining  A/W & Dr. Chen, Data Mining Chapter 3 Basic Data Mining Techniques Jason C. H. Chen, Ph.D. Professor of MIS School of Business.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
On the Ability of Graph Coloring Heuristics to Find Substructures in Social Networks David Chalupa By, Tejaswini Nallagatla.
Genetic Algorithm. Outline Motivation Genetic algorithms An illustrative example Hypothesis space search.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Intelligent Exploration for Genetic Algorithms Using Self-Organizing.
Machine Learning: Ensemble Methods
Machine Learning in Practice Lecture 18
Genetic-Algorithm-Based Instance and Feature Selection
Rule Induction for Classification Using
Presented by: Dr Beatriz de la Iglesia
Issues in Decision-Tree Learning Avoiding overfitting through pruning
Data Mining Lecture 11.
Advanced Artificial Intelligence Evolutionary Search Algorithm
Introduction to Data Mining, 2nd Edition by
Introduction to Data Mining, 2nd Edition by
Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen, Zne-Jung Lee
GATree Genetically Evolved Decision Trees
M. Kezunovic (P.I.) S. S. Luo D. Ristanovic Texas A&M University
Presentation transcript:

GATree: Genetically Evolved Decision Trees 전자전기컴퓨터공학과 데이터베이스 연구실 G 김태종

GATree SYSTEM Genetic Algorithms (GAs) have been widely used as an effective search technique, especially when the search space contains complex interacting parts. Decision tree induction is a very popular and practical method for pattern classification. This work is an attempt to overcome the use of greedy heuristics and search the decision tree space in a more natural way. More specifically, we make use of genetic algorithms to directly evolve binary decision trees in the conquest for the one that most closely matches the target concept.

Traditionally, GAs use binary strings to represent points in search space. However, such representations do not appear well suited for representing the space of concept descriptions that are generally symbolic in nature and with varying length and complexity. There are two different approaches one might take to resolve this issue. The first involves changing the fundamental GA operators so as to work well with the complex non-string objects, while the second attempts to construct string representations of solutions that minimize any changes to the basic GA philosophy. REPRESENTATION ISSUES

We stuck with the first approach for three fundamental reasons. First, it is natural to use a tree structure to represent decision trees and the mutation-crossover operators can be efficiently altered to match this structure. Second, it is not trivial to alter the basic mutation-crossover operators so as to be used with string representatives of decision trees and at the same time preserve trees structures. Finally, libraries of GA’s components emerge today that give the option of alternative internal representations and can substantially decrease the overhead of deriving the needed tuning of GA’s operators. REPRESENTATION ISSUES

Mutation chooses a random node of a desired tree and it replaces that node’s test-value with a new random chosen value. When the random node is a leaf, it replaces the installed class with a new random chosen class (Figure 1). The crossover operator chooses two random nodes and swaps those nodes’ sub-trees. Since predicted values rest only on leaves, the crossover operator does not affect the decision tree’s coherence (Figure 2). DATA PREPROCESSING AND GENETIC OPERATORS

PAYOFF FUNCTION A natural way to assign utility to a random decision tree is by using it to classify the known instance-set. Then we grant a scaled payoff to the best candidates. This is a way to avoid unnecessary test-values replications along a specific path while at the same time we derive comprehensible decision trees. Thus, the fitness function is balanced between accuracy and size: When the size of the tree is small the size factor is near one, while it decreases when the tree grows big. This way, the payoff is greater for smaller trees. PAYOFF FUNCTION

We have set the basic requirements for our genetic algorithm: The size of search space depends on tree size. Let D(n) be the number of topologically different binary decision trees of n leaves. SEARCH SPACE AND INDUCTION COSTS

The search space depends also on the amount of different attribute-values and classes of the underlying concept. a is the sum of the distinct values² of all features c is the distinct problem classes. An internal node can use any one of the a distinct values and that holds for every node. Since we allow values to be reused, a binary decision tree of n leaves has aⁿ‾¹ syntactically different trees regarding the attribute values. SEARCH SPACE AND INDUCTION COSTS

We search on a space containing a wide range of tree sizes. Assuming that the number of training instances is k, the maximum number of leaves is also k (one instance at every leaf). Thus, the size of the search space is: SEARCH SPACE AND INDUCTION COSTS

The algorithm’s parameters during the experiments are presented in Table 1. EXPERIMENTS

To ensure that the GA produces fit hypotheses we tested its performance with three synthetic datasets. All datasets had 3 attributes (A, B, C), that could take up to 26 distinct values (a…z) and 3 problem classes (c1, c2, c3). For those experiments we set the number of generations to 800. The exact activation rules of the first synthetic dataset are presented below: (31.0%) c1 <- B=(f or g or j) & C=(a or g or j) (28.0%) c2 <- C=(b or e) (41.0%) c3 <- B=(b or i) & C=(d or i) HYPOTHESES FITNESS

Mean Fitness refers to the average fitness score of all genomes inside the current population. Fitness is the fitness score of the best individual. Accuracy is the obtained classification accuracy using the best genome and Size is the number of nodes of the best individual. Figure 4 illustrates the final decision tree. HYPOTHESES FITNESS

Figure 5 presents the results for a more complex artificial dataset. The dataset was created using eight rules. the first two activation-rules were as below: (15.0%) c1 <- A=(a or b or t) & B=(a or h or q or x) (14.0%) c1 <- B=(f or l or s or w) & C=(c or e or f or k) HYPOTHESES FITNESS

Figure 6 presents the results for the most complex artificial concept we used. The dataset was created from twelve activation rules. The first two of them were as below: (13.0%) c1<- A=(i or k) & C=(a or c or e or h) (11.0%) c2<- A=(d or e or h or o) & B=(d or g or h) & C=(j or k or m) HYPOTHESES FITNESS

Consider the example data set given in Table 2 CONDITIONALLY DEPENDENT AND IRRELEVANT ATTRIBUTES

The less greedy strategy of GATree (which tries to minimize a tree’s size while at the same time maximize accuracy) easily discovers the desired decision tree (Figure 7) CONDITIONALLY DEPENDENT AND IRRELEVANT ATTRIBUTES

We created several artificial datasets with strong dependent and irrelevant attributes. CONDITIONALLY DEPENDENT AND IRRELEVANT ATTRIBUTES

The mean accuracy results of standard 5-fold cross validation are presented in Table 4. CONDITIONALLY DEPENDENT AND IRRELEVANT ATTRIBUTES

Table 5 presents the classification accuracy results while Table 6 presents the derived decision trees size for GATree and C4.5 (with pruning). EXPERIMENTS WITH STANDARD DATASETS

Conclusion In this work we have explored how GAs can be used to directly evolve decision trees. The whole approach is based on conceptual simplicity, adopting only necessary extensions to basic GAs and small a priori bias. Furthermore, experiments demonstrated the implications of adopting greedy procedural biases. Surely, the proposed approach has several childhood flaws. Still, those flows can be remedied by further work and this paper suggests a number of research topics towards this direction. Conclusion