Decision Tree. Classification Databases are rich with hidden information that can be used for making intelligent decisions. Classification is a form of.

Slides:



Advertisements
Similar presentations
COMP3740 CR32: Knowledge Management and Adaptive Systems
Advertisements

Learning from Observations Chapter 18 Section 1 – 3.
Data Mining Lecture 9.
Rule Generation from Decision Tree Decision tree classifiers are popular method of classification due to it is easy understanding However, decision tree.
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Data Mining Techniques: Classification. Classification What is Classification? –Classifying tuples in a database –In training set E each tuple consists.
IT 433 Data Warehousing and Data Mining
Hunt’s Algorithm CIT365: Data Mining & Data Warehousing Bajuna Salehe
Decision Tree Approach in Data Mining
Classification Techniques: Decision Tree Learning
Lecture outline Classification Decision-tree classification.
Classification and Prediction
Learning From Observations
Business Systems Intelligence: 5. Classification 1
Classification Continued
LEARNING DECISION TREES
Learning decision trees derived from Hwee Tou Ng, slides for Russell & Norvig, AI a Modern Approachslides Tom Carter, “An introduction to information theory.
Learning decision trees
Learning decision trees derived from Hwee Tou Ng, slides for Russell & Norvig, AI a Modern Approachslides Tom Carter, “An introduction to information theory.
Classification.
Gini Index (IBM IntelligentMiner)
A hybrid method for gene selection in microarray datasets Yungho Leu, Chien-Pan Lee and Ai-Chen Chang National Taiwan University of Science and Technology.
Chapter 7 Decision Tree.
Data Mining: Classification
1 Tutorial Document of ITM638 Data Warehousing and Data Mining Dr. Chutima Beokhaimook 24 th March 2012.
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
Inductive learning Simplest form: learn a function from examples
LEARNING DECISION TREES Yılmaz KILIÇASLAN. Definition - I Decision tree induction is one of the simplest, and yet most successful forms of learning algorithm.
Basic Data Mining Technique
Decision Trees. 2 Outline  What is a decision tree ?  How to construct a decision tree ? What are the major steps in decision tree induction ? How to.
Learning from Observations Chapter 18 Through
CHAPTER 18 SECTION 1 – 3 Learning from Observations.
Learning from Observations Chapter 18 Section 1 – 3, 5-8 (presentation TBC)
Learning from Observations Chapter 18 Section 1 – 3.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Classification COMP Seminar BCB 713 Module Spring 2011.
Classification and Prediction Compiled By: Umair Yaqub Lecturer Govt. Murray College Sialkot Readings: Chapter 6 – Han and Kamber.
CS690L Data Mining: Classification
Chapter 20 Data Analysis and Mining. 2 n Decision Support Systems  Obtain high-level information out of detailed information stored in (DB) transaction-processing.
1 Appendix D: Application of Genetic Algorithm in Classification Duong Tuan Anh 5/2014.
Decision Tree (Rule Induction)
Chapter 6 Classification and Prediction Dr. Bernard Chen Ph.D. University of Central Arkansas.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
Chapter 6. Classification and Prediction Classification by decision tree induction Bayesian classification Rule-based classification Classification by.
Classification and Prediction
DECISION TREE Ge Song. Introduction ■ Decision Tree: is a supervised learning algorithm used for classification or regression. ■ Decision Tree Graph:
An Introduction Student Name: Riaz Ahmad Program: MSIT( ) Subject: Data warehouse & Data Mining.
Decision Tree Learning Presented by Ping Zhang Nov. 26th, 2007.
CSC 8520 Spring Paula Matuszek DecisionTreeFirstDraft Paula Matuszek Spring,
1 Classification: predicts categorical class labels (discrete or nominal) classifies data (constructs a model) based on the training set and the values.
Decision Trees.
Fundamentals, Design, and Implementation, 9/e KDD and Data Mining Instructor: Dragomir R. Radev Winter 2005.
Chapter 18 Section 1 – 3 Learning from Observations.
Learning From Observations Inductive Learning Decision Trees Ensembles.
By N.Gopinath AP/CSE.  A decision tree is a flowchart-like tree structure, where each internal node (nonleaf node) denotes a test on an attribute, each.
10. Decision Trees and Markov Chains for Gene Finding.
Chapter 6 Decision Tree.
Learning from Observations
Learning from Observations
DECISION TREES An internal node represents a test on an attribute.
Introduce to machine learning
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Presented By S.Yamuna AP/CSE
Chapter 6 Classification and Prediction
Classification and Prediction
Classification by Decision Tree Induction
Learning from Observations
©Jiawei Han and Micheline Kamber
Learning from Observations
Decision trees One possible representation for hypotheses
Presentation transcript:

Decision Tree

Classification Databases are rich with hidden information that can be used for making intelligent decisions. Classification is a form of data analysis that can be used to extract models describing important data classes.

Data classification process Learning –Training data are analyzed by a classification algorithm. Classification –Test data are used to estimate the accuracy of the classification rules. –If the accuracy is considered acceptable, the rules can be applied to the classification of new data tuples.

Expressiveness Decision trees can express any function of the input attributes. E.g., for Boolean functions, truth table row → path to leaf: Trivially, there is a consistent decision tree for any training set with one path to leaf for each example (unless f nondeterministic in x) but it probably won't generalize to new examples Prefer to find more compact decision trees

What’s a decision tree A decision tree is a flow-chart-like tree structure, where each internal node denotes a test on an attribute, each branch represents an outcome of the test, and leaf nodes represent classes or class distributions. Decision trees can easily be converted to classification rules.

Example Training data NameAgeIncomeCredit_rating Snady<=30LowFair Bill<=30LowExcellent Courney31..40HighExcellent Susan>40MedFair Claire>40MedFair Andre31..40HighExcellent …… Classification rules If age =“31..40” and income=high Then Credit_rating=excellent Test data NameAgeIncomeCredit_rating Snady<=30LowFair Bill<=30LowExcellent Courney31..40HighExcellent …… Classification algorithm New data (John, , high) Credit rating? Classification rules Excellent

Example RIDAgeIncomeStudentCredit_ratingClass: buys_computer 1<=30HighNoFairNo 2<=30HighNoExcellentNo HighNoFairYes 4>40MediumNoFairYes 5>40LowYesFairYes 6>40LowYesExcellentNo LowYesExcellentYes 8<=30MediumNoFairNo 9<=30LowYesFairYes 10>40MediumYesFairYes 11<=30MediumYesExcellentYes MediumNoExcellentYes HighYesFairYes 14>40MediumNoExcellentNo

Example Age? Student? Credit_rating? No Yes No Yes <= >40 noyes excellent fair Buys computer

Decision tree induction (ID3) Attribute selection measure –The attribute with the highest information gain (or greatest entropy reduction) is chosen as the test attribute for the current node –The expected information needed to classify a given sample is given by

Example (cont.) RIDAgeIncomeStudentCredit_ratingClass: buys_computer 1<=30HighNoFairNo 2<=30HighNoExcellentNo HighNoFairYes 4>40MediumNoFairYes 5>40LowYesFairYes 6>40LowYesExcellentNo LowYesExcellentYes 8<=30MediumNoFairNo 9<=30LowYesFairYes 10>40MediumYesFairYes 11<=30MediumYesExcellentYes MediumNoExcellentYes HighYesFairYes 14>40MediumNoExcellentNo

Example (cont.) Compute the entropy of each attribute, e.g., age –For age=“<=30”: s 11 =2, S 21 =3, I(s 11,s 21 )=0.971 –For age=“31..40”: s 12 =4, s 22 =0, I(s 12,s 22 )=0 –For age=“>40”: s 13 =3, s 23 =2, I(s 13,s 23 )=0.971 <= >40

Example (cont.) The entropy according to age is E(age)=5/14*I(s 11,s 21 )+4/14*I(s 12,s 22 )+5/14*I(s 13,s 23 ) =0.694 The information gain would be Gain(age)=I(s 1,s 2 )-E(age)=0.246 Similarly, we can compute –Gain(income)=0.029 –Gain(student)=0.151 –Gain(credit_rating)=0.048

Example (cont.) IncomeStuden t Credit_ratin g class HighNoFairNo HighNoExcellentNo MediumNoFairNo LowYesFairYes MediumYesExcellentYes IncomeStuden t Credit_ratingclass MediumNoFairYes LowYesFairYes LowYesExcellentNo MediumYesFairYes MediumNoExcellentNo IncomeStuden t Credit_ratingclas s HighNoFairYes LowYesExcellentYes MediumNoExcellentYes HighYesFairYes Age? <= >40

Decision tree learning Aim: find a small tree consistent with the training examples Idea: (recursively) choose "most significant" attribute as root of (sub)tree