Presentation is loading. Please wait.

Presentation is loading. Please wait.

Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Introduction to Data Mining By Tan, Steinbach,

Similar presentations


Presentation on theme: "Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Introduction to Data Mining By Tan, Steinbach,"— Presentation transcript:

1 Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Introduction to Data Mining By Tan, Steinbach, Kumar Lecture 3 Basic Classification © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 1

2 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 2 Classification: Definition l Given a collection of records (training set ) –Each record contains a set of attributes, one of the attributes is the class. l Find a model for class attribute as a function of the values of other attributes. l Goal: previously unseen records should be assigned a class as accurately as possible. –A test set is used to determine the accuracy of the model. Usually, the given data set is divided into training and test sets, with training set used to build the model and test set used to validate it.

3 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 3 Illustrating Classification Task

4 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 4 Examples of Classification Task l Predicting tumor cells as benign or malignant l Classifying credit card transactions as legitimate or fraudulent l Classifying secondary structures of protein as alpha-helix, beta-sheet, or random coil l Categorizing news stories as finance, weather, entertainment, sports, etc

5 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 5 Classification Techniques l Decision Tree based Methods l Rule-based Methods l Memory based reasoning l Neural Networks l Naïve Bayes and Bayesian Belief Networks l Support Vector Machines

6 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 6 Example of a Decision Tree categorical continuous class Refund MarSt TaxInc YES NO YesNo Married Single, Divorced < 80K> 80K Splitting Attributes Training Data Model: Decision Tree

7 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 7 Another Example of Decision Tree categorical continuous class MarSt Refund TaxInc YES NO Yes No Married Single, Divorced < 80K> 80K There could be more than one tree that fits the same data!

8 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 8 Decision Tree Classification Task Decision Tree

9 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 9 Apply Model to Test Data Refund MarSt TaxInc YES NO YesNo Married Single, Divorced < 80K> 80K Test Data Start from the root of tree.

10 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 10 Apply Model to Test Data Refund MarSt TaxInc YES NO YesNo Married Single, Divorced < 80K> 80K Test Data

11 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 11 Apply Model to Test Data Refund MarSt TaxInc YES NO YesNo Married Single, Divorced < 80K> 80K Test Data

12 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 12 Apply Model to Test Data Refund MarSt TaxInc YES NO YesNo Married Single, Divorced < 80K> 80K Test Data

13 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 13 Apply Model to Test Data Refund MarSt TaxInc YES NO YesNo Married Single, Divorced < 80K> 80K Test Data

14 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 14 Apply Model to Test Data Refund MarSt TaxInc YES NO YesNo Married Single, Divorced < 80K> 80K Test Data Assign Cheat to “No”

15 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 15 Decision Tree Classification Task Decision Tree

16 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 16 Decision Tree Induction l Many Algorithms: –Hunt’s Algorithm (one of the earliest) –CART –ID3, C4.5 –SLIQ, SPRINT

17 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 17 General Structure of Hunt’s Algorithm l Let D t be the set of training records that reach a node t l General Procedure: –If D t contains records that belong the same class y t, then t is a leaf node labeled as y t –If D t is an empty set, then t is a leaf node labeled by the default class, y d –If D t contains records that belong to more than one class, use an attribute test to split the data into smaller subsets. Recursively apply the procedure to each subset. DtDt ?

18 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 18 Hunt’s Algorithm Don’t Cheat Refund Don’t Cheat Don’t Cheat YesNo Refund Don’t Cheat YesNo Marital Status Don’t Cheat Single, Divorced Married Taxable Income Don’t Cheat < 80K>= 80K Refund Don’t Cheat YesNo Marital Status Don’t Cheat Single, Divorced Married

19 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 19 Tree Induction l Greedy strategy. –Split the records based on an attribute test that optimizes certain criterion. l Issues –Determine how to split the records  How to specify the attribute test condition?  How to determine the best split? –Determine when to stop splitting

20 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 20 Tree Induction l Greedy strategy. –Split the records based on an attribute test that optimizes certain criterion. l Issues –Determine how to split the records  How to specify the attribute test condition?  How to determine the best split? –Determine when to stop splitting

21 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 21 How to Specify Test Condition? l Depends on attribute types –Nominal –Ordinal –Continuous l Depends on number of ways to split –2-way split –Multi-way split

22 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 22 Splitting Based on Nominal Attributes l Multi-way split: Use as many partitions as distinct values. l Binary split: Divides values into two subsets. Need to find optimal partitioning. CarType Family Sports Luxury CarType {Family, Luxury} {Sports} CarType {Sports, Luxury} {Family} OR

23 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 23 l Multi-way split: Use as many partitions as distinct values. l Binary split: Divides values into two subsets. Need to find optimal partitioning. l What about this split? Splitting Based on Ordinal Attributes Size Small Medium Large Size {Medium, Large} {Small} Size {Small, Medium} {Large} OR Size {Small, Large} {Medium}

24 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 24 Splitting Based on Continuous Attributes l Different ways of handling –Discretization to form an ordinal categorical attribute  Static – discretize once at the beginning  Dynamic – ranges can be found by equal interval bucketing, equal frequency bucketing (percentiles), or clustering. –Binary Decision: (A < v) or (A  v)  consider all possible splits and finds the best cut  can be more compute intensive

25 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 25 Splitting Based on Continuous Attributes

26 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 26 Tree Induction l Greedy strategy. –Split the records based on an attribute test that optimizes certain criterion. l Issues –Determine how to split the records  How to specify the attribute test condition?  How to determine the best split? –Determine when to stop splitting

27 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 27 How to determine the Best Split Before Splitting: 10 records of class 0, 10 records of class 1 Which test condition is the best?

28 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 28 How to determine the Best Split l Greedy approach: –Nodes with homogeneous class distribution are preferred l Need a measure of node impurity: Non-homogeneous, High degree of impurity Homogeneous, Low degree of impurity

29 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 29 Measures of Node Impurity l Gini Index l Entropy l Misclassification error

30 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 30 How to Find the Best Split B? YesNo Node N3Node N4 A? YesNo Node N1Node N2 Before Splitting: M0 M1 M2M3M4 M12 M34 Gain = M0 – M12 vs M0 – M34

31 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 31 Measure of Impurity: GINI l Gini Index for a given node t : (NOTE: p( j | t) is the relative frequency of class j at node t). –Maximum (1 - 1/n c ) when records are equally distributed among all classes, implying least interesting information –Minimum (0.0) when all records belong to one class, implying most interesting information

32 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 32 Examples for computing GINI P(C1) = 0/6 = 0 P(C2) = 6/6 = 1 Gini = 1 – P(C1) 2 – P(C2) 2 = 1 – 0 – 1 = 0 P(C1) = 1/6 P(C2) = 5/6 Gini = 1 – (1/6) 2 – (5/6) 2 = 0.278 P(C1) = 2/6 P(C2) = 4/6 Gini = 1 – (2/6) 2 – (4/6) 2 = 0.444

33 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 33 Splitting Based on GINI l Used in CART, SLIQ, SPRINT. l When a node p is split into k partitions (children), the quality of split is computed as, where,n i = number of records at child i, n = number of records at node p.

34 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 34 Binary Attributes: Computing GINI Index l Splits into two partitions l Effect of Weighing partitions: –Larger and Purer Partitions are sought for. B? YesNo Node N1Node N2 Gini(N1) = 1 – (5/7) 2 – (2/7) 2 = 0.4082 Gini(N2) = 1 – (1/5) 2 – (4/5) 2 = 0.32 Gini(Children) = 7/12 * 0.4082 + 5/12 * 0.32 = 0.3715

35 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 35 Categorical Attributes: Computing Gini Index l For each distinct value, gather counts for each class in the dataset l Use the count matrix to make decisions Multi-way splitTwo-way split (find best partition of values)

36 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 36 Continuous Attributes: Computing Gini Index l Use Binary Decisions based on one value l Several Choices for the splitting value –Number of possible splitting values = Number of distinct values l Each splitting value has a count matrix associated with it –Class counts in each of the partitions, A < v and A  v l Simple method to choose best v –For each v, scan the database to gather count matrix and compute its Gini index –Computationally Inefficient! Repetition of work.

37 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 37 Continuous Attributes: Computing Gini Index... l For efficient computation: for each attribute, –Sort the attribute on values –Linearly scan these values, each time updating the count matrix and computing gini index –Choose the split position that has the least gini index Split Positions Sorted Values

38 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 38 Alternative Splitting Criteria based on INFO l Entropy at a given node t: (NOTE: p( j | t) is the relative frequency of class j at node t). –Measures homogeneity of a node.  Maximum (log n c ) when records are equally distributed among all classes implying least information  Minimum (0.0) when all records belong to one class, implying most information –Entropy based computations are similar to the GINI index computations

39 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 39 Examples for computing Entropy P(C1) = 0/6 = 0 P(C2) = 6/6 = 1 Entropy = – 0 log 0 – 1 log 1 = – 0 – 0 = 0 P(C1) = 1/6 P(C2) = 5/6 Entropy = – (1/6) log 2 (1/6) – (5/6) log 2 (5/6) = 0.65 P(C1) = 2/6 P(C2) = 4/6 Entropy = – (2/6) log 2 (2/6) – (4/6) log 2 (4/6) = 0.92

40 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 40 Splitting Based on INFO... l Information Gain: Parent Node, p is split into k partitions; n i is number of records in partition i –Measures Reduction in Entropy achieved because of the split. Choose the split that achieves most reduction (maximizes GAIN) –Used in ID3 and C4.5 –Disadvantage: Tends to prefer splits that result in large number of partitions, each being small but pure.

41 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 41 Splitting Based on INFO... l Gain Ratio: Parent Node, p is split into k partitions n i is the number of records in partition i –Adjusts Information Gain by the entropy of the partitioning (SplitINFO). Higher entropy partitioning (large number of small partitions) is penalized! –Used in C4.5 –Designed to overcome the disadvantage of Information Gain

42 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 42 Splitting Criteria based on Classification Error l Classification error at a node t : l Measures misclassification error made by a node.  Maximum (1 - 1/n c ) when records are equally distributed among all classes, implying least interesting information  Minimum (0.0) when all records belong to one class, implying most interesting information

43 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 43 Examples for Computing Error P(C1) = 0/6 = 0 P(C2) = 6/6 = 1 Error = 1 – max (0, 1) = 1 – 1 = 0 P(C1) = 1/6 P(C2) = 5/6 Error = 1 – max (1/6, 5/6) = 1 – 5/6 = 1/6 P(C1) = 2/6 P(C2) = 4/6 Error = 1 – max (2/6, 4/6) = 1 – 4/6 = 1/3

44 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 44 Comparison among Splitting Criteria For a 2-class problem:

45 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 45 Misclassification Error vs Gini A? YesNo Node N1Node N2 Error(N1) = 1 – max[(3/3),(0/3)] = 1 - 1 = 0 Error(N2) = 1 – max[(4/7), (3/7)] = 1 – 4/7 = 3/7 Error(Children) = 3/10 * 0 + 7/10 * [3/7] = 0.3

46 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 46 Misclassification Error vs Gini A? YesNo Node N1Node N2 Gini(N1) = 1 – (3/3) 2 – (0/3) 2 = 0 Gini(N2) = 1 – (4/7) 2 – (3/7) 2 = 0.489 Gini(Children) = 3/10 * 0 + 7/10 * 0.489 = 0.342 Gini improves

47 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 47 Tree Induction l Greedy strategy. –Split the records based on an attribute test that optimizes certain criterion. l Issues –Determine how to split the records  How to specify the attribute test condition?  How to determine the best split? –Determine when to stop splitting

48 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 48 Stopping Criteria for Tree Induction l Stop expanding a node when all the records belong to the same class l Stop expanding a node when all the records have similar attribute values l Early termination (to be discussed later)

49 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 49 Decision Tree Based Classification l Advantages: –Inexpensive to construct –Extremely fast at classifying unknown records –Easy to interpret for small-sized trees –Accuracy is comparable to other classification techniques for many simple data sets

50 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 50 An example: Buys_Computer

51 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 51 Output: A Decision Tree for “buys_computer” age? overcast student?credit rating? <=30 >40 noyes 31... 40 no fairexcellent yesno

52 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 52 Attribute Selection Measure: Information Gain (ID3/C4.5) Select the attribute with the highest information gain Let p i be the probability that an arbitrary tuple in D belongs to class C i, estimated by |C i, D |/|D| Expected information (entropy) needed to classify a tuple in D: Information needed (after using A to split D into v partitions) to classify D: Information gained by branching on attribute A

53 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 53 Attribute Selection: Information Gain  Class P: buys_computer = “yes”  Class N: buys_computer = “no” means “age <=30” has 5 out of 14 samples, with 2 yes’es and 3 no’s. Hence Similarly,

54 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 54 Computing Information-Gain for Continuous-Value Attributes l Let attribute A be a continuous-valued attribute l Must determine the best split point for A –Sort the value A in increasing order –Typically, the midpoint between each pair of adjacent values is considered as a possible split point  (a i +a i+1 )/2 is the midpoint between the values of a i and a i+1 –The point with the minimum expected information requirement for A is selected as the split-point for A l Split: –D1 is the set of tuples in D satisfying A ≤ split-point, and D2 is the set of tuples in D satisfying A > split-point

55 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 55 Gain Ratio for Attribute Selection (C4.5) l Information gain measure is biased towards attributes with a large number of values l C4.5 (a successor of ID3) uses gain ratio to overcome the problem (normalization to information gain) –GainRatio(A) = Gain(A)/SplitInfo(A) l Ex. –gain_ratio(income) = 0.029/0.926 = 0.031 l The attribute with the maximum gain ratio is selected as the splitting attribute

56 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 56 Gini index (CART, IBM IntelligentMiner) l If a data set D contains examples from n classes, gini index, gini(D) is defined as where p j is the relative frequency of class j in D l If a data set D is split on A into two subsets D 1 and D 2, the gini index gini(D) is defined as l Reduction in Impurity: l The attribute provides the smallest gini split (D) (or the largest reduction in impurity) is chosen to split the node (need to enumerate all the possible splitting points for each attribute)

57 © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 57 Decision Tree Induction by C4.5 l Simple depth-first construction. l Uses Information Gain l Sorts Continuous Attributes at each node. l Needs entire data to fit in memory. l Unsuitable for Large Datasets. –Needs out-of-core sorting. l You can download the software from: http://www.cse.unsw.edu.au/~quinlan/c4.5r8.tar.gz http://www.cse.unsw.edu.au/~quinlan/c4.5r8.tar.gz


Download ppt "Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Introduction to Data Mining By Tan, Steinbach,"

Similar presentations


Ads by Google