More on Decision Trees. Numerical attributes Tests in nodes can be of the form x j > constant Divides the space into rectangles.

Slides:



Advertisements
Similar presentations
Selection Statements Make a decision based on conditions Allows the computer to be intelligent.
Advertisements

COMP3740 CR32: Knowledge Management and Adaptive Systems
The Software Infrastructure for Electronic Commerce Databases and Data Mining Lecture 4: An Introduction To Data Mining (II) Johannes Gehrke
Data Mining Lecture 9.
CHAPTER 9: Decision Trees
Mining High-Speed Data Streams
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Lower bound for sorting, radix sort COMP171 Fall 2005.
CS 206 Introduction to Computer Science II 04 / 27 / 2009 Instructor: Michael Eckmann.
Regression. So far, we've been looking at classification problems, in which the y values are either 0 or 1. Now we'll briefly consider the case where.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
UPGMA Algorithm.  Main idea: Group the taxa into clusters and repeatedly merge the closest two clusters until one cluster remains  Algorithm  Add a.
Lower bound for sorting, radix sort COMP171 Fall 2006.
ID3 Algorithm Abbas Rizvi CS157 B Spring What is the ID3 algorithm? ID3 stands for Iterative Dichotomiser 3 Algorithm used to generate a decision.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Decision Trees (2). Numerical attributes Tests in nodes are of the form f i > constant.
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
Ordered and Quantum Treemaps: Making effective use of 2D space to display hierarchies B. Bederson, B. Shneiderman, M. Wattenberg ACM Trans. On Graphics.
Supervised Learning I, Cont’d Reading: Bishop, Ch 14.4, 1.6, 1.5.
On the Application of Artificial Intelligence Techniques to the Quality Improvement of Industrial Processes P. Georgilakis N. Hatziargyriou Schneider ElectricNational.
ID3 Algorithm Allan Neymark CS157B – Spring 2007.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
Decision Trees Jyh-Shing Roger Jang ( 張智星 ) CSIE Dept, National Taiwan University.
Mathematical Induction II Lecture 21 Section 4.3 Mon, Feb 27, 2006.
Chapter 9 (modified) Abstract Data Types and Algorithms Nell Dale John Lewis.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.3: Decision Trees Rodney Nielsen Many of.
Equation Jeopardy Add Mixed Multiply/ Divide Fractions Q $100 Q $200 Q $300 Q $400 Q $500 Q $100 Q $200 Q $300 Q $400 Q $500 Final Jeopardy Subtract.
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
CS 206 Introduction to Computer Science II 04 / 22 / 2009 Instructor: Michael Eckmann.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
Hank Childs, University of Oregon Isosurfacing (Part 3)
R-Trees: A Dynamic Index Structure For Spatial Searching Antonin Guttman.
DECISION TREE Ge Song. Introduction ■ Decision Tree: is a supervised learning algorithm used for classification or regression. ■ Decision Tree Graph:
Red-Black trees Red-black trees are trees represented as binary trees at the expense of one extra bit per node. The idea is to represent 3- and 4-nodes.
1 Illustration of the Classification Task: Learning Algorithm Model.
Bootstrapped Optimistic Algorithm for Tree Construction
Part of a set or part of a whole. 3 4 =Numerator the number of parts = Denominator the number that equals the whole.
CS4445/B12 Provided by: Kenneth J. Loomis. genrecritics-reviewsratingIMAXlikes comedythumbs-upRFALSEno comedythumbs-upRTRUEno comedyneutralRFALSEno actionthumbs-downPG-13TRUEno.
1 Decision Trees. 2 OutlookTemp (  F) Humidity (%) Windy?Class sunny7570true play sunny8090true don’t play sunny85 false don’t play sunny7295false don’t.
Machine Learning in Practice Lecture 8 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Ensemble Learning, Boosting, and Bagging: Scaling up Decision Trees (with thanks to William Cohen of CMU, Michael Malohlava of 0xdata, and Manish Amde.
Chapter 3 Data Mining: Classification & Association Chapter 4 in the text box Section: 4.3 (4.3.1),
Data Mining Chapter 4 Algorithms: The Basic Methods - Constructing decision trees Reporter: Yuen-Kuei Hsueh Date: 2008/7/24.
By N.Gopinath AP/CSE.  A decision tree is a flowchart-like tree structure, where each internal node (nonleaf node) denotes a test on an attribute, each.
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
DECISION TREES An internal node represents a test on an attribute.
Decision Trees an introduction.
Decision Trees.
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Artificial Intelligence
Ch9: Decision Trees 9.1 Introduction A decision tree:
Binary Search Tree (BST)
Data Science Algorithms: The Basic Methods
Searching.
Classification by Decision Tree Induction
Minimizing DFA’s By Partitioning.
Chapter 1-Lesson 1 Position and Motion Vocabulary
Searching CLRS, Sections 9.1 – 9.3.
Machine Learning in Practice Lecture 23
Lower bound for sorting, radix sort
Publisher : TRANSACTIONS ON NETWORKING Author : Haoyu Song, Jonathan S
Simplifying rational expressions
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Mathematical Induction II
… 1 2 n A B V W C X 1 2 … n A … V … W … C … A X feature 1 feature 2
Data Mining CSCI 307, Spring 2019 Lecture 11
Data Mining CSCI 307, Spring 2019 Lecture 6
Presentation transcript:

More on Decision Trees

Numerical attributes Tests in nodes can be of the form x j > constant Divides the space into rectangles.

Numerical attributes Tests in nodes can be of the form x j > constant Divides the space into rectangles.

Considering splits The only thing we really need to do differently in our algorithm is to consider splitting between each data point in each dimension. So, in our bankruptcy domain, we'd consider 9 different splits in the R dimension –In general, we'd expect to consider m - 1 splits, if we have m data points; –But in our data set we have some examples with equal R values.

Considering splits II And there are another 6 possible splits in the L dimension –because L is an integer, really, there are lots of duplicate L values.

Bankruptcy Example

We consider all the possible splits in each dimension, and compute the average entropies of the children. And we see that, conveniently, all the points with L not greater than 1.5 are of class 0, so we can make a leaf there.

Bankruptcy Example Now, we consider all the splits of the remaining part of space. Note that we have to recalculate all the average entropies again, because the points that fall into the leaf node are taken out of consideration.

Bankruptcy Example Now the best split is at R > 0.9. And we see that all the points for which that's true are positive, so we can make another leaf.

Bankruptcy Example Continuing in this way, we finally obtain: