Decision Trees By Cole Daily CSCI 446.

Slides:



Advertisements
Similar presentations
DECISION TREES. Decision trees  One possible representation for hypotheses.
Advertisements

CHAPTER 9: Decision Trees
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Decision Tree Approach in Data Mining
Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Part I Introduction to Data Mining by Tan,
CART: Classification and Regression Trees Chris Franck LISA Short Course March 26, 2013.
Decision Tree.
A Quick Overview By Munir Winkel. What do you know about: 1) decision trees 2) random forests? How could they be used?
Chapter 7 – Classification and Regression Trees
CMPUT 466/551 Principal Source: CMU
Chapter 7 – Classification and Regression Trees
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
ID3 Algorithm Abbas Rizvi CS157 B Spring What is the ID3 algorithm? ID3 stands for Iterative Dichotomiser 3 Algorithm used to generate a decision.
CSci 8980: Data Mining (Fall 2002)
Decision Tree Algorithm
Ensemble Learning: An Introduction
© Vipin Kumar CSci 8980 Fall CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance Computing Research Center Department of Computer.
Lecture 5 (Classification with Decision Trees)
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Example of a Decision Tree categorical continuous class Splitting Attributes Refund Yes No NO MarSt Single, Divorced Married TaxInc NO < 80K > 80K.
ICS 273A Intro Machine Learning
Ensemble Learning (2), Tree and Forest
Microsoft Enterprise Consortium Data Mining Concepts Introduction to Directed Data Mining: Decision Trees Prepared by David Douglas, University of ArkansasHosted.
Introduction to Directed Data Mining: Decision Trees
Learning what questions to ask. 8/29/03Decision Trees2  Job is to build a tree that represents a series of questions that the classifier will ask of.
1 Data Mining Lecture 3: Decision Trees. 2 Classification: Definition l Given a collection of records (training set ) –Each record contains a set of attributes,
Chapter 9 – Classification and Regression Trees
Lecture 7. Outline 1. Overview of Classification and Decision Tree 2. Algorithm to build Decision Tree 3. Formula to measure information 4. Weka, data.
Computational Intelligence: Methods and Applications Lecture 19 Pruning of decision trees Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Machine Learning Queens College Lecture 2: Decision Trees.
Decision Tree Learning Debapriyo Majumdar Data Mining – Fall 2014 Indian Statistical Institute Kolkata August 25, 2014.
Scaling up Decision Trees. Decision tree learning.
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
For Wednesday No reading Homework: –Chapter 18, exercise 6.
CS690L Data Mining: Classification
1 Decision Tree Learning Original slides by Raymond J. Mooney University of Texas at Austin.
Decision Trees Example of a Decision Tree categorical continuous class Refund MarSt TaxInc YES NO YesNo Married Single, Divorced < 80K> 80K Splitting.
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Lecture Notes for Chapter 4 Introduction to Data Mining
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Classification and Regression Trees
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
Machine Learning Recitation 8 Oct 21, 2009 Oznur Tastan.
Eco 6380 Predictive Analytics For Economists Spring 2016 Professor Tom Fomby Department of Economics SMU.
Ensemble Learning, Boosting, and Bagging: Scaling up Decision Trees (with thanks to William Cohen of CMU, Michael Malohlava of 0xdata, and Manish Amde.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
BY International School of Engineering {We Are Applied Engineering} Disclaimer: Some of the Images and content have been taken from multiple online sources.
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
University of Waikato, New Zealand
Data Mining Practical Machine Learning Tools and Techniques
DECISION TREES An internal node represents a test on an attribute.
Bagging and Random Forests
Introduction to Machine Learning and Tree Based Methods
C4.5 - pruning decision trees
Chapter 13 – Ensembles and Uplift
Trees, bagging, boosting, and stacking
Ch9: Decision Trees 9.1 Introduction A decision tree:
Basic machine learning background with Python scikit-learn
Decision Tree Saed Sayad 9/21/2018.
Data Mining Classification: Basic Concepts and Techniques
Statistical Learning Dong Liu Dept. EEIS, USTC.
Ensemble learning.
Ensemble learning Reminder - Bagging of Trees Random Forest
Classification with CART
INTRODUCTION TO Machine Learning
INTRODUCTION TO Machine Learning 2nd Edition
Decision Trees Jeff Storey.
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
STT : Intro. to Statistical Learning
Presentation transcript:

Decision Trees By Cole Daily CSCI 446

Overview What are decision trees Visualizing a tree Types of trees Building a tree Pruning a tree Advantages and Disadvantages

What are Decision Trees? Predictive model Use previous experience Find the value of a target variable Efficient after built What are Decision Trees?

Visualizing a tree Types of nodes Root Decision Leaf

Types of Trees CART Decision stream Ensembles Classification Regression Decision stream Ensembles Boosted Trees Bootstrap Aggregate Rotation Forest

Building a Decision Tree Recent  Relevant  Time  Read  y  n 

Entropy “Measure of information in a single random variable” Ranges from 0 to 1 0 – No information possible 1 – Maximum information possible Maxes when possibilities are equally likely

Information Gain Measure of information gained from a split Sum the entropy of each possible value Subtract from the entropy before Good for Classification trees

General Method for Building a Tree Determine best split using some method Split Repeat until: Leaves are pure No splits possible Splitting is not useful Prune

The example Get information gain of each variable Recent = (- ⅗log2⅗ - ⅖log2⅖) - (((- ⅔log2 ⅔ – ⅓ log2 ⅓) + (-  ½log2 ½ - ½log2 ½)) / 2)  = .014  Recent doesn’t purify the leaf nodes Best split is Relevant Relevant true – read article Relevant false – 20% read article Recent  Relevant  Time  Read  y  n 

The Gini Coefficient Developed by statistician Corrado Gini Used to determine wealth distribution over a population Relative mean absolute difference 0 – Maximum equality 1 – Maximum inequality

Variance Reduction Used for continuous target variables Reduce the variance in possible values of each leaf node Good for regression trees Similar to Gini Coefficient

Pruning Stopping conditions: Overfitted to training data No more splits Splits are not useful Leaves are pure Overfitted to training data Inaccurate in practice Pruning

Pruning General pruning method: Reduced-Error Pruning Replace subtree with leaf node Check accuracy of tree If improved, keep leaf node If not, keep subtree General pruning method: Take most popular outcome in a subtree and replace subtree with it Fastest and easiest pruning method Doesn’t guarantee an optimal tree Reduced-Error Pruning Pruning

Cost-Complexity Pruning Generate a series of trees starting Consider a subtree and replace with a leaf node Subtree used decided by rate of error/number of leaves The subtree that minimizes this is used Use the new subtree as the starting tree and repeat End when only the root node is left Check accuracy of all trees

Advantages and Disadvantages Easy to visualize Simple concept Disadvantages Complex to build Lots of math Used mostly in data mining and artificial intelligence Can be applied any field that makes decisions Advantages and Disadvantages

Conclusion What are decision trees Visualizing Types Building Pruning Advantages and Disadvantages

Sources Brownlee, J. (2017, September 20). Classification And Regression Trees for Machine Learning. Retrieved November 20, 2018, from https://machinelearningmastery.com/classification-and-regression-trees-for- machine-learning/  Decision Tree AnalysisChoosing by Projecting "Expected Outcomes". (n.d.). Retrieved November 20, 2018, from https://www.mindtools.com/dectree.html  Decision tree implementation using Python. (2018, February 11). Retrieved November 20, 2018, from https://www.geeksforgeeks.org/decision-tree- implementation-python/  Decision tree learning. (2018, November 20). Retrieved November 20, 2018, from https://en.wikipedia.org/wiki/Decision_tree_learning  Decision tree pruning. (2018, November 16). Retrieved November 20, 2018, from https://en.wikipedia.org/wiki/Decision_tree_pruning  Elomma, T., & Kaariainen, M. (2001, January 09). Retrieved November 20, 2018, from https://www.jair.org/index.php/jair/article/view/10284/24526  Gini coefficient. (2018, November 12). Retrieved November 20, 2018, from https://en.wikipedia.org/wiki/Gini_coefficient  Variance reduction. (2018, March 21). Retrieved November 20, 2018, from https://en.wikipedia.org/wiki/Variance_reduction  Zhao, L., Nagamochi, H., & Ibaraki, T. (2005, January). Greedy Splitting Algorithms for Approximating Multiway Partition Problems. Retrieved November 20, 2018, from https://www.researchgate.net/profile/Liang_Zhao33/publication/220589424_ Greedy_splitting_algorithms_for_approximating_multiway_partition_problems/ links/004635267b9e5a3ed1000000/Greedy-splitting-algorithms-for- approximating-multiway-partition-problems.pdf?origin=publication_detail