Machine Learning II Decision Tree Induction CSE 473.

Slides:



Advertisements
Similar presentations
1 Machine Learning: Lecture 3 Decision Tree Learning (Based on Chapter 3 of Mitchell T.., Machine Learning, 1997)
Advertisements

Decision Trees Decision tree representation ID3 learning algorithm
Machine Learning III Decision Tree Induction
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
1er. Escuela Red ProTIC - Tandil, de Abril, Decision Tree Learning 3.1 Introduction –Method for approximation of discrete-valued target functions.
Classification Algorithms
ICS320-Foundations of Adaptive and Learning Systems
Classification Techniques: Decision Tree Learning
Università di Milano-Bicocca Laurea Magistrale in Informatica
Jesse Davis Machine Learning Jesse Davis
Decision Trees. DEFINE: Set X of Instances (of n-tuples x = ) –E.g., days decribed by attributes (or features): Sky, Temp, Humidity, Wind, Water, Forecast.
Part 7.3 Decision Trees Decision tree representation ID3 learning algorithm Entropy, information gain Overfitting.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 4: ID3.
Decision Tree Learning Learning Decision Trees (Mitchell 1997, Russell & Norvig 2003) –Decision tree induction is a simple but powerful learning paradigm.
Machine Learning CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanning.
Machine Learning: Symbol-Based
Machine Learning Lecture 10 Decision Trees G53MLE Machine Learning Dr Guoping Qiu1.
Decision Tree Learning
By Wang Rui State Key Lab of CAD&CG
Machine Learning Chapter 3. Decision Tree Learning
Artificial Intelligence 7. Decision trees
Mohammad Ali Keyvanrad
Decision tree learning Maria Simi, 2010/2011 Inductive inference with decision trees  Decision Trees is one of the most widely used and practical methods.
Machine Learning Lecture 10 Decision Tree Learning 1.
CpSc 810: Machine Learning Decision Tree Learning.
Learning from Observations Chapter 18 Through
Decision-Tree Induction & Decision-Rule Induction
For Wednesday No reading Homework: –Chapter 18, exercise 6.
CpSc 810: Machine Learning Concept Learning and General to Specific Ordering.
Outline Inductive bias General-to specific ordering of hypotheses
For Monday No new reading Homework: –Chapter 18, exercises 3 and 4.
CS 8751 ML & KDDDecision Trees1 Decision tree representation ID3 learning algorithm Entropy, Information gain Overfitting.
Machine Learning II Decision Tree Induction CSE 573.
CS 5751 Machine Learning Chapter 3 Decision Tree Learning1 Decision Trees Decision tree representation ID3 learning algorithm Entropy, Information gain.
1 Decision Tree Learning Original slides by Raymond J. Mooney University of Texas at Austin.
1 Decision Trees ExampleSkyAirTempHumidityWindWaterForecastEnjoySport 1SunnyWarmNormalStrongWarmSameYes 2SunnyWarmHighStrongWarmSameYes 3RainyColdHighStrongWarmChangeNo.
Decision Tree Learning
Decision Tree Learning Presented by Ping Zhang Nov. 26th, 2007.
Outline Logistics Review Machine Learning –Induction of Decision Trees (7.2) –Version Spaces & Candidate Elimination –PAC Learning Theory (7.1) –Ensembles.
Machine Learning Recitation 8 Oct 21, 2009 Oznur Tastan.
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
Friday’s Deliverable As a GROUP, you need to bring 2N+1 copies of your “initial submission” –This paper should be a complete version of your paper – something.
Learning From Observations Inductive Learning Decision Trees Ensembles.
Machine Learning & Datamining CSE 454. © Daniel S. Weld 2 Project Part 1 Feedback Serialization Java Supplied vs. Manual.
Review of Decision Tree Learning Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
CSE573 Autumn /09/98 Machine Learning Administrative –Last topic: Decision Tree Learning Reading: 5.1, 5.4 Last time –finished NLP sample system’s.
CSE573 Autumn /11/98 Machine Learning Administrative –Finish this topic –The rest of the time is yours –Final exam Tuesday, Mar. 17, 2:30-4:20.
Machine Learning Inductive Learning and Decision Trees
Università di Milano-Bicocca Laurea Magistrale in Informatica
CS 9633 Machine Learning Decision Tree Learning
Decision Tree Learning
Decision trees (concept learnig)
Machine Learning Lecture 2: Decision Tree Learning.
Introduce to machine learning
Decision trees (concept learnig)
Decision Tree Learning
CSE543: Machine Learning Lecture 2: August 6, 2014
Knowledge Representation
Introduction to Machine Learning Algorithms in Bioinformatics: Part II
Decision Tree Saed Sayad 9/21/2018.
Machine Learning Chapter 3. Decision Tree Learning
Machine Learning: Lecture 3
Decision Trees Decision tree representation ID3 learning algorithm
Machine Learning Chapter 3. Decision Tree Learning
Why Machine Learning Flood of data
Decision Trees Decision tree representation ID3 learning algorithm
Machine Learning Chapter 2
A task of induction to find patterns
A task of induction to find patterns
Machine Learning Chapter 2
Presentation transcript:

Machine Learning II Decision Tree Induction CSE 473

© Daniel S. Weld 2 Logistics PS 2 Due Review Session Tonight Exam on Wed Closed Book Questions like non-programming part of PSets PS 3 out in a Week

© Daniel S. Weld 3 Machine Learning Outline Machine learning: √Function approximation √Bias Supervised learning Classifiers A supervised learning technique in depth Induction of Decision Trees Ensembles of classifiers Overfitting

© Daniel S. Weld 4 Defined by restriction bias Terminology

© Daniel S. Weld 5 Past Insight Any ML problem may be cast as the problem of FUNCTION APPROXIMATION

© Daniel S. Weld 6 Types of Training Experience Credit assignment problem: Direct training examples: E.g. individual checker boards + correct move for each Supervised learning Indirect training examples : E.g. complete sequence of moves and final result Reinforcement learning Which examples: Random, teacher chooses, learner chooses

© Daniel S. Weld 7 Supervised Learning (aka Classification) FUNCTION APPROXIMATION The most simple case has experience: …

© Daniel S. Weld 8 Issues: Learning Algorithms

© Daniel S. Weld 9 Insight 2: Bias Learning occurs when PREJUDICE meets DATA! The nice word for prejudice is “bias”. What kind of hypotheses will you consider? What is allowable range of functions you use when approximating? What kind of hypotheses do you prefer?

© Daniel S. Weld 10 Two Strategies for ML Restriction bias: use prior knowledge to specify a restricted hypothesis space. Version space algorithm over conjunctions. Preference bias: use a broad hypothesis space, but impose an ordering on the hypotheses. Decision trees.

© Daniel S. Weld 11 Example: “Good day for sports” Attributes of instances Wind Temperature Humidity Outlook Feature = attribute with one value E.g. outlook = sunny Sample instance wind=weak, temp=hot, humidity=high, outlook=sunny

© Daniel S. Weld 12 Experience: “Good day for sports” Day OutlookTempHumidWindPlayTennis? d1shhwn d2shhsn d3ohhwy d4rmhw y d5rcnwy d6rcnsy d7ocnsy d8smhwn d9scnwy d10rmnwy d11smnsy d12omhsy d13ohnwy d14rmhsn

© Daniel S. Weld 13 Restricted Hypothesis Space conjunction

© Daniel S. Weld 14 Resulting Learning Problem

© Daniel S. Weld 15 Consistency Say an “example is consistent with a hypothesis” when they agree Hypothesis Sky Temp Humidity, Wind Examples:

© Daniel S. Weld 16 Naïve Algorithm

© Daniel S. Weld 17 More-General-Than Order {?, cold, high, ?, ?} >> {sun, cold, high, ? low}

© Daniel S. Weld 18 Ordering on Hypothesis Space

© Daniel S. Weld 19 Search Space Nodes? Operators?

© Daniel S. Weld 20 Two Strategies for ML Restriction bias: use prior knowledge to specify a restricted hypothesis space. Version space algorithm over conjunctions. Preference bias: use a broad hypothesis space, but impose an ordering on the hypotheses. Decision trees.

© Daniel S. Weld 21 Decision Trees Convenient Representation Developed with learning in mind Deterministic Expressive Equivalent to propositional DNF Handles discrete and continuous parameters Simple learning algorithm Handles noise well Classify as follows Constructive (build DT by adding nodes) Eager Batch (but incremental versions exist)

© Daniel S. Weld 22 Classification E.g. Learn concept “Edible mushroom” Target Function has two values: T or F Represent concepts as decision trees Use hill climbing search Thru space of decision trees Start with simple concept Refine it into a complex concept as needed

© Daniel S. Weld 23 Experience: “Good day for tennis” Day OutlookTempHumidWindPlayTennis? d1shhwn d2shhsn d3ohhwy d4rmhw y d5rcnwy d6rcnsy d7ocnsy d8smhwn d9scnwy d10rmnwy d11smnsy d12omhsy d13ohnwy d14rmhsn

© Daniel S. Weld 24 Decision Tree Representation Outlook Humidity Wind Yes No Sunny Overcast Rain High Strong Normal Weak Good day for tennis? Leaves = classification Arcs = choice of value for parent attribute Decision tree is equivalent to logic in disjunctive normal form G-Day  (Sunny  Normal)  Overcast  (Rain  Weak)

© Daniel S. Weld 25

© Daniel S. Weld 26

© Daniel S. Weld 27 DT Learning as Search Nodes Operators Initial node Heuristic? Goal? Decision Trees Tree Refinement: Sprouting the tree Smallest tree possible: a single leaf Information Gain Best tree possible (???)

© Daniel S. Weld 28 What is the Simplest Tree? Day OutlookTempHumidWindPlay? d1shhwn d2shhsn d3ohhwy d4rmhw y d5rcnwy d6rcnsy d7ocnsy d8smhwn d9scnwy d10rmnwy d11smnsy d12omhsy d13ohnwy d14rmhsn How good? [10+, 4-] Means: correct on 10 examples incorrect on 4 examples

© Daniel S. Weld 29 Successors Yes Outlook Temp Humid Wind Which attribute should we use to split?

© Daniel S. Weld 30 To be decided: How to choose best attribute? Information gain Entropy (disorder) When to stop growing tree?