INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.

Slides:



Advertisements
Similar presentations
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Advertisements

Cenni di Machine Learning: Decision Tree Learning Fabio Massimo Zanzotto.
Learning from Observations
Learning from Observations Chapter 18 Section 1 – 3.
ICS 178 Intro Machine Learning
DECISION TREES. Decision trees  One possible representation for hypotheses.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
CS 4700: Foundations of Artificial Intelligence
Learning Department of Computer Science & Engineering Indian Institute of Technology Kharagpur.
Decision making in episodic environments
Decision Tree Rong Jin. Determine Milage Per Gallon.
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Fall 2005.
Cooperating Intelligent Systems
Learning From Observations
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18.
ICS 273A Intro Machine Learning
CSCE 580 Artificial Intelligence Ch.18: Learning from Observations
Three kinds of learning
LEARNING DECISION TREES
Learning decision trees derived from Hwee Tou Ng, slides for Russell & Norvig, AI a Modern Approachslides Tom Carter, “An introduction to information theory.
Learning decision trees
Learning decision trees derived from Hwee Tou Ng, slides for Russell & Norvig, AI a Modern Approachslides Tom Carter, “An introduction to information theory.
Learning….in a rather broad sense: improvement of performance on the basis of experience Machine learning…… improve for task T with respect to performance.
ICS 273A Intro Machine Learning
Learning: Introduction and Overview
Part I: Classification and Bayesian Learning
Induction of Decision Trees (IDT) CSE 335/435 Resources: – –
Machine learning Image source:
Machine learning Image source:
Introduction to Machine Learning Reading for today: R&N Next lecture: R&N ,
Machine Learning CPS4801. Research Day Keynote Speaker o Tuesday 9:30-11:00 STEM Lecture Hall (2 nd floor) o Meet-and-Greet 11:30 STEM 512 Faculty Presentation.
Inductive learning Simplest form: learn a function from examples
Machine Learning CSE 681 CH2 - Supervised Learning.
LEARNING DECISION TREES Yılmaz KILIÇASLAN. Definition - I Decision tree induction is one of the simplest, and yet most successful forms of learning algorithm.
Data Mining: Classification & Predication Hosam Al-Samarraie, PhD. Centre for Instructional Technology & Multimedia Universiti Sains Malaysia.
Learning from observations
Learning from Observations Chapter 18 Through
CHAPTER 18 SECTION 1 – 3 Learning from Observations.
Lecture 7 : Intro to Machine Learning Rachel Greenstadt & Mike Brennan November 10, 2008.
Learning from Observations Chapter 18 Section 1 – 3, 5-8 (presentation TBC)
Learning from Observations Chapter 18 Section 1 – 3.
Homework Submission deadline: next class(May 28st) Handwritten report Please answer the following question on probability. –Suppose one is interested in.
L6. Learning Systems in Java. Necessity of Learning No Prior Knowledge about all of the situations. Being able to adapt to changes in the environment.
Decision Trees. What is a decision tree? Input = assignment of values for given attributes –Discrete (often Boolean) or continuous Output = predicated.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
DECISION TREE Ge Song. Introduction ■ Decision Tree: is a supervised learning algorithm used for classification or regression. ■ Decision Tree Graph:
Machine learning Image source:
CSC 8520 Spring Paula Matuszek DecisionTreeFirstDraft Paula Matuszek Spring,
Chapter 18 Section 1 – 3 Learning from Observations.
Decision Tree. Classification Databases are rich with hidden information that can be used for making intelligent decisions. Classification is a form of.
Learning From Observations Inductive Learning Decision Trees Ensembles.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Anifuddin Azis LEARNING. Why is learning important? So far we have assumed we know how the world works Rules of queens puzzle Rules of chess Knowledge.
Decision Tree Learning CMPT 463. Reminders Homework 7 is due on Tuesday, May 10 Projects are due on Tuesday, May 10 o Moodle submission: readme.doc and.
Learning from Observations
Learning from Observations
Learning from Observations
Machine learning Image source:
Machine Learning Inductive Learning and Decision Trees
Announcements (1) Background reading for next week is posted.
Introduction to Machine Learning
Introduce to machine learning
Presented By S.Yamuna AP/CSE
Data Mining Lecture 11.
Decision making in episodic environments
Learning from Observations
Learning from Observations
Decision trees One possible representation for hypotheses
Machine Learning: Decision Tree Learning
Decision Trees - Intermediate
Presentation transcript:

INTRODUCTION TO MACHINE LEARNING

$1,000,000

Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised learning  Reinforcement learning

Variants of Machine Learning Problems  What is being learned ?  Parameters, problem structure, hidden concepts, …  What information do we learn from ?  Labeled data, unlabeled data, rewards  What is the goal of learning ?  Prediction, diagnostics, summarization, …  How do we learn ?  Passive / active, online / offline  Outputs  Binary, discrete, continuous

Supervised Learning  Given a set of data points with an outcome, create a model to describe them  Classification  outcome is a discrete variable ( typically <10 outcomes )  Linear regression  outcome is continuous

Training Data Training set. Data set. Training data. Observations. InputOutput

What else do we have ?  In real life, we usually don ’ t have just a set of data  Also have background knowledge, theory about the underlying processes, etc.  We will assume just the data ( this is called inductive learning )  Cleaner and a good base case  More complex mechanisms needed to reason with prior knowledge

Inductive Learning

How can we pick the right function ?  There could be multiple models to generate the data  Examples do not completely describe the function  Search space is large  Would we know if we got the right answer ? Not the right way of thinking about the problem.

Inductive Learning

Avoiding Overfitting the Model 1. Divide the data that you have into a distinct training set and test set. 2. Use only the training set to train your model. 3. Verify performance using the test set. Measure error rate  Drawback of this method : the data withheld for the test set is not used for training  split of data means we didn ’ t train on half the data  split means we might not get a good idea of the accuracy

K - fold Cross - Validation

More Data Is Usually Better

Classification Algorithms  Decision trees  Neural networks  Logistic regression  Naïve Bayes  …

Decision Trees  …similar to a game of 20 questions  Decision trees are powerful and popular tools for classification and prediction.  Decision trees represent rules, which can be understood by humans and used in knowledge system such as database.

Learning decision trees Problem : decide whether to wait for a table at a restaurant, based on the following attributes : 1. Alternate : is there an alternative restaurant nearby ? 2. Bar : is there a comfortable bar area to wait in ? 3. Fri / Sat : is today Friday or Saturday ? 4. Hungry : are we hungry ? 5. Patrons : number of people in the restaurant ( None, Some, Full ) 6. Price : price range ($, $$, $$$) 7. Raining : is it raining outside ? 8. Reservation : have we made a reservation ? 9. Type : kind of restaurant ( French, Italian, Thai, Burger ) 10. WaitEstimate : estimated waiting time (0-10, 10-30, , >60)

Attribute - based representations  Examples described by attribute values ( Boolean, discrete, continuous )  E. g., situations where I will / won ' t wait for a table :  Classification of examples is positive ( T ) or negative ( F )

Decision trees  One possible representation for hypotheses

Expressiveness  Decision trees can express any function of the input attributes.  E. g., for Boolean functions, truth table row → path to leaf :  Trivially, there is a consistent decision tree for any training set with one path to leaf for each example ( unless f nondeterministic in x ) but it probably won ' t generalize to new examples  Prefer to find more compact decision trees

Decision tree learning  Aim : find a small tree consistent with the training examples  Idea : ( recursively ) choose " most significant " attribute as root of ( sub ) tree

Decision trees  One possible representation for hypotheses

Choosing an attribute  Idea : a good attribute splits the examples into subsets that are ( ideally ) " all positive " or " all negative "  Which is a better choice ?  Patrons