A task of induction to find patterns

Slides:



Advertisements
Similar presentations
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
Advertisements

1 Machine Learning: Lecture 3 Decision Tree Learning (Based on Chapter 3 of Mitchell T.., Machine Learning, 1997)
Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Part I Introduction to Data Mining by Tan,
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Intelligent Environments1 Computer Science and Engineering University of Texas at Arlington.
Classification Techniques: Decision Tree Learning
Spring 2003Data Mining by H. Liu, ASU1 3. Classification Methods Patterns and Models Regression, NBC k-Nearest Neighbors Decision Trees and Rules Large.
CII504 Intelligent Engine © 2005 Irfan Subakti Department of Informatics Institute Technology of Sepuluh Nopember Surabaya - Indonesia.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Machine Learning II Decision Tree Induction CSE 473.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
CES 514 – Data Mining Lecture 8 classification (contd…)
Induction of Decision Trees
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Classification and Prediction by Yen-Hsien Lee Department of Information Management College of Management National Sun Yat-Sen University March 4, 2003.
Machine Learning Lecture 10 Decision Trees G53MLE Machine Learning Dr Guoping Qiu1.
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Machine Learning Lecture 11 Summary G53MLE | Machine Learning | Dr Guoping Qiu1.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
9/03Data Mining – Classification G Dong 1 3. Classification Methods Patterns and Models Regression, NBC k-Nearest Neighbors Decision Trees and Rules Large.
CS690L Data Mining: Classification
Non-Bayes classifiers. Linear discriminants, neural networks.
Data Mining and Decision Support
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
DATA MINING TECHNIQUES (DECISION TREES ) Presented by: Shweta Ghate MIT College OF Engineering.
CSE573 Autumn /09/98 Machine Learning Administrative –Last topic: Decision Tree Learning Reading: 5.1, 5.4 Last time –finished NLP sample system’s.
CSE573 Autumn /11/98 Machine Learning Administrative –Finish this topic –The rest of the time is yours –Final exam Tuesday, Mar. 17, 2:30-4:20.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
CS 9633 Machine Learning Support Vector Machines
Machine Learning Inductive Learning and Decision Trees
DECISION TREES An internal node represents a test on an attribute.
Learning with Perceptrons and Neural Networks
Classification Algorithms
CSE543: Machine Learning Lecture 2: August 6, 2014
School of Computer Science & Engineering
CS 9633 Machine Learning Concept Learning
Prepared by: Mahmoud Rafeek Al-Farra
Data Science Algorithms: The Basic Methods
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Supervised Learning Seminar Social Media Mining University UC3M
Cost-Sensitive Learning
Data Mining Lecture 11.
A task of induction to find patterns
Classification Techniques: Bayesian Classification
Cost-Sensitive Learning
Machine Learning: Lecture 3
Machine Learning Ensemble Learning: Voting, Boosting(Adaboost)
Naïve Bayes Classifier
network of simple neuron-like computing elements
COSC 4335: Other Classification Techniques
Artificial Intelligence Lecture No. 28
Lecture Notes for Chapter 4 Artificial Neural Networks
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
A task of induction to find patterns
Artificial Intelligence 9. Perceptron
COSC 4335: Part2: Other Classification Techniques
Machine Learning Chapter 2
A task of induction to find patterns
Using Bayesian Network in the Construction of a Bi-level Multi-classifier. A Case Study Using Intensive Care Unit Patients Data B. Sierra, N. Serrano,
David Kauchak CS158 – Spring 2019
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Machine Learning Chapter 2
Data Mining CSCI 307, Spring 2019 Lecture 6
Presentation transcript:

A task of induction to find patterns Classification A task of induction to find patterns 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu Outline Data and its format Problem of Classification Learning a classifier Different approaches Key issues 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu Data and its format Data attribute-value pairs with/without class Data type continuous/discrete nominal Data format flat 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu Sample data 6/25/2019 CSE 591: Data Mining by H. Liu

Induction from databases Inferring knowledge from data The task of deduction infer information that is a logical consequence of querying a database Who conducted this class before? Which courses are attended by Mary? Deductive databases: extending the RDBMS RDBMS - relational database management systems RDBMS offer simple operators for the deduction of information, such as join 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu Classification It is one type of induction data with class labels Examples - If weather is rainy then no golf If Induction is different from deduction and DBMS does not not support induction; The result of induction is higher-level information or knowledge: general statements about data There are many approaches. Refer to the lecture notes for CS3244 available at the Co-Op. We focus on three approaches here, other examples: Other approaches Instance-based learning other neural networks Concept learning (Version space, Focus, Aq11, …) Genetic algorithms Reinforcement learning 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu Different approaches There exist many techniques Decision trees Neural networks K-nearest neighbors Naïve Bayesian classifiers Support Vector Machines Ensemble methods Co-training and many more ... 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu A decision tree Outlook Humidity Wind sunny overcast rain YES high normal strong weak NO Issues How to build such a tree from the data? What are the criteria for performance measurement correctness conciseness What are the key components? test stopping criterion 6/25/2019 CSE 591: Data Mining by H. Liu

Inducing a decision tree There are many possible trees let’s try it on the golfing data How to find the most compact one that is consistent with the data? Why the most compact? Occam’s razor principle Issue of efficiency w.r.t. optimality 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu Information gain and Entropy - Information gain - the difference between the node before and after splitting 6/25/2019 CSE 591: Data Mining by H. Liu

Building a compact tree The key to building a decision tree - which attribute to choose in order to branch. The heuristic is to choose the attribute with the maximum IG. Another explanation is to reduce uncertainty as much as possible. 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu Learn a decision tree Outlook sunny overcast rain Humidity Wind YES high normal strong weak NO YES NO YES 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu K-Nearest Neighbor One of the most intuitive classification algorithm An unseen instance’s class is determined by its nearest neighbor The problem is it is sensitive to noise Instead of using one neighbor, we can use k neighbors 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu K-NN New problems lazy learning large storage An example How good is k-NN? 6/25/2019 CSE 591: Data Mining by H. Liu

Naïve Bayes Classifier This is a direct application of Bayes’ rule P(C|X) = P(X|C)P(C)/P(X) X - a vector of x1,x2,…,xn That’s the best classifier you can build But, there are problems 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu NBC (2) Assume conditional independence between xi’s We have An example How good is it in reality? 6/25/2019 CSE 591: Data Mining by H. Liu

Classification via Neural Networks Squash  A perceptron 6/25/2019 CSE 591: Data Mining by H. Liu

What can a perceptron do? Neuron as a computing device To separate a linearly separable points Nice things about a perceptron distributed representation local learning weight adjusting 6/25/2019 CSE 591: Data Mining by H. Liu

CSE 591: Data Mining by H. Liu Linear threshold unit Basic concepts: projection, thresholding W vectors evoke 1 W = [.11 .6] L= [.7 .7] .5 6/25/2019 CSE 591: Data Mining by H. Liu

E.g. 1: solution region for AND problem Find a weight vector that satisfies all the constraints AND problem 0 0 0 0 1 0 1 0 0 1 1 1 6/25/2019 CSE 591: Data Mining by H. Liu

E.g. 2: Solution region for XOR problem? 0 0 0 0 1 1 1 0 1 1 1 0 6/25/2019 CSE 591: Data Mining by H. Liu

Learning by error reduction Perceptron learning algorithm If the activation level of the output unit is 1 when it should be 0, reduce the weight on the link to the ith input unit by r*Li, where Li is the ith input value and r a learning rate If the activation level of the output unit is 0 when it should be 1, increase the weight on the link to the ith input unit by r*Li Otherwise, do nothing 6/25/2019 CSE 591: Data Mining by H. Liu

Multi-layer perceptrons Using the chain rule, we can back-propagate the errors for a multi-layer perceptrons. Output layer Hidden layer Differences between DT and NN Speed Accuracy Comprehensibility Which one to use Many successful applications of both approaches Input layer 6/25/2019 CSE 591: Data Mining by H. Liu