These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.

Slides:



Advertisements
Similar presentations
1 Classification using instance-based learning. 3 March, 2000Advanced Knowledge Management2 Introduction (lazy vs. eager learning) Notion of similarity.
Advertisements

Machine Learning Instance Based Learning & Case Based Reasoning Exercise Solutions.
Data Mining Classification: Alternative Techniques
K-means method for Signal Compression: Vector Quantization
Preventing Overfitting Problem: We don’t want to these algorithms to fit to ``noise’’ The generated tree may overfit the training data –Too many branches,
Instance Based Learning
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
Kansas State University Department of Computing and Information Sciences Laboratory for Knowledge Discovery in Databases (KDD) KDD Group Research Seminar.
1er. Escuela Red ProTIC - Tandil, de Abril, Instance-Based Learning 4.1 Introduction Instance-Based Learning: Local approximation to the.
Lecture 6 K-Nearest Neighbor Classifier
. Markov Chains as a Learning Tool. 2 Weather: raining today40% rain tomorrow 60% no rain tomorrow not raining today20% rain tomorrow 80% no rain tomorrow.
Classification and Decision Boundaries
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
Instance Based Learning
K nearest neighbor and Rocchio algorithm
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
Bayesian classifiers.
Instance based learning K-Nearest Neighbor Locally weighted regression Radial basis functions.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Nearest Neighbor Models (Reading: Chapter.
Instance Based Learning
1 Classification: Definition Given a collection of records (training set ) Each record contains a set of attributes, one of the attributes is the class.
Instance-Based Learning
Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
1 Nearest Neighbor Learning Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Intro AI.
Aprendizagem baseada em instâncias (K vizinhos mais próximos)
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Instance Based Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán) 1.
Nearest Neighbour Condensing and Editing David Claus February 27, 2004 Computer Vision Reading Group Oxford.
INSTANCE-BASE LEARNING
CS Instance Based Learning1 Instance Based Learning.
Module 04: Algorithms Topic 07: Instance-Based Learning
Supervised Learning and k Nearest Neighbors Business Intelligence for Managers.
Rainbow Tool Kit Matt Perry Global Information Systems Spring 2003.
GEOMETRIC VIEW OF DATA David Kauchak CS 451 – Fall 2013.
K Nearest Neighborhood (KNNs)
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
11/12/2012ISC471 / HCI571 Isabelle Bichindaritz 1 Prediction.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
1 Instance Based Learning Ata Kaban The University of Birmingham.
David Claus and Christoph F. Eick: Nearest Neighbor Editing and Condensing Techniques Nearest Neighbor Editing and Condensing Techniques 1.Nearest Neighbor.
CpSc 881: Machine Learning Instance Based Learning.
CpSc 810: Machine Learning Instance Based Learning.
Outline K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Lazy Learners K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
KNN Classifier.  Handed an instance you wish to classify  Look around the nearby region to see what other classes are around  Whichever is most common—make.
K nearest neighbors algorithm Parallelization on Cuda PROF. VELJKO MILUTINOVIĆ MAŠA KNEŽEVIĆ 3037/2015.
Meta-learning for Algorithm Recommendation Meta-learning for Algorithm Recommendation Background on Local Learning Background on Algorithm Assessment Algorithm.
CS Machine Learning Instance Based Learning (Adapted from various sources)
K-Nearest Neighbor Learning.
Classification Algorithms Covering, Nearest-Neighbour.
1 Learning Bias & Clustering Louis Oliphant CS based on slides by Burr H. Settles.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
1 Instance Based Learning Soongsil University Intelligent Systems Lab.
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Data Science Algorithms: The Basic Methods
Data Mining: Concepts and Techniques (3rd ed
Instance Based Learning (Adapted from various sources)
K Nearest Neighbor Classification
Nearest-Neighbor Classifiers
Classification Algorithms
Chap 8. Instance Based Learning
Advanced Mathematics Hossein Malekinezhad.
Machine Learning: UNIT-4 CHAPTER-1
Nearest Neighbors CSC 576: Data Mining.
Nearest Neighbor Classifiers
Presentation transcript:

These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified Pros: –Classification hypothesis is developed locally for each instance to be classified Cons: –Running time (no model is built, so each classification actually builds a local model from scratch)

These slides are based on Tom Mitchell’s book “Machine Learning” K-Nearest Neighbors Classification of new instances is based on classifications of (one or more) known instances nearest to them –K=1  1-NN (using a single nearest neighbor) –Frequently, K > 1 Assumption: all instances correspond to points in the n-dimensional space R n –Dimensions = features (aka attributes)

These slides are based on Tom Mitchell’s book “Machine Learning” Metrics Nearest neighbors are identified using a metric defined for this high-dimensional space Let x be an arbitrary instance with feature vector Euclidean metric is frequently used for real-valued features:

These slides are based on Tom Mitchell’s book “Machine Learning” Pseudo-code for KNN Training algorithm –For each training example, add the example to the list Training Classification algorithm (R n  V) –Let V ={v 1, …, v l } be a set of classes –Given a query instance x q to be classified Let X={x 1, …, x k } denote the k instances from Training that are nearest to x q Return v i such that |vote i | is largest

These slides are based on Tom Mitchell’s book “Machine Learning” Distance-weighted KNN Weighting contribution of each of the k neighbors according to their distance to the query point x q –Give greater weight to closer neighbors –Return v i such that |w i | is largest

These slides are based on Tom Mitchell’s book “Machine Learning” Distance-weighted KNN (cont’d) If x q exactly matches one of the training instances x i, and d(x q, x i )=0, then we simply take class(x i ) to be the classification of x q

These slides are based on Tom Mitchell’s book “Machine Learning” Remarks on KNN Highly effective learning algorithm The distance between instances is calculated based on all features –If some features are irrelevant, or redundant, or noisy, then KNN suffers from the curse of dimensionality –In such a case, feature selection must be performed prior to invoking KNN

These slides are based on Tom Mitchell’s book “Machine Learning” Home assignment #4: Feature selection Compare the following algorithms 1.ID3 – regular ID3 with internal feature selection 2.KNN.all – KNN that uses all the features available 3.KNN.FS – KNN with a priori feature selection (IG) Two datasets: –Spam –Handwritten digits You don’t have to understand the physical meaning of all the coefficients involved !

These slides are based on Tom Mitchell’s book “Machine Learning” Cross-validation Averaging the accuracy of a learning algorithm over a number of experiments N-fold cross-validation: –Partition the available data D into N disjoint subsets T 1, …, T N of equal size (|D| / N) –For n from 1 to N do Training = D \ T i, Testing = T i Induce a classifier using Training, test it on Testing, and measure the accuracy A i –Return (∑ A i ) / N (cross-validated accuracy)