INSTANCE-BASE LEARNING

Slides:



Advertisements
Similar presentations
1 Classification using instance-based learning. 3 March, 2000Advanced Knowledge Management2 Introduction (lazy vs. eager learning) Notion of similarity.
Advertisements

Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
K-means method for Signal Compression: Vector Quantization
Support Vector Machines
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.
Preventing Overfitting Problem: We don’t want to these algorithms to fit to ``noise’’ The generated tree may overfit the training data –Too many branches,
Instance Based Learning
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
Statistical Classification Rong Jin. Classification Problems X Input Y Output ? Given input X={x 1, x 2, …, x m } Predict the class label y  Y Y = {-1,1},
1er. Escuela Red ProTIC - Tandil, de Abril, Instance-Based Learning 4.1 Introduction Instance-Based Learning: Local approximation to the.
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
Instance Based Learning
Università di Milano-Bicocca Laurea Magistrale in Informatica Corso di APPRENDIMENTO E APPROSSIMAZIONE Lezione 8 - Instance based learning Prof. Giancarlo.
K nearest neighbor and Rocchio algorithm
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Instance based learning K-Nearest Neighbor Locally weighted regression Radial basis functions.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Nearest Neighbor Models (Reading: Chapter.
Instance Based Learning
Instance-Based Learning
These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.
1 Nearest Neighbor Learning Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Intro AI.
Aprendizagem baseada em instâncias (K vizinhos mais próximos)
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Instance Based Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán) 1.
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor.
Instance-Based Learners So far, the learning methods that we have seen all create a model based on a given representation and a training set. Once the.
CS Instance Based Learning1 Instance Based Learning.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
DATA MINING LECTURE 10 Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
11/12/2012ISC471 / HCI571 Isabelle Bichindaritz 1 Prediction.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Statistical Inference (By Michael Jordon) l Bayesian perspective –conditional perspective—inferences.
1 Instance Based Learning Ata Kaban The University of Birmingham.
Instance Based Learning
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
CpSc 881: Machine Learning Instance Based Learning.
CpSc 810: Machine Learning Instance Based Learning.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
Outline K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Lazy Learners K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Tuesday, November 23, 1999.
KNN Classifier.  Handed an instance you wish to classify  Look around the nearby region to see what other classes are around  Whichever is most common—make.
Machine Learning ICS 178 Instructor: Max Welling Supervised Learning.
DATA MINING LECTURE 10b Classification k-nearest neighbor classifier
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Thursday, 05 April 2007 William.
Meta-learning for Algorithm Recommendation Meta-learning for Algorithm Recommendation Background on Local Learning Background on Algorithm Assessment Algorithm.
CS Machine Learning Instance Based Learning (Adapted from various sources)
K-Nearest Neighbor Learning.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.
CS 8751 ML & KDDInstance Based Learning1 k-Nearest Neighbor Locally weighted regression Radial basis functions Case-based reasoning Lazy and eager learning.
1 Instance Based Learning Soongsil University Intelligent Systems Lab.
1 Instance Based Learning Soongsil University Intelligent Systems Lab.
Instance Based Learning
Data Mining: Concepts and Techniques (3rd ed
Instance Based Learning (Adapted from various sources)
K Nearest Neighbor Classification
Prepared by: Mahmoud Rafeek Al-Farra
یادگیری بر پایه نمونه Instance Based Learning Instructor : Saeed Shiry
Instance Based Learning
Chap 8. Instance Based Learning
Machine Learning: UNIT-4 CHAPTER-1
Nearest Neighbors CSC 576: Data Mining.
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Presentation transcript:

INSTANCE-BASE LEARNING Instance-based learning methods simply store the training examples instead of learning explicit description of the target function. Generalizing the examples is postponed until a new instance must be classified. When a new instance is encountered, its relationship to the stored examples is examined in order to assign a target function value for the new instance. Instance-based learning includes nearest neighbor, locally weighted regression and case-based reasoning methods. Instance-based methods are sometimes referred to as lazy learning methods because they delay processing until a new instance must be classified. A key advantage of lazy learning is that instead of estimating the target function once for the entire instance space, these methods can estimate it locally and differently for each new instance to be classified.

k-Nearest Neighbor Learning k-Nearest Neighbor Learning algorithm assumes all instances correspond to points in the n-dimensional space Rn The nearest neighbors of an instance are defined in terms of Euclidean distance. Euclidean distance between the instances xi = <xi1,…,xin> and xj = <xj1,…,xjn> are: For a given query instance xq, f(xq) is calculated the function values of k-nearest neighbor of xq

k-Nearest Neighbor Learning Store all training examples <xi,f(xi)> Calculate f(xq) for a given query instance xq using k-nearest neighbor Nearest neighbor: (k=1) Locate the nearest traing example xn, and estimate f(xq) as f(xq)  f(xn) k-Nearest neighbor: Locate k nearest traing examples, and estimate f(xq) as If the target function is real-valued, take mean of f-values of k nearest neighbors. f(xq) = If the target function is discrete-valued, take a vote among f-values of k nearest neighbors. CS464 Introduction to Machine Learning

When To Consider Nearest Neighbor Instances map to points in Rn Less than 20 attributes per instance Lots of training data Advantages Training is very fast Learn complex target functions Can handle noisy data Does not loose any information Disadvantages Slow at query time Easily fooled by irrelevant attributes CS464 Introduction to Machine Learning

Distance-Weighted kNN CS464 Introduction to Machine Learning

Curse of Dimensionality CS464 Introduction to Machine Learning

Locally Weighted Regression KNN forms local approximation to f for each query point xq Why not form an explicit approximation f(x) for region surrounding xq  Locally Weighted Regression Locally weighted regression uses nearby or distance-weighted training examples to form this local approximation to f. We might approximate the target function in the neighborhood surrounding x, using a linear function, a quadratic function, a multilayer neural network. The phrase "locally weighted regression" is called local because the function is approximated based only on data near the query point, weighted because the contribution of each training example is weighted by its distance from the query point, and regression because this is the term used widely in the statistical learning community for the problem of approximating real-valued functions. CS464 Introduction to Machine Learning

Locally Weighted Regression Given a new query instance xq, the general approach in locally weighted regression is to construct an approximation f that fits the training examples in the neighborhood surrounding xq. This approximation is then used to calculate the value f(xq), which is output as the estimated target value for the query instance. CS464 Introduction to Machine Learning

Locally Weighted Linear Regression Kernel function K is the function of distance that is used to determine the weight of each training example. CS464 Introduction to Machine Learning

Radial Basis Functions One approach to function approximation that is closely related to distance-weighted regression and also to artificial neural networks is learning with radial basis functions. The learned hypothesis is a function of the form CS464 Introduction to Machine Learning

Radial Basis Function Networks Each hidden unit produces an activation determined by a Gaussian function centered at some instance xu. Therefore, its activation will be close to zero unless the input x is near xu. The output unit produces a linear combination of the hidden unit activations. CS464 Introduction to Machine Learning

CS464 Introduction to Machine Learning Case-based reasoning Instance-based methods lazy classification based on classifications of near (similar) instances data: points in n-dim. space Case-based reasoning as above, but data represented in symbolic form New distance metrics required CS464 Introduction to Machine Learning

CS464 Introduction to Machine Learning Lazy & eager learning Lazy: generalize at query time kNN, CBR Eager: generalize before seeing query Radial basis, ID3, … Difference eager must create global approximation lazy can create many local approximation lazy can represent more complex functions using same H (H = linear functions) CS464 Introduction to Machine Learning