 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 5.

Slides:



Advertisements
Similar presentations
On Comparing Classifiers : Pitfalls to Avoid and Recommended Approach
Advertisements

Machine Learning: Intro and Supervised Classification
CHAPTER 2: Supervised Learning. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Learning a Class from Examples.
Learning Algorithm Evaluation
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Other Classification Techniques 1.Nearest Neighbor Classifiers 2.Support Vector Machines.
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Spring 2004.
Evaluation (practice). 2 Predicting performance  Assume the estimated error rate is 25%. How close is this to the true error rate?  Depends on the amount.
Evaluation.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Assessing and Comparing Classification Algorithms Introduction Resampling and Cross Validation Measuring Error Interval Estimation and Hypothesis Testing.
Model Evaluation Metrics for Performance Evaluation
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Fall 2005.
Northwestern University Winter 2007 Machine Learning EECS Machine Learning Lecture 13: Computational Learning Theory.
Measuring Model Complexity (Textbook, Sections ) CS 410/510 Thurs. April 27, 2007 Given two hypotheses (models) that correctly classify the training.
© sebastian thrun, CMU, The KDD Lab Intro: Outcome Analysis Sebastian Thrun Carnegie Mellon University
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Fall 2004.
Evaluation.
Learning From Data Chichang Jou Tamkang University.
Evaluating Hypotheses
CS 4700: Foundations of Artificial Intelligence
Experimental Evaluation
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
General Mining Issues a.j.m.m. (ton) weijters Overfitting Noise and Overfitting Quality of mined models (some figures are based on the ML-introduction.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Part I: Classification and Bayesian Learning
On Comparing Classifiers: Pitfalls to Avoid and Recommended Approach Published by Steven L. Salzberg Presented by Prakash Tilwani MACS 598 April 25 th.
Computer Vision Lecture 8 Performance Evaluation.
1 Machine Learning: Lecture 5 Experimental Evaluation of Learning Algorithms (Based on Chapter 5 of Mitchell T.., Machine Learning, 1997)
1  The goal is to estimate the error probability of the designed classification system  Error Counting Technique  Let classes  Let data points in class.
CLassification TESTING Testing classifier accuracy
COMP3503 Intro to Inductive Modeling
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Error estimation Data Mining II Year Lluís Belanche Alfredo Vellido.
Evaluating Hypotheses Reading: Coursepack: Learning From Examples, Section 4 (pp )
Ensemble Classification Methods Rayid Ghani IR Seminar – 9/26/00.
Experimental Evaluation of Learning Algorithms Part 1.
Learning from Observations Chapter 18 Through
1 CS 391L: Machine Learning: Experimental Evaluation Raymond J. Mooney University of Texas at Austin.
CpSc 810: Machine Learning Evaluation of Classifier.
1 Machine Learning: Lecture 8 Computational Learning Theory (Based on Chapter 7 of Mitchell T.., Machine Learning, 1997)
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
Evaluating Results of Learning Blaž Zupan
CpSc 881: Machine Learning Evaluating Hypotheses.
Concept learning, Regression Adapted from slides from Alpaydin’s book and slides by Professor Doina Precup, Mcgill University.
Machine Learning Chapter 5. Evaluating Hypotheses
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
1 CSI5388 Current Approaches to Evaluation (Based on Chapter 5 of Mitchell T.., Machine Learning, 1997)
Chapter5: Evaluating Hypothesis. 개요 개요 Evaluating the accuracy of hypotheses is fundamental to ML. - to decide whether to use this hypothesis - integral.
Chapter 6 Cross Validation.
1 Evaluation of Learning Models Literature: Literature: T. Mitchel, Machine Learning, chapter 5 T. Mitchel, Machine Learning, chapter 5 I.H. Witten and.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 5: Credibility: Evaluating What’s Been Learned.
Goal of Learning Algorithms  The early learning algorithms were designed to find such an accurate fit to the data.  A classifier is said to be consistent.
Classification Ensemble Methods 1
Validation methods.
CS 8751 ML & KDDComputational Learning Theory1 Notions of interest: efficiency, accuracy, complexity Probably, Approximately Correct (PAC) Learning Agnostic.
Physical Science and You Chapter One: Studying Physics and Chemistry Chapter Two: Experiments and Variables Chapter Three: Key Concepts in Physical Science.
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Evaluating Classifiers. Reading for this topic: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)
7. Performance Measurement
Computational Learning Theory
CSE 4705 Artificial Intelligence
Data Mining Practical Machine Learning Tools and Techniques
Computational Learning Theory
Computational Learning Theory
Machine Learning: Lecture 6
Evaluating Hypothesis
Lecture 14 Learning Inductive inference
Introduction to Machine learning
Machine Learning: Lecture 5
Presentation transcript:

 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 5. Evaluation of Empirical Inductive Learners

 2003, G.Tecuci, Learning Agents Laboratory 2 Overview Introduction Computational learning theory Empirical evaluation: Single partitioning Recommended reading Empirical evaluation: Resampling

 2003, G.Tecuci, Learning Agents Laboratory 3 Introduction Suppose we have collected a body of training examples, adopted a learning bias, implemented the learning algorithm, executed the algorithm, and learned the concept “c” represented by the examples. There are several questions we may ask about this process: Can we believe the we have learned the right concept? What is the likelihood that “c” will correctly classify previously unseen examples? How can we have confidence that the concept “c” is approximately correct? There are two possible answers, a theoretical answer and an experimental one.

 2003, G.Tecuci, Learning Agents Laboratory 4 The Computational Learning Theory The Computational Learning Theory, pioneered by Valiant, is concerned with finding theoretical answers to the previous questions. In this theory, learning is viewed as function reconstruction Given: a set of input-output pairs {x, f(x)} for a boolean function f. Determine: an expression f1 that provides a good approximation of the boolean function ff : {0,1} n --> {0,1} The Valiant framework provides bounds on the number of training examples required for a given bias, in order to have high confidence that the learned hypothesis f1 is approximately correct. That is, how many training examples would one need so that the probability that the error rate of f1 is less than  is greater than 1 -  : Probability (error rate of f1 ≤  ) ≥ 1 - 

 2003, G.Tecuci, Learning Agents Laboratory 5 The Computational Learning Theory (cont.) This style of analysis is called probably approximately correct (PAC) learning. The basic idea is to analyze the expressiveness of the hypothesis space. If a restricted hypothesis space H is very small Then it is unlikely that a learning algorithm could by chance succeed in finding a hypothesis f1  H consistent with the training examples. Therefore, it is more likely that f1, if it is found, is a good approximation of the correct hypothesis.

 2003, G.Tecuci, Learning Agents Laboratory 6 The Computational Learning Theory (cont.) The theoretical analysis has provided insight into the relationship between: - the number of training examples, - the bias of the learning algorithm, and - the confidence that we can have in the hypothesis f1 produced by the algorithm. This analysis has been successful only for simple learning algorithms. Most applied work in machine learning employs experimental techniques for determining the correctness of f1.

 2003, G.Tecuci, Learning Agents Laboratory 7 Overview Introduction Computational learning theory Empirical evaluation: Single partitioning Recommended reading Empirical evaluation: Resampling

 2003, G.Tecuci, Learning Agents Laboratory 8 Simple partitioning: the holdout method 1. The available examples are randomly broken into two disjoint groups: the training set and the testing set; 2. The concept is learned by using only the examples from the training set; 3. The learned concept is then used to classify examples from the testing set; 4. The obtained results are compared with the correct classification to produce an error rate.

 2003, G.Tecuci, Learning Agents Laboratory 9 Discussion How does the number of examples affects the result of the evaluation? How does the distribution of examples affects the result of the evaluation? How to evaluate if there are very few examples? How to reuse examples?

 2003, G.Tecuci, Learning Agents Laboratory 10 Overview Introduction Computational learning theory Empirical evaluation: Single partitioning Recommended reading Empirical evaluation: Resampling

 2003, G.Tecuci, Learning Agents Laboratory 11 Resampling: the leave-one-out method Let us consider that the number of available examples is 'n'. A concept is learned from n-1 examples and is tested on the remaining example. This is repeated n times, each time leaving out a different example. The error rate is the total number of errors on the single test case divided by n.

 2003, G.Tecuci, Learning Agents Laboratory 12 Discussion How is the error estimate likely to compare with single partitioning? What is a likely problem with the leave one out method and how could it be avoided? What about the repeatability of the experimental results? Why is this important?

 2003, G.Tecuci, Learning Agents Laboratory 13 Resampling: the cross-validation method In k-fold cross validation, the cases are randomly divided into k (usually 10) mutually disjoint sets of approximately equal size (of at least 30 examples). The concept is learned from the examples in k-1 sets, and is tested on the examples from the remaining set. This is repeated k times, once for each set (i.e. each set is once used as a test set). The average error rates over all k sets is the cross- validated error rate.

 2003, G.Tecuci, Learning Agents Laboratory 14 Resampling vs single partitioning Resampling is a powerful idea. With a single train and test partition, too few cases in the training group can lead to the learning of a poor concept, while too few test cases can lead to erroneous error estimates. Resampling allows for more accurate estimates of the error rates while training on most cases. Resampling allows the duplication of the analysis conditions in future experiments on the same data.

 2003, G.Tecuci, Learning Agents Laboratory 15 Discussion How could we compare two learning algorithms? What can be said about the result of the comparison?

 2003, G.Tecuci, Learning Agents Laboratory 16 Other types of experiments Determine other characteristics of the learning methods: the speed of learning; the asymptotic behavior and the number of examples needed to approximate this behavior; predictive accuracy versus concept complexity; the influence of different types of noise on the predictive accuracy; the influence of different biases on the predictive accuracy; etc.

 2003, G.Tecuci, Learning Agents Laboratory 17 Recommended reading Mitchell T.M., Machine Learning, Chapter 5: Evaluating Hypotheses, pp , McGraw Hill, Weiss, S.M., Kapouleas, I., An Experimental Comparison of Pattern Recognition, Neural Nets, and Machine Learning Classification Methods, in Readings in Machine Learning. Kibler D., Langley P., Machine Learning as an Experimental Science, in Readings in Machine Learning.