Chapter 11: Learning Introduction

Slides:



Advertisements
Similar presentations
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Advertisements

Machine Learning: Intro and Supervised Classification
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Combining Inductive and Analytical Learning Ch 12. in Machine Learning Tom M. Mitchell 고려대학교 자연어처리 연구실 한 경 수
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Spring 2004.
CII504 Intelligent Engine © 2005 Irfan Subakti Department of Informatics Institute Technology of Sepuluh Nopember Surabaya - Indonesia.
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering CSCE 580 Artificial Intelligence Ch.5 [P]: Propositions and Inference Sections.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
01 -1 Lecture 01 Artificial Intelligence Topics –Introduction –Knowledge representation –Knowledge reasoning –Machine learning –Applications.
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Fall 2005.
Machine Learning II Decision Tree Induction CSE 473.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Learning from Observations Chapter 18 Section 1 – 4.
18 LEARNING FROM OBSERVATIONS
An Introduction to Machine Learning In the area of AI (earlier) machine learning took a back seat to Expert Systems Expert system development usually consists.
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Fall 2004.
CSE 5731 Lecture 26 Introduction to Machine Learning CSE 573 Artificial Intelligence I Henry Kautz Fall 2001.
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
Machine Learning CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanning.
Three kinds of learning
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Machine Learning: Symbol-Based
1 Learning: Overview and Decision Trees  Learning: Intro and General concepts  Decision Trees Basic Concepts Finding the best DT.
Part I: Classification and Bayesian Learning
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Learning CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
COMP3503 Intro to Inductive Modeling
For Friday Read chapter 18, sections 3-4 Homework: –Chapter 14, exercise 12 a, b, d.
Classification. An Example (from Pattern Classification by Duda & Hart & Stork – Second Edition, 2001)
Machine Learning CSE 681 CH2 - Supervised Learning.
Lecture 10: 8/6/1435 Machine Learning Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Learning from Observations Chapter 18 Through
CPS 270: Artificial Intelligence Machine learning Instructor: Vincent Conitzer.
Learning from observations
Learning Agents MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
CS Inductive Bias1 Inductive Bias: How to generalize on novel data.
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Machine Learning Margaret H. Dunham Department of Computer Science and Engineering Southern.
Data Mining and Decision Support
Learning From Observations Inductive Learning Decision Trees Ensembles.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
CSE573 Autumn /11/98 Machine Learning Administrative –Finish this topic –The rest of the time is yours –Final exam Tuesday, Mar. 17, 2:30-4:20.
Who am I? Work in Probabilistic Machine Learning Like to teach 
Learning from Observations
Learning from Observations
Chapter 5: Representing Knowledge
CS 9633 Machine Learning Concept Learning
Chapter 10: Using Uncertain Knowledge
CS 9633 Machine Learning Inductive-Analytical Methods
Chapter 10 (part 3): Using Uncertain Knowledge
Presented By S.Yamuna AP/CSE
Introduction to Data Mining, 2nd Edition by
Prepared by: Mahmoud Rafeek Al-Farra
Machine Learning: Lecture 3
Overview of Machine Learning
Dr. Unnikrishnan P.C. Professor, EEE
Supervised vs. unsupervised Learning
Classification and Prediction
Why Machine Learning Flood of data
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Learning from Observations
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
CS 5310 Data Mining Hong Lin.
Lecture 14 Learning Inductive inference
Chapter 12: Building Situated Robots
Chapter 1: Computational Intelligence and Knowledge
Representations & Reasoning Systems (RRS) (2.2)
Instructor: Vincent Conitzer
Presentation transcript:

Chapter 11: Learning Introduction D. Poole, A. Mackworth, and R. Goebel, Computational Intelligence: A Logical Approach, Oxford University Press, January 1998 Introduction

CS 426 Artificial Intelligence, Ch 11: Learning Learning is the ability to improve one’s behavior based on experience. The range of behaviors is expanded: the agent can do more. The accuracy on tasks is improved: the agent can do things better. The speed is improved: the agent can do things faster. Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

Components of a learning problem The following components are part of any learning problem: Task. The behavior or task that’s being improved. For example: classification, acting in an environment Data. The experiences that are being used to improve performance in the task. Measure of improvement. How can the improvement be measured? For example: increasing accuracy in prediction, new skills that were not present initially, improved speed. Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

CS 426 Artificial Intelligence, Ch 11: Learning Learning Task Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

Learning Architecture Induction: infer internal representation based on examples: classify unseen objects Deduction: deriving consequences of a knowledge base Abduction: hypothesizing what may be true about a particular case Decision trees Neural networks Case bases Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

Choosing a representation The richer the representation, the more useful it is for subsequent problem solving. The richer the representation, the more difficult it is to learn. Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

CS 426 Artificial Intelligence, Ch 11: Learning Common Learning Tasks Supervised classification. Given a set of pre-classified training examples, classify a new instance. Unsupervised learning. Find natural classes for examples. Reinforcement learning. Determine what to do based on rewards and punishments (feedback after actions: Credit Assignment Problem). Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

Example of Classification Data Set of properties or features examples Data that the infobot obtained from observing a user deciding whether to read/skip articles depending on whether the author was known or not, whether the article started a new thread or was a follow-up, the length of the article or whether it was read at home or at work. We want to classify new examples on property “Action” based on the examples’ “Author”, “Thread”, “Length”, and “Where”. Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

CS 426 Artificial Intelligence, Ch 11: Learning Feedback Learning tasks can be characterized by the feedback given to the learner. Supervised learning. What has to be learned is specified for each example. Unsupervised learning. No classifications are given; the learner has to discover categories and regularities in the data. Reinforcement learning. Feedback occurs after a sequence of actions. Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

CS 426 Artificial Intelligence, Ch 11: Learning Measuring Success The measure of success is not how well the agent performs on the training examples, but how well the agent performs for new examples (apparent versus true error rate). Consider two agents: P claims the negative examples seen (during training) are the only negative examples. Every other instance is positive. N claims the positive examples seen are the only positive examples. Every other instance is negative. Both agents correctly classify every training example, but disagree on every other new example: success in learning should not be judged on correctly classifying the training set but being able to classify unseen examples. Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

CS 426 Artificial Intelligence, Ch 11: Learning Bias The tendency to prefer one hypothesis over another is called a bias. Saying a hypothesis is better than the agents N’s or P’s hypothesis isn’t something that’s obtained from the training data: but is external to the data. To have any inductive process make predictions on unseen data, you need a bias (external information). What constitutes a good bias is an empirical question about which biases work best in practice. Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

CS 426 Artificial Intelligence, Ch 11: Learning Learning as Search Given a representation and a bias, the problem of learning can be reduced to one of search. Learning is search through the space of possible representations: Looking for the representation or representations that best fits the data, given the bias. These search spaces are typically prohibitively large for systematic search. Use of a hill climbing search. A learning algorithm is made of a search space, an evaluation function (heuristic), and a search method. Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning

CS 426 Artificial Intelligence, Ch 11: Learning Noise Data aren’t perfect: some of the attributes (features) are assigned the wrong value the attributes given are inadequate to predict the classification (poor features feature extraction) there are examples with missing attributes Apparent error rates tend to be biased optimistically. The true error rate is almost invariably higher than the apparent error rate. This happens because of overfitting (overlearning) which occurs because of random correlations in the training set. Dr. Djamel Bouchaffra CS 426 Artificial Intelligence, Ch 11: Learning