CpSc 810: Machine Learning Design a learning system.

Slides:



Advertisements
Similar presentations
Machine learning Overview
Advertisements

1 Machine Learning: Lecture 1 Overview of Machine Learning (Based on Chapter 1 of Mitchell T.., Machine Learning, 1997)
Reinforcement Learning
1 Classification using instance-based learning. 3 March, 2000Advanced Knowledge Management2 Introduction (lazy vs. eager learning) Notion of similarity.
ChooseMove=16  19 V =30 points. LegalMoves= 16  19, or SimulateMove = 11  15, or …. 16  19,
Slides from: Doug Gray, David Poole
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
CS 484 – Artificial Intelligence1 Announcements Project 1 is due Tuesday, October 16 Send me the name of your konane bot Midterm is Thursday, October 18.
1er. Escuela Red ProTIC - Tandil, de Abril, Introduction How to program computers to learn? Learning: Improving automatically with experience.
Feature Selection Presented by: Nafise Hatamikhah
Machine Learning CPSC 315 – Programming Studio Spring 2009 Project 2, Lecture 5.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Fall 2004.
Machine Learning CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanning.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
1 Some rules  No make-up exams ! If you miss with an official excuse, you get average of your scores in the other exams – at most once.  WP only-if you.
A Brief Survey of Machine Learning
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Introduction to Machine Learning Approach Lecture 5.
Chapter 5 Data mining : A Closer Look.
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Introduction to machine learning
Radial Basis Function Networks
CS 391L: Machine Learning Introduction
Inductive learning Simplest form: learn a function from examples
COMP3503 Intro to Inductive Modeling
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
For Friday Read chapter 18, sections 3-4 Homework: –Chapter 14, exercise 12 a, b, d.
1 Machine Learning What is learning?. 2 Machine Learning What is learning? “That is what learning is. You suddenly understand something you've understood.
Machine Learning Chapter 11.
Computing & Information Sciences Kansas State University Wednesday, 13 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 9 of 42 Wednesday, 13 September.
Machine Learning CSE 681 CH2 - Supervised Learning.
Computing & Information Sciences Kansas State University Lecture 10 of 42 CIS 530 / 730 Artificial Intelligence Lecture 10 of 42 William H. Hsu Department.
Overview of Supervised Learning Overview of Supervised Learning2 Outline Linear Regression and Nearest Neighbors method Statistical Decision.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
CS 445/545 Machine Learning Winter, 2012 Course overview: –Instructor Melanie Mitchell –Textbook Machine Learning: An Algorithmic Approach by Stephen Marsland.
MECN 3500 Inter - Bayamon Lecture 3 Numerical Methods for Engineering MECN 3500 Professor: Dr. Omar E. Meza Castillo
Applying Neural Networks Michael J. Watts
Well Posed Learning Problems Must identify the following 3 features –Learning Task: the thing you want to learn. –Performance measure: must know when you.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
Computing & Information Sciences Kansas State University Wednesday, 12 Sep 2007CIS 530 / 730: Artificial Intelligence Lecture 9 of 42 Wednesday, 12 September.
Learning from observations
Non-Bayes classifiers. Linear discriminants, neural networks.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Chapter 3.
CpSc 881: Machine Learning Evaluating Hypotheses.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 9 of 42 Wednesday, 14.
Chapter 1: Introduction. 2 목 차목 차 t Definition and Applications of Machine t Designing a Learning System  Choosing the Training Experience  Choosing.
CS Inductive Bias1 Inductive Bias: How to generalize on novel data.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Data Mining and Decision Support
Copyright Paula Matuszek Kinds of Machine Learning.
1 Introduction to Machine Learning Chapter 1. cont.
Introduction Machine Learning: Chapter 1. Contents Types of learning Applications of machine learning Disciplines related with machine learning Well-posed.
Well Posed Learning Problems Must identify the following 3 features –Learning Task: the thing you want to learn. –Performance measure: must know when you.
Computing & Information Sciences Kansas State University CIS 530 / 730: Artificial Intelligence Lecture 09 of 42 Wednesday, 17 September 2008 William H.
Machine Learning & Datamining CSE 454. © Daniel S. Weld 2 Project Part 1 Feedback Serialization Java Supplied vs. Manual.
Overfitting, Bias/Variance tradeoff. 2 Content of the presentation Bias and variance definitions Parameters that influence bias and variance Bias and.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
1 Machine Learning Patricia J Riddle Computer Science 367 6/26/2016Machine Learning.
Supervise Learning Introduction. What is Learning Problem Learning = Improving with experience at some task – Improve over task T, – With respect to performance.
Chapter 7. Classification and Prediction
Applying Neural Networks
Spring 2003 Dr. Susan Bridges
A Simple Artificial Neuron
Data Mining Lecture 11.
Overview of Machine Learning
Why Machine Learning Flood of data
Machine Learning Algorithms – An Overview
Presentation transcript:

CpSc 810: Machine Learning Design a learning system

2 Copy Right Notice Most slides in this presentation are adopted from slides of text book and various sources. The Copyright belong to the original authors. Thanks!

3 Well posed learning problems General: given task T, performance measure P, design an algorithm which improves on T as measured by P with experience E Checkers playing: T= play checkers, P = percentage of games won against opponent, E = playing practice against itself Handwritten digit classification: T = map image to {0,...,9}, P = percent of correctly classified images, E = digits from the US postal office

4 Not very interesting learning problems Task: learn to sort a given set of numbers... we can solve this exactly in short time without experience just knowing the definition of sorted lists Task: learn the speed of a ball falling from the leaning tower of Pisa... we know a (sufficiently precise and efficiently computable) physical model ML: for tasks which cannot efficiently be modeled exactly, but experience is available

5 ill-posed learning problems Task: make something reasonable with my ... no objective given, simply delete ? Task: learn to draw a curve through given points... we will not be happy with every curve Task: learn an arbitrary binary series... we cannot learn random series ML: for tasks which are learnable, i.e. structure is available (inductive bias), well- defined target

6 Design a learning system Choose the training experience Choose the target function Choose a representation Choose the parameter fitting algorithm Evaluate the entire system... try to win the world-championship

7 Design a learning system Choose training experience  different learning paradigms: reinforcement learning, supervised learning, unsupervised learning  different availability of data: online/offline examples

8 Design a learning system Choose the target function  make it as easy as possibly  integrate all prior information  learn only the aspects which you have to learn

9 Design a learning system Choose a representation of the target function representation of the input data: integrate all relevant features  make it easy but not more than necessary  curse of dimensionality representation of the function: different types of functions: classification, regression, density estimation, novelty detection, visualization,... different models: symbolic (logical rules, decision tree, prolog program,...), subsymbolic (neural network, statistical estimator,...), parameterized, lazy model,...

10 Design a learning system Estimate the parameters optimize some error/objective on the given data linear/quadratic optimization gradient descent greedy algorithm, heuristics discrete optimization methods such as genetic algorithms statistical methods such as EM mimic biological learning algorithms Hebbian learning imitation learning

11 Design a learning system Evaluation does the system behave well in practice generalization to data not used for training make sure the underlying regularity and not only the given (finite set of) data is learned

12 Learning to Play Checkers T= play checkers, P = percentage of games won against opponent What is experience? What exactly should be learned? How shall it be represented? What specific algorithm to learn it

13 Learning to Play Checkers: Choosing the Training Experience Direct versus Indirect Experience Indirect Experience gives rise to the credit assignment problem and is thus more difficult Teacher versus Learner Controlled Experience the teacher might provide training examples; the learner might suggest interesting examples and ask the teacher for their outcome; or the learner can be completely on its own with no access to correct outcomes How Representative is the Experience? Is the training experience representative of the task the system will actually have to solve? It is best if it is, but such a situation cannot systematically be achieved Our task: train checker to play games against itself.

14 Learning to Play Checkers: Choosing the Target Function Given a set of legal moves, we want to learn how to choose the best move since the best move is not necessarily known, this is an optimization problem ChooseMove: B --> M is called a Target Function ChooseMove, however, is difficult to learn. An easier and related target function to learn is V: B --> R, which assigns a numerical score to each board. The better the board, the higher the score. Operational versus Non-Operational Description of a Target Function An operational description must be given Task: discovering an operational description of the ideal target function V. Function Approximation The actual function can often not be learned and must be approximated

15 Learning to Play Checkers: Choosing the Target Function A possible definition for Target Function V If b is a final board state that is won, then V(b)=100 If b is a final board state that is lost, then V(b)=-100 If b is a final board state that is drawn, then V(b)=0 If b is not a final state in the game, then V(b)=V(b’), where b’ is the best final board state that can be achieved starting from b and playing optimally until the end of game This function give correct values, but is not operational! Why?

16 Choosing a Representation for the Target Function Expressiveness versus Training set size The more expressive the representation of the target function, the closer to the “truth” we can get. However, the more expressive the representation, the more training examples are necessary to choose among the large number of “representable” possibilities.

17 Choosing a Representation for the Target Function A simple representation: for a given board, its state is represented by the board features. Bp(b): number of black pieces on board b. Rp(b): number of red pieces on board b. Bk(b): number of black kings on board b. Rk(b): number of red kings on board b. Bt(b) number of red pieces threatened by black (i.e., which can be taken on black’s next turn) Rt(b): number of black pieces threatened by red

18 Choosing a Representation for the Target Function The target function will be calculated as a linear combination of the board features: Where w 0 through w 6 are numerical coefficients, or weights, to be obtained by learning algorithm. Weights w 1 through w 6 will determine the relative importance of different board features.

19 Obtain Training Examples A training example will contain a specific board state b and the training value V train (b) for b., +100> This is a board state b in which black has won the game (Rp=0). How to assign training values to the intermediate board states One simple approach: Where Successor(b) denotes the next board state following b.

20 Training the System Defining a criterion for success What is the error that needs to be minimized? One common approach is the squared error Choose an algorithm capable of finding weights of a linear function that minimize that error the Least Mean Square (LMS) training rule.

21 LMS Weigh Tuning Rule Do repeatedly Select a training example at random Use the current weights to calculated Compute error(b) For each weight w i, update it as Where x i is the value of the feature i and c is some small constant, e.g. 0.1, to moderate the rate of learning.

22 Final Design for Checkers Learning The Performance System: Takes a new board as input and outputs a trace of the game it played against itself. The Critic: Takes the trace of a game as input and outputs a set of training examples of the target function The Generalizer: Takes training examples as input and outputs a hypothesis which estimates the target function. Good generalization to new cases is crucial. The Experiment Generator: Takes the current hypothesis (currently learned function) as input and outputs a new problem (an initial board state) for the performance system to explore

23 Final Design for Checkers Learning

24 Design Choices

25 Issues in Machine Learning What algorithms are available for learning? How well do they perform? How much training data is sufficient to learn with high confidence? When is it useful to use prior knowledge? Are some training examples more useful than others? What are best tasks for a system to learn? What is the best way for a system to represent its knowledge?