Presentation is loading. Please wait.

Presentation is loading. Please wait.

Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.

Similar presentations


Presentation on theme: "Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward."— Presentation transcript:

1 Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward network Back-prop –Other types of networks

2 Components of a “Well-Posed” Learning Problem Task: the domain of the problem Experience: information about the domain Performance measure: a metric to judge how well the trained system can solve the problem Learner: a computer program whose performance on the task improves (according to the metric) with more experience

3 Example: Classification Task: Predict whether the user might like a movie or not Experience: database of movies the user has seen and the user’s ratings for them Performance Measure: percent of times the system correctly predicts the user’s preference

4 Example: Speech Recognition Task: take dictations from the user Experience: a collection of recordings of acoustic utterances with their transcriptions Performance Measure: percent of words correctly identified

5 Example: Function Modeling Task: approximate an unknown function f(x) Experience: a set of data points: {x i, f(x i )} Performance Measure: average error rate between f(x), the target function, and h(x), the function the system learned, over m test points e.g.

6 Designing a Learner Training experience –Kind of feedback? –A representative sample? –Learner has control? Target function –Specify expected behavior Function representation –Specify form and parameters Learning algorithm

7 Artificial Neural Networks Inspired by neurobiology A network is made up of massively interconnect “neurons” Good for some learning problems –Noisy training examples (contain errors) –Target function input can be best described by a vector (e.g., robot sensor data) –Target function is continuous (differentiable)

8 Perceptron w0w0 w1w1 wnwn 1 x1x1 xnxn O={-1,+1}  O = g(In) = g(xw) = +1 : In >  -1: otherwise n weighted inputs: In = w 0 +x 1 w 1 + x 2 w 2 + … + x n w n = x w An activation function, g(In) …

9 Training a Perceptron Quantify error –compare output with correct answer Update weights to minimize error  is a constant, the learning rate

10 How Powerful Are Perceptrons? A perceptron can represent simple Boolean functions –AND, OR, NOT A network of perceptron can represent any Boolean function A perceptron cannot represent XOR –Why?

11 Linearly Separable Refer to pictures from R&N Fig. 19.9

12 Gradient Descent Guarantees convergence Approximates non-linearly separable functions Search through the weight space Define error as a continuous function of the weights

13

14 Multilayer Network x1x1 x2x2 x n … Input units Hidden units Output units uiui … ujuj w ij OjOj … w ni

15 Training a Multilayer Network Need to update weights to minimize error, but… –How to assign portions of “blame” to each weights fairly? –In a multilayer network, a weight may (eventually) contribute to multiple outputs –Need to back-propagate the error

16 Back-Propagation Between a hidden unit and an output unit: Between an input unit and a hidden unit:

17 Artificial Neural Network Summary Expressiveness: Can approximate any function of a set of attributes Computational efficiency: May take a long time to train to convergence Generalization: generalizes well Sensitivity to noise: very tolerant Transparency: can be used like a black box Prior knowledge: difficult to incorporate


Download ppt "Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward."

Similar presentations


Ads by Google