Download presentation
Presentation is loading. Please wait.
Published byPenelope Moore Modified over 8 years ago
1
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -
2
Solving the differential equation: or in the general form: What is the solution of this type of equation: Try:
3
THE PERCEPTRON: (Classification) Threshold unit: where is the output for input pattern, are the synaptic weights and is the desired output w 1 w 2 w 3 w 4 w 5 x1x2y 111 100 010 000 AND
4
x1x2y 111 100 010 000 AND -1.5 1 0 1 1 Linearly seprable
5
x1x2y 111 101 011 000 OR -0.5 1 0 1 1 Linearly separable
6
Perceptron learning rule: w 1 w 2 w 3 w 4 w 5 Convergence proof: Hertz, Krough, Palmer (HKP) - did you receive the email? Assignment 3a: program in matlab a preceptron with a perceptron learning rule and solve the OR, AND and XOR problems. (Due before Feb 27) Show Demo
7
Summary – what can perceptrons do and how?
8
Linear single layer network: ( approximation, curve fitting) Linear unit: where is the output for input pattern, are the synaptic weights and is the desired output w 1 w 2 w 3 w 4 w 5 Minimize mean square error: or *
9
Linear single layer network: ( approximation, curve fitting) Linear unit: where is the output for input pattern, are the synaptic weights and is the desired output w 1 w 2 w 3 w 4 w 5 Minimize mean square error:
10
The best solution is obtained when E is minimal. For linear neurons there is an exact solution for this called the pseudo-inverse (see HKP). Looking for a solution by gradient descent: E w -gradient Chain rule
11
and Since: Error: Therefore: Which types of problems can a linear network solve?
12
Sigmoidal neurons: Which types of problems can a sigmoidal networks solve? Assignment 3b – Implement a one layer linear and sigmoidal network, fit a 1D a linear, a sigmoid and a quadratic function, for both networks. for example:
13
Multi layer networks: Can solve non linearly separable classification problems. Can approximate any arbitrary function, given ‘enough’ units in the hidden layer. Hidden layer Output layer Input layer
14
Note: is not a vector but a matrix
15
Solving linearly inseparable problems x1x2y 110 101 011 000 XOR Hint: XOR = or and not and
16
How do we learn a multi-layer network The credit assignment problem !.50 -.5 x1x2y 110 101 011 000 XOR 0.5 -0.5 1 -1 1 0.5
17
Gradient descent/ Back Propagation, the solution to the credit assignment problem: Where: From hidden layer to output weights: {
18
Where: For input to hidden layer: and {
19
For input to hidden layer: and Assignment 3c: Program a 2 layer network in matlab, solve the XOR problem. Fit the curve: x(x- 1) between 0 and 1, how many hidden units did you need?
20
Formal neural networks can accomplish many tasks, for example: Perform complex classification Learn arbitrary functions Account for associative memory Some applications: Robotics, Character recognition, Speech recognition, Medical diagnostics. This is not Neuroscience, but is motivated loosely by neuroscience and carries important information for neuroscience as well. For example: Memory, learning and some aspects of development are assumed to be based on synaptic plasticity.
21
What did we learn today? Is BackProp biologically realistic?
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.