Presentation is loading. Please wait.

Presentation is loading. Please wait.

A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -

Similar presentations


Presentation on theme: "A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -"— Presentation transcript:

1 A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -

2 Solving the differential equation: or in the general form: What is the solution of this type of equation: Try:

3 THE PERCEPTRON: (Classification) Threshold unit: where is the output for input pattern, are the synaptic weights and is the desired output w 1 w 2 w 3 w 4 w 5 x1x2y 111 100 010 000 AND

4 x1x2y 111 100 010 000 AND -1.5 1 0 1 1 Linearly seprable

5 x1x2y 111 101 011 000 OR -0.5 1 0 1 1 Linearly separable

6 Perceptron learning rule: w 1 w 2 w 3 w 4 w 5 Convergence proof: Hertz, Krough, Palmer (HKP) - did you receive the email? Assignment 3a: program in matlab a preceptron with a perceptron learning rule and solve the OR, AND and XOR problems. (Due before Feb 27) Show Demo

7 Summary – what can perceptrons do and how?

8 Linear single layer network: ( approximation, curve fitting) Linear unit: where is the output for input pattern, are the synaptic weights and is the desired output w 1 w 2 w 3 w 4 w 5 Minimize mean square error: or *

9 Linear single layer network: ( approximation, curve fitting) Linear unit: where is the output for input pattern, are the synaptic weights and is the desired output w 1 w 2 w 3 w 4 w 5 Minimize mean square error:

10 The best solution is obtained when E is minimal. For linear neurons there is an exact solution for this called the pseudo-inverse (see HKP). Looking for a solution by gradient descent: E w -gradient Chain rule

11 and Since: Error: Therefore: Which types of problems can a linear network solve?

12 Sigmoidal neurons: Which types of problems can a sigmoidal networks solve? Assignment 3b – Implement a one layer linear and sigmoidal network, fit a 1D a linear, a sigmoid and a quadratic function, for both networks. for example:

13 Multi layer networks: Can solve non linearly separable classification problems. Can approximate any arbitrary function, given ‘enough’ units in the hidden layer. Hidden layer Output layer Input layer

14 Note: is not a vector but a matrix

15 Solving linearly inseparable problems x1x2y 110 101 011 000 XOR Hint: XOR = or and not and

16 How do we learn a multi-layer network The credit assignment problem !.50 -.5 x1x2y 110 101 011 000 XOR 0.5 -0.5 1 -1 1 0.5

17 Gradient descent/ Back Propagation, the solution to the credit assignment problem: Where: From hidden layer to output weights: {

18 Where: For input to hidden layer: and {

19 For input to hidden layer: and Assignment 3c: Program a 2 layer network in matlab, solve the XOR problem. Fit the curve: x(x- 1) between 0 and 1, how many hidden units did you need?

20 Formal neural networks can accomplish many tasks, for example: Perform complex classification Learn arbitrary functions Account for associative memory Some applications: Robotics, Character recognition, Speech recognition, Medical diagnostics. This is not Neuroscience, but is motivated loosely by neuroscience and carries important information for neuroscience as well. For example: Memory, learning and some aspects of development are assumed to be based on synaptic plasticity.

21 What did we learn today? Is BackProp biologically realistic?


Download ppt "A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -"

Similar presentations


Ads by Google