Download presentation

Presentation is loading. Please wait.

Published byCole Hardiman Modified over 3 years ago

1
Artificial Neural Networks Brian Talecki CSC 8520 Villanova University

2
ANN - Artificial Neural Network A set of algebraic equations and functions which determine the best output given a set of inputs. An artificial neural network is modeled on a very simplified version of the a human neuron which make up the human nervous system. Although the brain operates at 1 millionth the speed of modern computers, it functions faster than computers because of the parallel processing structure of the nervous system.

3
Human Nerve Cell picture from: G5AIAI Introduction to AI by Graham Kendall www.cs.nott.ac.uk/~gxk/courses/g5aiai

4
At the synapse – the nerve cell releases a chemical compounds called neurotransmitters, which excite or inhibit a chemical / electrical discharge in the neighboring nerve cells. The summation of the responses of the adjacent neurons will elicit the appropriate response in the neuron.

5
Brief History of ANN McCulloch and Pitts (1943) designed the first neural network Hebb (1949) who developed the first learning rule. If two neurons were active at the same time then the strength between them should be increased. Rosenblatt (1958) – introduced the concept of a perceptron which performed pattern recognition. Widrow and Hoff (1960) introduced the concept of the ADALINE (ADAptive Linear Element). The training rule was based on the idea of Least-Mean-Squares learning rule which minimizing the error between the computed output and the desired output. Minsky and Papert (1969) stated that the perceptron was limited in its ability to recognize features that were separated by linear boundaries. “Neural Net Winter” Kohonen and Anderson – independently developed neural networks that acted like memories. Webros(1974) – developed the concept of back propagation of an error to train the weights of the neural network. McCelland and Rumelhart (1986) published the paper on back propagation algorithm. “Rebirth of neural networks”. Today - they are everywhere a decision can be made. Source : G5AIAI - Introduction to Artificial Intelligence Graham Kendall:

6
Basic Neural Network Inputs – normally a vector of measured parameters Bias – may/may not be added f() – transfer or activation function Outputs = f( ∑ W p + b) f() W ∑ Outputs InputsInputs b - Bias ∑ Wp +b T

7
Activation Functions Source: Supervised Neural Network Introduction CISC 873. Data Mining Yabin Meng

8
Log Sigmoidal Function Source: Artificial Neural Networks Colin P. Fahey http://www.colinfahey.com/2003apr20_neuron/index.htm

9
Hard Limit Function 1.0 x y

10
Log Sigmoid and Derivative Source : The Scientist and Engineer’s Guide to Digital Signal Processing by Steven Smith

11
Derivative of the Log Sigmoidal Function s(x) = (1 + e ) s’(x) = -(1+e ) * (-e ) = e * (1+ e ) = ( e ) * ( 1 ) (1+ e ) ( 1 + e ) = (1 + e – 1) * ( 1 ) ( 1+ e ) ( 1 + e ) = (1 - ( 1 ) ) * ( 1 ) (1+ e ) (1 + e ) s’(x) = (1-s(x)) * s(x) -x -2 -x -2 -x Derivative is important for the back error propagation algorithm used to train multilayer neural networks.

12
Example : Single Neuron Given : W = 1.3, p = 2.0, b = 3.0 Wp + b = 1.3(2.0) + 3.0 = 5.6 Linear: f(5.6) = 5.6 Hard limit f(5.6) = 1.0 Log Sigmoidal f(5.6) = 1/(1+exp(-5.6) = 1/(1+0.0037) =.9963

13
Simple Neural Network One neuron with a linear activation function => Straight Line Recall the equation of a straight Line : y = mx +b m is the slope (weight), b is the y-intercept (bias). Bad Good Decision Boundary p2 p1 Mp1 + b >= p2 Mp1 + b < p2

14
Perceptron Learning Extend our simple perceptron to two inputs and hard limit activation function F() W bias Output W1 W2 o = f ( ∑ W p + b) W is the weight matrix p is the input vector o is our scalar output p1 p2 Hard limit function ∑ T

15
Rules of Matrix Math Addition/Subtraction 1 2 3 9 8 7 10 10 10 4 5 6 +/- 6 5 4 = 10 10 10 7 8 9 3 2 1 10 10 10 Multiplication by a scalar Transpose a 1 2 = a 2a 1 = 1 2 3 4 3a 4a 2 Matrix Multiplication 2 4 5 = 18, 5 2 4 = 10 20 2 2 4 8 T

16
Data Points for the AND Function q1 = 0, o1 = 0 0 q2 = 1, o2 = 0 0 q3 = 0, o3 = 0 1 q4 = 1, o4 = 1 1 Truth Table P1 P2 O 0 0 0 0 1 0 1 0 0 1 1 1

17
Weight Vector and the Decision Boundary W = 1.0 1.0 Magnitude and Direction Decision Boundary is the line where W p = b or W p – b = 0 TT W p < b W p > b T T As we adjust the weights and biases of the neural network, we change the magnitude and direction of the weight vector or the slope and intercept of the decision boundary

18
Perceptron Learning Rule Adjusting the weights of the Perceptron Perceptron Error : Difference between the desired and derived outputs. e = Desired – Derived When e = 1 W new = W old + p When e = -1 W new = W old - p When e = 0 W new = W old Simplifing W new = W old + λ * ep b new = b old + e λ is the learning rate ( = 1 for the perceptron).

19
AND Function Example Start with W1 = 1, W2 = 1, and b = -1 W p + b => t - a = e 1 1 0 + -1 => 0 - 0 = 0 N/C 0 1 1 0 + -1 => 0 - 1 = -1 1 1 0 1 + -2 => 0 - 0 = 0 N/C 0 1 0 1 + -2 => 1 - 0 = 1 1 T

20
W p + b => t - a = e 2 1 0 + -1 => 0 - 0 = 0 N/C 0 2 1 0 + -1 => 0 - 1 = -1 1 2 0 1 + -2 => 0 - 1 = -1 0 1 0 1 + -3 => 1 - 0 = 1 1 T

21
W p + b => t - a = e 2 1 0 + -2 => 0 - 0 = 0 N/C 0 2 1 0 + -2 => 0 - 0 = 0 N/C 1 2 1 1 + -2 => 0 - 1 = -1 0 1 1 1 + -3 => 1 - 0 = 1 1 T

22
W p + b => t - a = e 2 2 0 + -2 => 0 - 0 = 0 N/C 0 2 2 0 + -2 => 0 - 1 = -1 1 2 1 1 + -3 => 0 - 0 = 0 N/C 0 2 1 1 + -3 => 1 - 1 = 0 N/C 1 T

23
W p + b => t - a = e 2 1 0 + -3 => 0 - 0 = 0 N/C 0 2 1 0 + -3 => 0 - 0 = 0 N/C 1 Done ! T 2 f() 1 Hardlim() p1 p2 Σ -3

24
XOR Function Truth Table X Y Z = (X and not Y) or (not X and Y) 0 0 0 0 1 1 1 0 1 1 1 0 1 0 No single decision boundary can separate the favorable and unfavorable outcomes. z x y We will need a more complicated neural net to realize this function Circuit Diagram

25
XOR Function – Multilayer Perceptron W5 W6 W1 W3 W2 W4 f1() f() z b2 b11 b12 Σ Σ x y Z = f (W5*f1(W1*x + W4*y+b11) +W6*f1(W2*x + W3*y+b12)+b2) Weights of the neural net are independent of each other, so that we can compute the partial derivatives of z with respect to the weights of the network. i.e. δz / δW1, δz / δW2, δz / δW3, δz / δW4, δz / δW5, δz / δW6

26
Back Propagation Diagram Neural Networks and Logistic Regression by Lucila Ohno-Machado Decision Systems Group, Brigham and Women’s Hospital, Department of Radiology

27
Back Propagation Algorithm This algorithm to train Artificial Neural Networks (ANN) depends to two basic concepts : a) Reduced the Sum Squared Error, SSE, to an acceptable value. b) Reliable data to train your network under your supervision. Simple case : Single input no bias neural net. zx W1 n1 f1 W2 a1n2 T = desired output f2

28
BP Equations n1 = W1 * x a1 = f1(n1) = f1(W1 * x) n2 = W2 * a1 = W2 * f1(n1) = W2 * f1(W1 * x) z = f2(n2) = f2(W2 * f1(W1 * x)) SSE = ½ (z – T) Lets now take the partial derivatives δSSE/ δW2 = (z - T) * δ(z - T)/ δW2 = (z – T) * δz/ δW2 = (z - T) * δf2(n2)/δW2 Chain Rule δf2(n2)/δW2 = (δf2(n2)/δn2)* (δn2/δW2) = (δf2(n2)/δn2)* a1 δSSE/ δW2 = (z - T) * (δf2(n2)/δn2)* a1 Define λ to our learning rate (0 < λ < 1, typical λ = 0.2) Compute our new weight: W2(k+1) = W2(k) - λ (δSSE/ δW2) = W2(k) - λ ((z - T) * (δf2(n2)/δn2)* a1) 2

29
Sigmoid function: δf2(n2)/δn2 = f2(n2)(1 – f2(n2)) = z(1 – z) Therefore: W2(k+1) = W2(k) - λ ((z - T) * ( z(1 –z) )* a1) Analysis for W1 n1 = W1 * x a1 = f1(W1*x) n2 = W2 * f1(n1) = W2 * f1(W1 * x) δSSE/ δW1 = (z - T) * δ(z -T )/ δW1 = (z - T) * δz/ δW1 = (z - T) * δf2(n2)/δW1 δf2(n2)/δW1 = (δf2(n2)/δn2)* (δn2/δW1) -> Chain Rule δn2/δW1 = W2 * (δf1(n1)/δW1) = W2 * (δf1(n1)/δn1) * (n1/δW1) -> Chain Rule = W2 * (δf1(n1)/δn1) * x δSSE/ δW1 = (z - T ) * (δf2(n2)/δn2)* W2 * (δf1(n1)/δn1) * x W1(k+1) = W1(k) - λ ( (z - T ) * (δf2(n2)/δn2)* W2 * (δf1(n1)/δn1) * x) δf2(n2)/δn2 = z (1 – z) and δf1(n1)/δn1 = a1 ( 1 – a1)

30
Gradient Descent Local minimum Global minimum Error Training time Neural Networks and Logistic Regression by Lucila Ohno-Machado Decision Systems Group, Brigham and Women’s Hospital, Department of Radiology

31
2-D Diagram of Gradient Descent Source : Back Propagation algorithm by Olena Lobunets www.essex.ac.uk/ccfea/Courses/ workshops03-04/Workshop4/Workshop%204.ppt

32
Learning by Example Training Algorithm: backpropagation of errors using gradient descent training. Colors: –Red: Current weights –Orange: Updated weights –Black boxes: Inputs and outputs to a neuron –Blue: Sensitivities at each layer Source : A Brief Overview of Neural Networks Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch campus.umr.edu/smartengineering/ EducationalResources/Neural_Net.ppt

33
First Pass 0.5 1 0.6225 0.6508 Error=1-0.6508=0.3492 G3=(1)(0.3492)=0.3492 G2= (0.6508)(1- 0.6508)(0.3492)(0.5)=0.0397 G1= (0.6225)(1- 0.6225)(0.0397)(0.5)(2)=0.0093 Gradient of the neuron= G =slope of the transfer function ×[Σ{ (weight of the neuron to the next neuron) × ( output of the neuron)}] Gradient of the output neuron = slope of the transfer function × error

34
Weight Update 1 New Weight=Old Weight + {(learning rate)(gradient)(prior output)} 0.5+(0.5)(0.3492)(0.6508) 0.6136 0.5124 0.6136 0.5124 0.5047 0.5+(0.5)(0.0397)(0.6225) 0.5+(0.5)(0.0093)(1) Source : A Brief Overview of Neural Networks Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch campus.umr.edu/smartengineering/ EducationalResources/Neural_Net.ppt

35
Second Pass 0.5047 0.5124 0.6136 0.5047 0.5124 1 0.5047 0.6391 0.6236 0.8033 0.6545 0.8033 Error=1-0.8033=0.1967 G3=(1)(0.1967)=0.1967 G2= (0.6545)(1- 0.6545)(0.1967)(0.6136)=0.0273 G1= (0.6236)(1- 0.6236)(0.5124)(0.0273)(2)=0.0066 Source : A Brief Overview of Neural Networks Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch campus.umr.edu/smartengineering/ EducationalResources/Neural_Net.ppt

36
Weight Update 2 New Weight=Old Weight + {(learning rate)(gradient)(prior output)} 0.6136+(0.5)(0.1967)(0.6545) 0.6779 0.5209 0.6779 0.5209 0.508 0.5124+(0.5)(0.0273)(0.6236) 0.5047+(0.5)(0.0066)(1) Source : A Brief Overview of Neural Networks Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch campus.umr.edu/smartengineering/ EducationalResources/Neural_Net.ppt

37
Third Pass 0.508 0.5209 0.6779 0.508 0.5209 1 0.508 0.6504 0.6243 0.8909 0.6571 0.8909 Source : A Brief Overview of Neural Networks Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch campus.umr.edu/smartengineering/ EducationalResources/Neural_Net.ppt

38
Weight Update Summary W1: Weights from the input to the input layer W2: Weights from the input layer to the hidden layer W3: Weights from the hidden layer to the output layer Source : A Brief Overview of Neural Networks Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch campus.umr.edu/smartengineering/ EducationalResources/Neural_Net.ppt

39
ECG Interpretation Neural Networks and Logistic Regression by Lucila Ohno-Machado Decision Systems Group, Brigham and Women’s Hospital, Department of Radiology

40
Other Applications of ANN Lip Reading Using Artificial Neural Network Ahmad Khoshnevis, Sridhar Lavu, Bahar Sadeghi and Yolanda Tsang ELEC502 Course Project www-dsp.rice.edu/~lavu/research/doc/502lavu.ps AI Techniques in Power Electronics and Drives Dr. Marcelo G. Simões Colorado School of Mines egweb.mines.edu/msimoes/tutorial Car Classification with Neural Networks Koichi Sato & Sangho Park hercules.ece.utexas.edu/course/ ee380l/1999sp/present/carclass.ppt Face Detection and Neural Networks Todd Wittman www.ima.umn.edu/~whitman/faces/face_detection2.ppt A Neural Network for Detecting and Diagnosing Tornadic Circulations V Lakshmanan, Gregory Stumpf, Arthur Witt www.cimms.ou.edu/~lakshman/Papers/mdann_talk.ppt www.cimms.ou.edu/~lakshman/Papers/mdann_talk.ppt

41
Bibliography A Brief Overview of Neural Networks Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch campus.umr.edu/smartengineering/ EducationalResources/Neural_Net.ppt Neural Networks and Logistic Regression Lucila Ohno-Machado Decision Systems Group, Brigham and Women’s Hospital,Department of Radiology dsg.harvard.edu/courses/hst951/ppt/hst951_0320.ppt G5AIAI Introduction to AI by Graham Kendall Schooll of Computer Science and IT, University of Nottingham www.cs.nott.ac.uk/~gxk/courses/g5aiai The Scientist and Engineer's Guide to Digital Signal Processing Steven W. Smith, Ph.D. California Technical Publishing www.dspguide.com Neural Network Design Martin Hagen, Howard B. Demuth, and Mark Beale Campus Publishing Services, Boulder Colorado 80309-0036 ECE 8412 lectures notes by Dr. Anthony Zygmont Department of Electrical Engineering Villanova University January 2003 Supervised Neural Network Introduction CISC 873. Data Mining Yabin Meng meng@cs.queensu.ca

Similar presentations

OK

Computational Intelligence Semester 2 Neural Networks Lecture 2 out of 4.

Computational Intelligence Semester 2 Neural Networks Lecture 2 out of 4.

© 2018 SlidePlayer.com Inc.

All rights reserved.

To make this website work, we log user data and share it with processors. To use this website, you must agree to our Privacy Policy, including cookie policy.

Ads by Google

Ppt on aircraft emergencies videos Ppt on chapter 12 electricity for kids Seminar ppt on brain machine interface Ppt on project report writing Ppt on java virtual machine Ppt on tsunami in 2004 Ppt on gunn diodes Ppt on word association test definition Ppt on electric motor class 10 Animated ppt on magnetism