Download presentation

1
**Memristor in Learning Neural Networks**

Shaodi Wang Parts of slides from Elham Zamanidoost and Ligang Gao Puneet Gupta

2
Characteristics Ag Pt + - Shaodi Wang

3
Neural Network Shaodi Wang

4
**Learning in Neural Network**

Supervised Learning - Training set contains input and output *Feed-forward network *Recurrent network Unsupervised Learning -Training set contains input only *self-organizing network Shaodi Wang

5
**Multi Layer Perceptron**

Hidden layer(s) perform classification of features Sigmoid activation function Back Propagation Learning: Apply gradient decent over the entire network As before, we have: For every output neuron: For every hidden neuron: Shaodi Wang

6
Gradient Descent Define cost function as sum of errors over entire training set, and errors as: Now train the network in order to minimize the cost. This means that we need to minimize the error. Hence, we need a continuous activation function to calculate the derivative. Sigmoid activation function: *Gradient Descent Learning where Shaodi Wang

7
**Recurrent Network Characteristics:**

- Nodes connect back to other nodes or themselves - Information flow is bidirectional Fully recurrent network: there is a pair of directed connections between every pair of neurons in the network Shaodi Wang

8
**Hopfield Network Characteristics:**

- A RNN in which all connections are symmetric - Binary threshold activation function (CAM) - No unit has a connection with itself and Wi,j =Wj,i (symmetric) - symmetric weights guarantee that the energy function decreases monotonically Hebbian learning: Increase weight between two nodes if both have same activity, otherwise decrease. Synchronous training: the outputs for all the nodes are calculated before applied to the other nodes Asynchronous training: randomly choose a node and calculate its output Shaodi Wang

9
Self Organized Map The purpose of SOM is to map a multidimensional input space onto a topology preserving map of neurons Preserve a topological so that neighboring neurons respond to « similar »input patterns The topological structure is often a 2 or 3 dimensional space Each neuron is assigned a weight vector with the same dimensionality of the input space Input patterns are compared to each weight vector and the closest wins (Euclidean Distance) Shaodi Wang

10
Thanks Shaodi Wang

Similar presentations

OK

Hazırlayan NEURAL NETWORKS Backpropagation Network PROF. DR. YUSUF OYSAL.

Hazırlayan NEURAL NETWORKS Backpropagation Network PROF. DR. YUSUF OYSAL.

© 2018 SlidePlayer.com Inc.

All rights reserved.

To ensure the functioning of the site, we use **cookies**. We share information about your activities on the site with our partners and Google partners: social networks and companies engaged in advertising and web analytics. For more information, see the Privacy Policy and Google Privacy & Terms.
Your consent to our cookies if you continue to use this website.

Ads by Google

Ppt on object-oriented concepts in java with examples Ppt on development of indian railways What does appt only means in spanish Ppt on standing order act 1946 calendar Ppt on keeping it from harold free download Ppt on ms access query Ppt on line drawing algorithms Good ppt on transportation Ppt on front office management in hospital Ppt on ram and rom memory