CS 678 –Relaxation and Hopfield Networks1 Relaxation and Hopfield Networks Totally connected recurrent relaxation networks Bidirectional weights (symmetric)

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Chapter3 Pattern Association & Associative Memory
NEURAL NETWORKS Backpropagation Algorithm
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
CS 678 –Boltzmann Machines1 Boltzmann Machine Relaxation net with visible and hidden units Learning algorithm Avoids local minima (and speeds up learning)
Artificial Neural Networks
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
1 Neural networks 3. 2 Hopfield network (HN) model A Hopfield network is a form of recurrent artificial neural network invented by John Hopfield in 1982.
Machine Learning Neural Networks
Simple Neural Nets For Pattern Classification
Recurrent Neural Networks
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Chapter 6: Multilayer Neural Networks
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Radial Basis Function (RBF) Networks
8/10/ RBF NetworksM.W. Mak Radial Basis Function Networks 1. Introduction 2. Finding RBF Parameters 3. Decision Surface of RBF Networks 4. Comparison.
CS623: Introduction to Computing with Neural Nets (lecture-10) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
Computer Science and Engineering
Multi-Layer Perceptrons Michael J. Watts
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
1 Chapter 6: Artificial Neural Networks Part 2 of 3 (Sections 6.4 – 6.6) Asst. Prof. Dr. Sukanya Pongsuparb Dr. Srisupa Palakvangsa Na Ayudhya Dr. Benjarath.
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
Neural Networks - Berrin Yanıkoğlu1 Applications and Examples From Mitchell Chp. 4.
NEURAL NETWORKS FOR DATA MINING
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Artificial Intelligence Techniques Multilayer Perceptrons.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Optimization with Neural Networks Presented by: Mahmood Khademi Babak Bashiri Instructor: Dr. Bagheri Sharif University of Technology April 2007.
Non-Bayes classifiers. Linear discriminants, neural networks.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 32: sigmoid neuron; Feedforward.
Neural Networks - Berrin Yanıkoğlu1 Applications and Examples From Mitchell Chp. 4.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Chapter 18 Connectionist Models
Chapter 6 Neural Network.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Deep Belief Network Training Same greedy layer-wise approach First train lowest RBM (h 0 – h 1 ) using RBM update algorithm (note h 0 is x) Freeze weights.
J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control Artificial Neural Networks II - Outline Cascade Nets and Cascade-Correlation.
Machine Learning Supervised Learning Classification and Regression
Neural networks.
Fall 2004 Backpropagation CS478 - Machine Learning.
Neural Networks.
Fall 2004 Perceptron CS478 - Machine Learning.
Real Neurons Cell structures Cell body Dendrites Axon
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
CSSE463: Image Recognition Day 17
Machine Learning Today: Reading: Maria Florina Balcan
ECE 471/571 - Lecture 17 Back Propagation.
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 17
Boltzmann Machine (BM) (§6.4)
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 17
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

CS 678 –Relaxation and Hopfield Networks1 Relaxation and Hopfield Networks Totally connected recurrent relaxation networks Bidirectional weights (symmetric) Inputs/Outputs? Auto-Associator Nodes represent features of task Weights create an energy landscape When let go relaxes to closest energy minima Hopfield and n-body problems

Hopfield Network Basic node activation – g(H) can also be (-1,1) or the threshold function with -1,1 outputs A "happy" node CS 678 –Relaxation and Hopfield Networks2

Hopfield Network Basic node activation – g(H) can also be (-1,1) or the threshold function with -1,1 outputs A "happy" node – One whose activation and weights correlate with the activations of its neighbors CS 678 –Relaxation and Hopfield Networks3

Hopfield Network Basic node activation – g(H) can also be (-1,1) or the threshold function with -1,1 outputs A "happy" node – One whose activation and weights correlate with the activations of its neighbors A "happy" network – One whose nodes are overall happy CS 678 –Relaxation and Hopfield Networks4

Network Energy Energy is just the negative of "happiness" Appropriate activation correlations, given the weights, lead to low stable energy states CS 678 –Relaxation and Hopfield Networks5

Node Update When do nodes update? Continuously in the "real" model – Slow in software, fast in hardware Typically implemented with "random" update – One node updates at a time but in random order – Synchronous or same order update can lead to oscillating states Processing – How do we initiate node activations – Random – Subset of nodes set to initial state, others random – Subset of nodes clamped, others set to random or initial state Will relax to nearest stable minima – gradient descent CS 678 –Relaxation and Hopfield Networks6

Relaxation What are the stable minima for the following Hopfield Network assuming bipolar states? Each unit is a threshold unit which outputs 1 if net > 0, else -1 CS 678 –Relaxation and Hopfield Networks7

Relaxation What are the stable minima for the following Hopfield Network assuming bipolar states? Each unit is a threshold unit which outputs 1 if net > 0, else -1 CS 678 –Relaxation and Hopfield Networks => => => =>

Relaxation What are the stable minima for the following Hopfield Network assuming bipolar states? Each unit is a threshold unit which outputs 1 if net > 0, else -1 Energy (-1,1,-1) = = -2 (lowest possible for this net) CS 678 –Relaxation and Hopfield Networks => => => =>

Hopfield as a CAM Hopfield can function as a content addressable memory (CAM) – aka Associative Memory For example, could memorize m faces, and then when shown a partial/noisy face, it would return the most similar face Network has nodes equal to the number of bits in the patterns Set the weights according to This just says that for each pattern, increment each weight between bits which have the same values, and decrement the weight between bits which have different values CS 678 –Relaxation and Hopfield Networks10

Hopfield CAM examples What would the weights be set to for an associative memory Hopfield net which is programmed to remember the following patterns? Do you think the net would do well? (i.e. only have stable states are for the learned patterns) 1) ) CS 678 –Relaxation and Hopfield Networks11

Hopfield as an Associative Memory Number of stable storable patterns ≈.15n for random bit patterns For larger percentage of n, many local minima arise There are more efficient ways to build CAMs Limited by weights just set as first order node correlations What would Hopfield weights be for the following? CS 678 –Relaxation and Hopfield Networks12

Execution Approach However, note a fundamentally different approach to execution than, for example, MLPs Rather than find a decision surface for classification, Hopfield returns the nearest memory/stable state Thus, appropriately adjusting and filling in the blanks of an initial state, such that it becomes an "appropriate" final state Real potential for Hopfield/relaxation was demonstrated when used as an optimization engine (rather than just a memory) – TSP – "Good" solutions in fast time CS 678 –Relaxation and Hopfield Networks13

TSP Setup We want shortest cycle with no repeat cites n cities require n 2 nodes 2 n 2 possible binary final states for the network n! legal paths n!/2n distinct legal paths We want a legal path which is also short How do we set weights; energy equation or intuition? CS 678 –Relaxation and Hopfield Networks14

CS 678 –Relaxation and Hopfield Networks15

CS 678 –Relaxation and Hopfield Networks16

CS 678 –Relaxation and Hopfield Networks17

CS 678 –Relaxation and Hopfield Networks18 For n = 30 cities, there are 4.4 × distinct legal paths Typically finds one of the best 10 7 Thus pruning How do we deal with local minima?

Hopfield Summary Highly Parallel, numerous hardware/optic implementations Significant biological plausabilities Not great as an associative memory (saturation issues), but for optimization there are no saturation issues – set weights to fit the problem Stability Issues – Local Minima – Original Hopfield net relaxes to a valid TSP about 20% of time – Our extensions to TSP (MCA and RR as part of Multcons): increase valid tours to 78%, decrease avg valid tour length by 64% Hopfield did not give a learning algorithm – We did BP style training, other efforts Basic Hopfield has no hidden nodes – Boltzmann, Multcons CS 678 –Relaxation and Hopfield Networks19

Multcons Multi-layer temporal constraint satisfaction networks Speech Recognition, Free Market Economics Multi-layer relaxation network Weights are asymmetric (don't observe cycles) Weights are learned – Backpropagation style learning with potential hidden nodes between nodes and layers All nodes/aspects interact before relaxing to a final utterance CS 678 –Relaxation and Hopfield Networks20

CS 678 –Relaxation and Hopfield Networks21 Speech utterance first processed by an MLP to give initial phoneme estimates at different sample times Then network allowed to start relaxing

CS 678 –Relaxation and Hopfield Networks22

CS 678 –Relaxation and Hopfield Networks23

Multcons Notes There are many nodes in the network. However, in software unlikely nodes (low activations) are quickly pruned allowing high efficiency Weights are learned with perceptron style updates or Backpropagation with hidden nodes between layers to learn higher order combinations – Train on an many example utterances. For each utterance, if a node does not relax to the proper target, then adjust its incoming weights to make that node wore likely to hit the target next time (using perceptron or BP learning rules with appropriate learning rates). BP learning update rule is slightly changed due to MCA vs Sigmoid activation function CS 678 –Relaxation and Hopfield Networks24

CS 678 –Relaxation and Hopfield Networks25