Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam

Slides:



Advertisements
Similar presentations
Bioinspired Computing Lecture 16
Advertisements

Chapter3 Pattern Association & Associative Memory
Pattern Association.
Computational Intelligence
Introduction to Neural Networks Computing
B.Macukow 1 Lecture 12 Neural Networks. B.Macukow 2 Neural Networks for Matrix Algebra Problems.
Ming-Feng Yeh1 CHAPTER 13 Associative Learning. Ming-Feng Yeh2 Objectives The neural networks, trained in a supervised manner, require a target signal.
1 Neural networks 3. 2 Hopfield network (HN) model A Hopfield network is a form of recurrent artificial neural network invented by John Hopfield in 1982.
Simple Neural Nets For Pattern Classification
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Pattern Association A pattern association learns associations between input patterns and output patterns. One of the most appealing characteristics of.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
18 1 Hopfield Network Hopfield Model 18 3 Equations of Operation n i - input voltage to the ith amplifier a i - output voltage of the ith amplifier.
November 30, 2010Neural Networks Lecture 20: Interpolative Associative Memory 1 Associative Networks Associative networks are able to store a set of patterns.
Neural Networks Lab 5. What Is Neural Networks? Neural networks are composed of simple elements( Neurons) operating in parallel. Neural networks are composed.
December 7, 2010Neural Networks Lecture 21: Hopfield Network Convergence 1 The Hopfield Network The nodes of a Hopfield network can be updated synchronously.
Prelude A pattern of activation in a NN is a vector A set of connection weights between units is a matrix Vectors and matrices have well-understood mathematical.
CS623: Introduction to Computing with Neural Nets (lecture-10) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
Supervised Hebbian Learning
Chapter 6 Associative Models. Introduction Associating patterns which are –similar, –contrary, –in close proximity (spatial), –in close succession (temporal)
© Negnevitsky, Pearson Education, Lecture 7 Artificial neural networks: Supervised learning Introduction, or how the brain works Introduction, or.
Jochen Triesch, UC San Diego, 1 Short-term and Long-term Memory Motivation: very simple circuits can store patterns of.
Associative-Memory Networks Input: Pattern (often noisy/corrupted) Output: Corresponding pattern (complete / relatively noise-free) Process 1.Load input.
IE 585 Introduction to Neural Networks. 2 Modeling Continuum Unarticulated Wisdom Articulated Qualitative Models Theoretic (First Principles) Models Empirical.
10/6/20151 III. Recurrent Neural Networks. 10/6/20152 A. The Hopfield Network.
1 Chapter 6: Artificial Neural Networks Part 2 of 3 (Sections 6.4 – 6.6) Asst. Prof. Dr. Sukanya Pongsuparb Dr. Srisupa Palakvangsa Na Ayudhya Dr. Benjarath.
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
Hebbian Coincidence Learning
Neural Networks and Fuzzy Systems Hopfield Network A feedback neural network has feedback loops from its outputs to its inputs. The presence of such loops.
IE 585 Associative Network. 2 Associative Memory NN Single-layer net in which the weights are determined in such a way that the net can store a set of.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators.
CSC321: Introduction to Neural Networks and machine Learning Lecture 16: Hopfield nets and simulated annealing Geoffrey Hinton.
Modelling Language Evolution Lecture 1: Introduction to Learning Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.
Pencil-and-Paper Neural Networks Prof. Kevin Crisp St. Olaf College.
Ming-Feng Yeh1 CHAPTER 16 AdaptiveResonanceTheory.
Optimization with Neural Networks Presented by: Mahmood Khademi Babak Bashiri Instructor: Dr. Bagheri Sharif University of Technology April 2007.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
13 1 Associative Learning Simple Associative Network.
Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam en Universiteit Utrecht
CS623: Introduction to Computing with Neural Nets (lecture-12) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
Lecture 9 Model of Hopfield
ECE 471/571 - Lecture 16 Hopfield Network 11/03/15.
Neural Network to solve Traveling Salesman Problem Amit Goyal Koustubh Vachhani Ankur Jain 01D05007.
Computational Intelligence Winter Term 2015/16 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
Assocative Neural Networks (Hopfield) Sule Yildirim 01/11/2004.
Lecture 39 Hopfield Network
J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control Artificial Neural Networks II - Outline Cascade Nets and Cascade-Correlation.
Hopfield Networks MacKay - Chapter 42.
CSC321 Lecture 18: Hopfield nets and simulated annealing
Ch7: Hopfield Neural Model
i. Only one word can occur at a given position
FUNDAMENTAL CONCEPT OF ARTIFICIAL NETWORKS
Neural Networks Chapter 4
Adaptive Resonance Theory
Creating Data Representations
Recurrent Networks A recurrent network is characterized by
Adaptive Resonance Theory
Lecture 39 Hopfield Network
Computational Intelligence
Computational Intelligence
CS623: Introduction to Computing with Neural Nets (lecture-11)
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam

Overview Interactive activation model Hopfield networks Constraint satisfaction Attractors Traveling salesman problem Hebb rule and Hopfield networks Bidirectional associative networks Linear associative networks

Much of perception is dealing with ambiguity LAB

Many interpretations are processed in parallel CAB

The final interpretation must satisfy many constraints In the recognition of letters and words: i. Only one word can occur at a given position ii. Only one letter can occur at a given position iii. A letter-on-a-position activates a word iv. A feature-on-a-position activates a letter

Letter detection Press the left button when the letter L is present Press as fast as you can

C##

L##

LAP

L..C...A...P..B LAPCAPCAB

i. Only one word can occur at a given position LAPCAPCAB L..C...A...P..B

ii. Only one letter can occur at a given position L..C...A...P..B LAPCAPCAB

iii. A letter-on-a-position activates a word L..C...A...P..B LAPCAPCAB

iv. A feature-on-a-position activates a letter L..C...A...P..B LAPCAPCAB

L..C...A...P..B LAPCAPCAB Recognition of a letter is a process of constraint satisfaction

L..C...A...P..B LAPCAPCAB Recognition of a letter is a process of constraint satisfaction

L..C...A...P..B LAPCAPCAB Recognition of a letter is a process of constraint satisfaction

L..C...A...P..B LAPCAPCAB Recognition of a letter is a process of constraint satisfaction

L..C...A...P..B LAPCAPCAB Recognition of a letter is a process of constraint satisfaction

Hopfield (1982) Bipolar activations –-1 or 1 Symmetric weights (no self weights) –w ij = w ji Asynchronous update rule –Select one neuron randomly and update it Simple threshold rule for updating

Energy of a Hopfield network Energy E = - ½  i,j w ji a i a j  E = - ½  i (w ji a i + w ij a i )  a j = -  i w ji a i  a j Net input to node j is  i w ji a i = net j Thus, we can write  E = - net j  a j

Given a net input, net j, find a j so that -net j a j is minimized If net j is positive set a j to 1 If net j is negative set a j to -1 If net j is zero, don’t care (leave a j as is) This activation rule ensures that the energy never increases Hence, eventually the energy will reach a minimum value

Attractor An attractor is a stationary network state (configuration of activation values) This is a state where it is not possible to minimize the energy any further by just flipping one activation value It may be possible to reach a deeper attractor by flipping many nodes at once Conclusion: The Hopfield rule does not guarantee that an absolute energy minimum will be reached

Attractor Local minimum Global minimum

Example: 8-Queens problem Place 8 queens on a chess board such that they are not able to take each other This implies the following three constraints: –1 queen per column –1 queen per row –1 queen on any diagonal This encoding of the constraints ensures that the attractors of the network correspond to valid solutions

The constraints are satisfied by inhibitory connections Column Row Diagonals

Problem: how to ensure that exactly 8 nodes are 1? A term may be added to control for this in the activation rule Binary nodes may be used with a bias It is also possible to use continuous valid nodes with Hopfield networks (e.g, between 0 and 1)

Traveling Salesman Problem

The energy minimization question can also be turned around Given a i and a j, how should we set the weight w ji = w ji so that the energy is minimized?  E = - ½   w ji a i a j, so that –when a i a j = 1,  w ji must be positive –when a i a j = -1,  w ji must be negative For example,  w ji =  a i a j, where  is a learning constant

Hebb and Hopfield When used with Hopfield type activation rules, the Hebb learning rule places patterns at attractors If a network has n nodes, 0.15n random patterns can be reliably stored by such a system For complete retrieval it is typically necessary to present the network with over 90% of the original pattern

Bidirectional Associative Memories (BAM, Kosko 1988) Uses binary nodes (0 or 1) Symmetric weights Input and output layer Layers are updated in order, using threshold activation rule Nodes within a layer are updated synchronously

BAM BAM is in fact a Hopfield network with two layers of nodes Within a layer, weights are 0 These neurons are not dependent on each other (no mutual inputs) If updated synchronously, there is therefore no danger of increasing the network energy BAM is similar to the core of Grossberg’s Adaptive Resonance Theory (Lecture 4)

Linear Associative Networks Invented by Kohonen (1972), Nakano (1972), and by Anderson (1972) Two layers Linear activation rule –Activation is equal to net input Can store patterns Their behavior is mathematically tractable using matrix algebra

Associating an input vector p with an output vector q Storage: W =  qp T with  = (p T p) -1 Recall: Wp =  qp T p =  p T pq = q

Inner product p T p gives a scalar p pTpT  = (p T p) -1 = 1/27

Outer product qp T gives a matrix q output vector p T input vector W  weight matrix multiplied by learning constant

Final weight matrix W =  qp T

Recall: Wp = q 0.11      1 =      1 = 2 Weight matrix Input vector Output vector

Storing n patterns Storage: W k =  k q k p k T, with  k = p k T p k W = W 1 + W 2 + … + W k + … + W n Recall: Wp k =  k q k p k T p k + Error = q + Error Error = W 1 p k + … + W h p k + … + W n p k is 0 only if p h T p k = 0 for all h  k

Characteristics of LANs LANs work only well, if the input patterns are (nearly) orthogonal If an input pattern overlaps with others, then recall will be contaminated with the output patterns of those overlapping patterns It is, therefore, important that input patterns are orthogonal (i.e., have little overlap)

LANs have limited representational power For each three-layer LAN, there exists an equivalent two layer LAN Proof: Suppose that q = Wp and r = Vq, than we have r = Vq = VWp = Xp with X = VW p q r W V p r X

Summing up There is a wide variety of ways to store and retrieve patterns in neural networks based on the Hebb rule –Willshaw network (associator) –BAM –LAN –Hopfield network In Hopfield networks, stored patterns can be viewed as attractors

Summing up Finding an attractor is a process of constraint satisfaction. It can can be used as: –A recognition model –A memory retrieval model –A way of solving the traveling salesman problem and other difficult problems