Introduction to Neural Networks John Paxton Montana State University Summer 2003.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Bioinspired Computing Lecture 16
Chapter3 Pattern Association & Associative Memory
Pattern Association.
Feedback Networks and Associative Memories
Introduction to Neural Networks Computing
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
1 Neural networks 3. 2 Hopfield network (HN) model A Hopfield network is a form of recurrent artificial neural network invented by John Hopfield in 1982.
Neural NetworksNN 11 Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Simple Neural Nets For Pattern Classification
Pattern Association A pattern association learns associations between input patterns and output patterns. One of the most appealing characteristics of.
Correlation Matrix Memory CS/CMPE 333 – Neural Networks.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Before we start ADALINE
November 30, 2010Neural Networks Lecture 20: Interpolative Associative Memory 1 Associative Networks Associative networks are able to store a set of patterns.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Data Mining with Neural Networks (HK: Chapter 7.5)
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
CHAPTER 3 Pattern Association.
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Neural Networks Chapter 2 Joost N. Kok Universiteit Leiden.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
CS532 Neural Networks Dr. Anwar Majid Mirza Lecture No. 3 Week2, January 22 nd, 2008 National University of Computer and Emerging.
CS623: Introduction to Computing with Neural Nets (lecture-10) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
Chapter 6 Associative Models. Introduction Associating patterns which are –similar, –contrary, –in close proximity (spatial), –in close succession (temporal)
© Negnevitsky, Pearson Education, Lecture 7 Artificial neural networks: Supervised learning Introduction, or how the brain works Introduction, or.
Neural Networks. Plan Perceptron  Linear discriminant Associative memories  Hopfield networks  Chaotic networks Multilayer perceptron  Backpropagation.
Associative-Memory Networks Input: Pattern (often noisy/corrupted) Output: Corresponding pattern (complete / relatively noise-free) Process 1.Load input.
-Artificial Neural Network- Hopfield Neural Network(HNN) 朝陽科技大學 資訊管理系 李麗華 教授.
1 Chapter 6: Artificial Neural Networks Part 2 of 3 (Sections 6.4 – 6.6) Asst. Prof. Dr. Sukanya Pongsuparb Dr. Srisupa Palakvangsa Na Ayudhya Dr. Benjarath.
© Negnevitsky, Pearson Education, Lecture 8 (chapter 6) Artificial neural networks: Supervised learning The perceptron Quick Review The perceptron.
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
Hebbian Coincidence Learning
Fuzzy BSB-neuro-model. «Brain-State-in-a-Box Model» (BSB-model) Dynamic of BSB-model: (1) Activation function: (2) 2.
Neural Networks and Fuzzy Systems Hopfield Network A feedback neural network has feedback loops from its outputs to its inputs. The presence of such loops.
IE 585 Associative Network. 2 Associative Memory NN Single-layer net in which the weights are determined in such a way that the net can store a set of.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 8: Neural Networks.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Chapter 2 Single Layer Feedforward Networks
What is Unsupervised Learning? Learning without a teacher. No feedback to indicate the desired outputs. The network must by itself discover the relationship.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Supervised learning network G.Anuradha. Learning objectives The basic networks in supervised learning Perceptron networks better than Hebb rule Single.
Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam en Universiteit Utrecht
Previous Lecture Perceptron W  t+1  W  t  t  d(t) - sign (w(t)  x)] x Adaline W  t+1  W  t  t  d(t) - f(w(t)  x)] f’ x Gradient.
CS623: Introduction to Computing with Neural Nets (lecture-12) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
Lecture 9 Model of Hopfield
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
ECE 471/571 - Lecture 16 Hopfield Network 11/03/15.
Computational Intelligence Winter Term 2015/16 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
Assocative Neural Networks (Hopfield) Sule Yildirim 01/11/2004.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control Artificial Neural Networks II - Outline Cascade Nets and Cascade-Correlation.
Chapter 6 Associative Models
Ch7: Hopfield Neural Model
Chapter 2 Single Layer Feedforward Networks
ECE 471/571 - Lecture 15 Hopfield Network 03/29/17.
ECE 471/571 - Lecture 19 Hopfield Network.
Ch2: Adaline and Madaline
Computational Intelligence
Ch6: AM and BAM 6.1 Introduction AM: Associative Memory
Computational Intelligence
CS623: Introduction to Computing with Neural Nets (lecture-11)
Presentation transcript:

Introduction to Neural Networks John Paxton Montana State University Summer 2003

Chapter 3: Pattern Association Aristotle’s observed that human memory associates –similar items –contrary items –items close in proximity –items close in succession (a song)

Terminology and Issues Autoassociative Networks Heteroassociative Networks Feedforward Networks Recurrent Networks How many patterns can be stored?

Hebb Rule for Pattern Association Architecture x1x1 xnxn ymym y1y1 w 11 w nm

Algorithm 1.set w ij = 0 1 <= i <= n, 1 <= j <= m 2.for each training pair s:t 3.x i = s i 4.y j = t j 5.w ij (new) = w ij (old) + x i y j

Example s 1 = ( ), s 2 = (-1 1 1) t 1 = (1 -1), t 2 = (-1 1) w 11 = 1*1 + (-1)(-1) = 2 w 12 = 1*(-1) + (-1)1 = -2 w 21 = (-1)1+ 1(-1) = -2 w 22 = (-1)(-1) + 1(1) = 2 w 31 = (-1)1 + 1(-1) = -2 w 32 = (-1)(-1) + 1*1 = 2

Matrix Alternative s 1 = ( ), s 2 = (-1 1 1) t 1 = (1 -1), t 2 = (-1 1) =

Final Network f(y in ) = 1 if y in > 0, 0 if y in = 0, else -1 x1x1 x2x2 x3x3 y2y2 y1y

Properties Weights exist if input vectors are linearly independent Orthogonal vectors can be learned perfectly High weights imply strong correlations

Exercises What happens if ( ) is tested? This vector has one mistake. What happens if ( ) is tested? This vector has one piece of missing data. Show an example of training data that is not learnable. Show the learned network.

Delta Rule for Pattern Association Works when patterns are linearly independent but not orthogonal Introduced in the 1960s for ADALINE Produces a least squares solution

Activation Functions Delta Rule (1) w ij (new) = w ij (old) +  (t j – y j )*x i *1 Extended Delta Rule (f’(y in.j )) w ij (new) = w ij (old) +  (t j – y j )*x i *f’(y in.j )

Heteroassociative Memory Net Application: Associate characters. A a B b

Autoassociative Net Architecture x1x1 xnxn ynyn y1y1 w 11 w nn

Training Algorithm Assuming that the training vectors are orthogonal, we can use the Hebb rule algorithm mentioned earlier. Application: Find out whether an input vector is familiar or unfamiliar. For example, voice input as part of a security system.

Autoassociate Example = =

Evaluation What happens if (1 1 1) is presented? What happens if (0 1 1) is presented? What happens if (0 0 1) is presented? What happens if (-1 1 1) is presented? What happens if ( ) is presented? Why are the diagonals set to 0?

Storage Capacity 2 vectors (1 1 1), ( ) Recall is perfect =

Storage Capacity 3 vectors: (1 1 1), ( ), (1 -1 1) Recall is no longer perfect =

Theorem Up to n-1 bipolar vectors of n dimensions can be stored in an autoassociative net.

Iterative Autoassociative Net 1 vector: s = (1 1 -1) s t * s = (1 0 0) -> (0 1 -1) (0 1 -1) -> (2 1 -1) -> (1 1 -1) (1 1 -1) -> (2 2 -2) -> (1 1 -1)

Testing Procedure 1.initialize weights using Hebb learning 2.for each test vector do 3.set x i = s i 4.calculate t i 5.set s i = t i 6. go to step 4 if the s vector is new

Exercises 1 piece of missing data: (0 1 -1) 2 pieces of missing data: (0 0 -1) 3 pieces of missing data: (0 0 0) 1 mistake: ( ) 2 mistakes: ( )

Discrete Hopfield Net content addressable problems pattern association problems constrained optimization problems w ij = w ji w ii = 0

Characteristics Only 1 unit updates its activation at a time Each unit continues to receive the external signal An energy (Lyapunov) function can be found that allows the net to converge, unlike the previous system Autoassociative

Architecture y1y1 y2y2 y3 x 1 x 3 x2x2

Algorithm 1.initialize weights using Hebb rule 2.for each input vector do 3.y i = x i 4.do steps 5-6 randomly for each y i 5.y in.i = x i +  y j w ji 6.calculate f(y in.i ) 7.go to step 2 if the net hasn’t converged

Example training vector: (1 -1) y1y1 y2y2 x 1 x 2

Example input (0 -1) update y 1 = 0 + (-1)(-1) = 1 update y 2 = (-1) = -2 -> -1 input (1 -1) update y 2 = (-1) = -2 -> -1 update y 1 = (-1) = 2 -> 1

Hopfield Theorems Convergence is guaranteed. The number of storable patterns is approximately n / (2 * log n) where n is the dimension of a vector

Bidirectional Associative Memory (BAM) Heteroassociative Recurrent Net Kosko, 1988 Architecture xnxn x1x1 ymym y1y1

Activation Function f(y in ) = 1, if y in > 0 f(y in ) = 0, if y in = 0 f(y in ) = -1 otherwise

Algorithm 1.initialize weights using Hebb rule 2.for each test vector do 3.present s to x layer 4.present t to y layer 5.while equilibrium is not reached 6.compute f(y in.j ) 7.compute f(x in.j )

Example s1 = (1 1), t1 = (1 -1) s2 = (-1 -1), t2 = (-1 1)

Example Architecture x2x2 x1x1 y2y2 y1y present (1 1) to x -> 1 -1 present (1 -1) to y -> 1 1

Hamming Distance Definition: Number of different corresponding bits in two vectors For example, H[(1 -1), (1 1)] = 1 Average Hamming Distance is ½.

About BAMs Observation: Encoding is better when the average Hamming distance of the inputs is similar to the average Hamming distance of the outputs. The memory capacity of a BAM is min(n-1, m-1).