1 Pertemuan 18 HOPFIELD DESIGN Matakuliah: H0434/Jaringan Syaraf Tiruan Tahun: 2005 Versi: 1.

Slides:



Advertisements
Similar presentations
1 Pertemuan > > Matakuliah: >/ > Tahun: > Versi: >
Advertisements

Pertemuan 03 Teori Peluang (Probabilitas)
Teknik Enkoding Pertemuan 08 Matakuliah: H0484/Jaringan Komputer Tahun: 2007.
1 Pertemuan 12 Binary Search Tree Matakuliah: T0026/Struktur Data Tahun: 2005 Versi: 1/1.
1 Pertemuan 22 Radix Sort Matakuliah: T0016/Algoritma dan Pemrograman Tahun: 2005 Versi: versi 2.
1 Pertemuan 15 The Business Owner’s View Matakuliah: A0194/Pengendalian Rekayasa Ulang Informasi Tahun: 2005 Versi: 1/5.
1 Pertemuan 6 Elementary UDP Sockets Matakuliah: H0483 / Network Programming Tahun: 2005 Versi: 1.0.
An Illustrative Example.
Linear Transformations
1 Pertemuan 10 Arsitektur Jaringan Model OSI Matakuliah: H0174/Jaringan Komputer Tahun: 2006 Versi: 1/0.
Performance Optimization
1 Pertemuan 26 Object Relational Database Management System (Lanjutan) Matakuliah: M0174/OBJECT ORIENTED DATABASE Tahun: 2005 Versi: 1/0.
1 Pertemuan 16 Object Query Language (Lanjutan bagian 3) Matakuliah: M0174/OBJECT ORIENTED DATABASE Tahun: 2005 Versi: 1/0.
1 Pertemuan 23 Object database design (Lanjutan bagian 2) Matakuliah: M0174/OBJECT ORIENTED DATABASE Tahun: 2005 Versi: 1/0.
1 Pertemuan > > Matakuliah: >/ > Tahun: > Versi: >
1 Pertemuan 07 Pendugaan Parameter Matakuliah: I0262 – Statistik Probabilitas Tahun: 2007 Versi: Revisi.
1 Pertemuan 13 BACK PROPAGATION Matakuliah: H0434/Jaringan Syaraf Tiruan Tahun: 2005 Versi: 1.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
1 Pertemuan 09 Design and Production Matakuliah: T0553/Sistem Multimedia Tahun: 2005 Versi: 5.
1 Pertemuan 23 Managing The Effectiveness of The Audit Department Matakuliah:A0274/Pengelolaan Fungsi Audit Sistem Informasi Tahun: 2005 Versi: 1/1.
1 Pertemuan 15 ADAPTIVE RESONANCE THEORY Matakuliah: H0434/Jaringan Syaraf Tiruan Tahun: 2005 Versi: 1.
1 Pertemuan 02 LAN Matakuliah: H0451/Praktikum Jaringan Komputer Tahun: 2006 Versi: 1/0.
IP Addressing Pertemuan 18 Matakuliah: H0484/Jaringan Komputer Tahun: 2007.
1 Pertemuan 14 Object Query Language (Lanjutan bagian 1) Matakuliah: M0174/OBJECT ORIENTED DATABASE Tahun: 2005 Versi: 1/0.
1 Pertemuan 5 The structure part of object data model Matakuliah: M0174/OBJECT ORIENTED DATABASE Tahun: 2005 Versi: 1/0.
1 Pertemuan 9 Making an outline Matakuliah: G1072 – Reading 1 Tahun: 2005 Versi: revisi 0.
1 Pertemuan 17 Seleksi DBMS Matakuliah: >/ > Tahun: > Versi: >
1 Pertemuan 21 Audit Reporting Matakuliah:A0274/Pengelolaan Fungsi Audit Sistem Informasi Tahun: 2005 Versi: 1/1.
1 Pertemuan 11 QUIZ Matakuliah: J0274/Akuntansi Manajemen Tahun: 2005 Versi: 01/00.
11 - Innovation Matakuliah: G0622/Bahasa Inggris 1 Tahun: 2005 Versi: 1.01.
1 Pertemuan 20 Time & Condition Clauses with Future reference Matakuliah: G0134 – Grammar III Tahun: 2005 Versi: revisi 1.
1 Pertemuan 18 I wish, If only Matakuliah: G0134 – Grammar III Tahun: 2005 Versi: revisi 1.
1 Pertemuan 7 The Object Definition Language Matakuliah: M0174/OBJECT ORIENTED DATABASE Tahun: 2005 Versi: 1/0.
1 Pertemuan 17 HOPFIELD NETWORK Matakuliah: H0434/Jaringan Syaraf Tiruan Tahun: 2005 Versi: 1.
18 1 Hopfield Network Hopfield Model 18 3 Equations of Operation n i - input voltage to the ith amplifier a i - output voltage of the ith amplifier.
6 1 Linear Transformations. 6 2 Hopfield Network Questions.
1 Pertemuan #3 Clocks and Realtime Matakuliah: H0232/Sistem Waktu Nyata Tahun: 2005 Versi: 1/5.
12 1 Variations on Backpropagation Variations Heuristic Modifications –Momentum –Variable Learning Rate Standard Numerical Optimization –Conjugate.
1 Pertemuan 9 JARINGAN LEARNING VECTOR QUANTIZATION Matakuliah: H0434/Jaringan Syaraf Tiruan Tahun: 2005 Versi: 1.
1 Pertemuan 25 Making It Happen Matakuliah: A0194/Pengendalian Rekayasa Ulang Informasi Tahun: 2005 Versi: 1/5.
1 Pertemuan 8 The Object Definition Language (Lanjutan) Matakuliah: M0174/OBJECT ORIENTED DATABASE Tahun: 2005 Versi: 1/0.
9 1 Performance Optimization. 9 2 Basic Optimization Algorithm p k - Search Direction  k - Learning Rate or.
6 1 Linear Transformations. 6 2 Hopfield Network Questions The network output is repeatedly multiplied by the weight matrix W. What is the effect of this.
7 1 Supervised Hebbian Learning. 7 2 Hebb’s Postulate “When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part.
Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam
1 Pertemuan 26 Making It Happen Matakuliah: A0194/Pengendalian Rekayasa Ulang Informasi Tahun: 2005 Versi: 1/5.
1 Pertemuan 22 Implementasi Sistem File Matakuliah: T0316/sistem Operasi Tahun: 2005 Versi/Revisi: 5.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Variations on Backpropagation.
5.1 Eigenvectors and Eigenvalues 5. Eigenvalues and Eigenvectors.
Signal & Weight Vector Spaces
Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam en Universiteit Utrecht
1 Pertemuan 16 The Business Owner’s View Matakuliah: A0194/Pengendalian Rekayasa Ulang Informasi Tahun: 2005 Versi: 1/5.
5 5.1 © 2016 Pearson Education, Ltd. Eigenvalues and Eigenvectors EIGENVECTORS AND EIGENVALUES.
Reduced echelon form Matrix equations Null space Range Determinant Invertibility Similar matrices Eigenvalues Eigenvectors Diagonabilty Power.
IP Addressing Pertemuan 06
Pertemuan 20 The Business Views of the Technology Architecture
Table Pertemuan 10 Matakuliah : L0182 / Web & Animation Design
Pertemuan 13 Perencanaan Perdagangan Saham
Pertemuan 22 The Business Views of the Technology Architecture
Pertemuan <<6>> Tempat Penyimpanan Data dan Indeks
Pertemuan 7 JARINGAN INSTAR DAN OUTSTAR
Matakuliah : Web Design
Variations on Backpropagation.
Hopfield Network.
Linear Transformations
Variations on Backpropagation.
Section 3: Second Order Methods
Linear Transformations
Supervised Hebbian Learning
Presentation transcript:

1 Pertemuan 18 HOPFIELD DESIGN Matakuliah: H0434/Jaringan Syaraf Tiruan Tahun: 2005 Versi: 1

2 Learning Outcomes Pada akhir pertemuan ini, diharapkan mahasiswa akan mampu : Menunjukkan cara menentukan bobot pada Jaringan Hopfield

3 Outline Materi Content addressable memory. Hebb Rule.

4 Hopfield Design Choose the weight matrix W and the bias vector b so that V takes on the form of a function you want to minimize. The Hopfield network will minimize the following Lyapunov function:

5 Content-Addressable Memory Content-Addressable Memory - retrieves stored memories on the basis of part of the contents. Prototype Patterns: Proposed Performance Index: (bipolar vectors) For orthogonal prototypes, if we evaluate the performance index at a prototype: J(a) will be largest when a is not close to any prototype pattern, and smallest when a is equal to a prototype pattern.

6 Hebb Rule If we use the supervised Hebb rule to compute the weight matrix: the Lyapunov function will be: This can be rewritten: Therefore the Lyapunov function is equal to our performance index for the content addressable memory.

7 Hebb Rule Analysis If we apply prototype p j to the network: Therefore each prototype is an eigenvector, and they have a common eigenvalue of S. The eigenspace for the eigenvalue = S is therefore: An S-dimensional space of all vectors which can be written as linear combinations of the prototype vectors.

8 Weight Matrix Eigenspace The entire input space can be divided into two disjoint sets: where X  is the orthogonal complement of X. For vectors a in the orthogonal complement we have: Therefore, The eigenvalues of W are S and 0, with corresponding eigenspaces of X and X . For the Hessian matrix the eigenvalues are -S and 0, with the same eigenspaces.

9 Lyapunov Surface The high-gain Lyapunov function is a quadratic function. Therefore, the eigenvalues of the Hessian matrix determine its shape. Because the first eigenvalue is negative, V will have negative curvature in X. Because the second eigenvalue is zero, V will have zero curvature in X . Because V has negative curvature in X, the trajectories of the Hopfield network will tend to fall into the corners of the hypercube {a: -1 < a i < 1} that are contained in X.

10 Example

11 Zero Diagonal Elements We can zero the diagonal elements of the weight matrix: The prototypes remain eigenvectors of this new matrix, but the corresponding eigenvalue is now (S-Q): The elements of X  also remain eigenvectors of this new matrix, with a corresponding eigenvalue of (-Q): The Lyapunov surface will have negative curvature in X and positive curvature in X , in contrast with the original Lyapunov function, which had negative curvature in X and zero curvature in X .

12 Example If the initial condition falls exactly on the line a 1 = -a 2, and the weight matrix W is used, then the network output will remain constant. If the initial condition falls exactly on the line a 1 = -a 2, and the weight matrix W’ is used, then the network output will converge to the saddle point at the origin.