Unsupervised learning

Slides:



Advertisements
Similar presentations
Computational Neuroscience 03 Lecture 8
Advertisements

Factor Analysis Continued
B.Macukow 1 Lecture 12 Neural Networks. B.Macukow 2 Neural Networks for Matrix Algebra Problems.
PCA + SVD.
Chapter 7 Supervised Hebbian Learning.
Linear Transformations
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Self Organization: Hebbian Learning CS/CMPE 333 – Neural Networks.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Synapses are everywhere neurons synapses Synapse change continuously –From msec –To hours (memory) Lack HH type model for the synapse.
Hypercubes and Neural Networks bill wolfe 9/21/2005.
Implementing a reliable neuro-classifier
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
AN INTERACTIVE TOOL FOR THE STOCK MARKET RESEARCH USING RECURSIVE NEURAL NETWORKS Master Thesis Michal Trna
Supervised Hebbian Learning. Hebb’s Postulate “When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part in firing.
6 1 Linear Transformations. 6 2 Hopfield Network Questions.
Artificial neural networks.
Tutorial 10 Iterative Methods and Matrix Norms. 2 In an iterative process, the k+1 step is defined via: Iterative processes Eigenvector decomposition.
PART 5 Supervised Hebbian Learning. Outline Linear Associator The Hebb Rule Pseudoinverse Rule Application.
PCA NETWORK Unsupervised Learning NEtWORKS. PCA is a Representation Network useful for signal, image, video processing.
COMP305. Part I. Artificial neural networks.. Topic 3. Learning Rules of the Artificial Neural Networks.
Techniques for studying correlation and covariance structure
Principal Component Analysis. Philosophy of PCA Introduced by Pearson (1901) and Hotelling (1933) to describe the variation in a set of multivariate data.
Section 8.3 – Systems of Linear Equations - Determinants Using Determinants to Solve Systems of Equations A determinant is a value that is obtained from.
Principle Component Analysis (PCA) Networks (§ 5.8) PCA: a statistical procedure –Reduce dimensionality of input vectors Too many features, some of them.
Unsupervised learning
Deep Learning – Fall 2013 Instructor: Bhiksha Raj Paper: T. D. Sanger, “Optimal Unsupervised Learning in a Single-Layer Linear Feedforward Neural Network”,
The BCM theory of synaptic plasticity.
PSY105 Neural Networks 4/5 4. “Traces in time” Assignment note: you don't need to read the full book to answer the first half of the question. You should.
Artificial Neural Network Yalong Li Some slides are from _24_2011_ann.pdf.
6 1 Linear Transformations. 6 2 Hopfield Network Questions The network output is repeatedly multiplied by the weight matrix W. What is the effect of this.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
HEBB’S THEORY The implications of his theory, and their application to Artificial Life.
Neural Networks and Fuzzy Systems Hopfield Network A feedback neural network has feedback loops from its outputs to its inputs. The presence of such loops.
7 1 Supervised Hebbian Learning. 7 2 Hebb’s Postulate “When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -
The BCM theory of synaptic plasticity. The BCM theory of cortical plasticity BCM stands for Bienestock Cooper and Munro, it dates back to It was.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent (If f(x) is more complex we usually cannot.
Adaptive Algorithms for PCA PART – II. Oja’s rule is the basic learning rule for PCA and extracts the first principal component Deflation procedure can.
Unsupervised Learning Motivation: Given a set of training examples with no teacher or critic, why do we learn? Feature extraction Data compression Signal.
1 Financial Informatics –XVII: Unsupervised Learning 1 Khurshid Ahmad, Professor of Computer Science, Department of Computer Science Trinity College, Dublin-2,
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
Fast Learning in Networks of Locally-Tuned Processing Units John Moody and Christian J. Darken Yale Computer Science Neural Computation 1, (1989)
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Adaptive Cooperative Systems Chapter 8 Synaptic Plasticity 8.11 ~ 8.13 Summary by Byoung-Hee Kim Biointelligence Lab School of Computer Sci. & Eng. Seoul.
0 Chapter 4: Associators and synaptic plasticity Fundamentals of Computational Neuroscience Dec 09.
Jochen Triesch, UC San Diego, 1 Part 3: Hebbian Learning and the Development of Maps Outline: kinds of plasticity Hebbian.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
CSE 554 Lecture 8: Alignment
Principle Component Analysis (PCA) Networks (§ 5.8)
Real Neurons Cell structures Cell body Dendrites Axon
A Simple Artificial Neuron
Unsupervised Learning: Principle Component Analysis
Simple learning in connectionist networks
Degree and Eigenvector Centrality
Financial Informatics –XVII: Unsupervised Learning
Hebb and Perceptron.
A principled way to principal components analysis
Bayesian belief networks 2. PCA and ICA
Techniques for studying correlation and covariance structure
Feature space tansformation methods
X.2 Linear Discriminant Analysis: 2-Class
Parametric Methods Berlin Chen, 2005 References:
Simple learning in connectionist networks
Introduction to Neural Network
Supervised Hebbian Learning
An extended Hebbian model of unsupervised learning
Presentation transcript:

Unsupervised learning The Hebb rule – Neurons that fire together wire together. PCA RF development with PCA

Classical Conditioning The fundamental question that draws most of my scientific attention is how experience shapes the brain. During this talks I will speak about two central concepts in Neuroscience receptive field plasticity and synaptic plasticity. Synaptic plasticity is the change in synaptic efficacy that occurs due to pre and postsynaptic activity. One of the first people to suggest that synaptic plasticity serves as the basis for learning is the Canadian Psychologist Donald Hebb who said Is his famous book in 1949 the following sentence … In this slide I will try to illustrate with a simple example how synaptic plasticity is connected to behavior. Today there is strong evidence for the existence of synaptic plasticity and it is considered the major candidate for The basis of learning memory and some aspects of development. Therefore it is very important to know What are the rules that govern synaptic plasticity. The central theme in my work and throughout this talk is to Find out what these rules are. I think it is imperative that such work would involve a combination of theoretical and Experimental work.

Classical Conditioning and Hebb’s rule Ear A Nose B Tongue The fundamental question that draws most of my scientific attention is how experience shapes the brain. During this talks I will speak about two central concepts in Neuroscience receptive field plasticity and synaptic plasticity. Synaptic plasticity is the change in synaptic efficacy that occurs due to pre and postsynaptic activity. One of the first people to suggest that synaptic plasticity serves as the basis for learning is the Canadian Psychologist Donald Hebb who said Is his famous book in 1949 the following sentence … In this slide I will try to illustrate with a simple example how synaptic plasticity is connected to behavior. Today there is strong evidence for the existence of synaptic plasticity and it is considered the major candidate for The basis of learning memory and some aspects of development. Therefore it is very important to know What are the rules that govern synaptic plasticity. The central theme in my work and throughout this talk is to Find out what these rules are. I think it is imperative that such work would involve a combination of theoretical and Experimental work. “When an axon in cell A is near enough to excite cell B and repeatedly and persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficacy in firing B is increased” D. O. Hebb (1949)

The generalized Hebb rule: where xi are the inputs and y the output is assumed linear: Results in 2D

Example of Hebb in 2D w (Note: here inputs have a mean of zero)

On the board: Solve simple linear first order ODE Fixed points and their stability for non linear ODE. Describe potential and gradient descent

In the simplest case, the change in synaptic weight w is: where x are input vectors and y is the neural response. Assume for simplicity a linear neuron: So we get: Now take an average with respect to the distribution of inputs, get:

(in matrix notation) If <x>=0 , Q is the covariance function. If a small change Δw occurs over a short time Δt then: (in matrix notation) If <x>=0 , Q is the covariance function. What is then the solution of this simple first order linear ODE ? (Show on board)

Mathematics of the generalized Hebb rule The change in synaptic weight w is: where x are input vectors and y is the neural response. Assume for simplicity a linear neuron: So we get:

Taking an average of the distribution of inputs And using and We obtain

In matrix form Where J is a matrix of ones, e is a vector in direction (1,1,1 … 1), and or Where

The equation therefore has the form If k1 is not zero, this has a fixed point, however it is usually not stable. If k1=0 then have:

The Hebb rule is unstable – how can it be stabilized while preserving its properties? The stabilized Hebb (Oja) rule. Where: Appoximate to first order in η: Now insert Get: Normalize

} y Therefore The Oja rule therefore has the form:

Average In matrix form:

Using this rule the weight vector converges to the eigen-vector of Q with the highest eigen-value. It is often called a principal component or PCA rule. The exact dynamics of the Oja rule have been solved by Wyatt and Elfaldel 1995 Variants of networks that extract several principal components have been proposed (e.g: Sanger 1989)

Therefore a stabilized Hebb (Oja neuron) carries out Eigen-vector, or principal component analysis (PCA).

Using this rule the weight vector converges to the eigen-vector of Q with the highest eigen-value. It is often called a principal component or PCA rule. Another way to look at this: Where for the Oja rule: At the f.p: where So the f.p is an eigen-vector of Q. The condition means that w is normalized. Why? Could there be other choices for β?

Show that the Oja rule converges to the state |w^2|=1 The Oja rule in matrix form: Multiply by w, get Bonus question for H.W: The equivalence above That the direction of the PC is the direction of maximal variance (HKP – pg 202)

Show that the f.p of the Oja rule is such that the largest eigen-vector with the largest eigen-value (PC) is stable while others are not (from HKP – pg 202). Start with: Assume w=ua+εub where ua and ub are eigen-vectors with eigen-values λa,b

Get: (show on board) Therefore: That is stable only when λa> λb for every b.