How the Brain Learns: Rules and Outcomes Psychology 209 January 28, 2013.

Slides:



Advertisements
Similar presentations
© Negnevitsky, Pearson Education, Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Advertisements

The Primary Visual Cortex
History, Part III Anatomical and neurochemical correlates of neuronal plasticity Two primary goals of learning and memory studies are –i. neural centers.
Chapter 3: Neural Processing and Perception. Lateral Inhibition and Perception Experiments with eye of Limulus –Ommatidia allow recordings from a single.
Activity-Dependent Development I April 23, 2007 Mu-ming Poo 1.Development of OD columns 2.Effects of visual deprivation 3. The critical period 4. Hebb’s.
Un Supervised Learning & Self Organizing Maps. Un Supervised Competitive Learning In Hebbian networks, all neurons can fire at the same time Competitive.
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
Spike Timing-Dependent Plasticity Presented by: Arash Ashari Slides mostly from: 1  Woodin MA, Ganguly K, and Poo MM. Coincident pre-
Learning crossmodal spatial transformations through STDP Gerhard Neumann Seminar B, SS 06.
Spike timing dependent plasticity Homeostatic regulation of synaptic plasticity.
Neural Mechanisms of Memory Storage Molecular, synaptic, and cellular events store information in the nervous system. New learning and memory formation.
Slides are based on Negnevitsky, Pearson Education, Lecture 8 Artificial neural networks: Unsupervised learning n Introduction n Hebbian learning.
1 Activity-Dependent Development Plasticity 1.Development of OD columns 2.Effects of visual deprivation 3. The critical period 4. Hebb’s hypothesis 5.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Adult Cortical Plasticity 1.Maps in somatic sensory and motor cortex 2.Reorganization of cortical maps following sensory deprivation 3.Synaptic basis of.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Synapses are everywhere neurons synapses Synapse change continuously –From msec –To hours (memory) Lack HH type model for the synapse.
The three main phases of neural development 1. Genesis of neurons (and migration). 2. Outgrowth of axons and dendrites, and synaptogenesis. 3. Refinement.
Artificial neural networks.
Module : Development of the Nervous System Lecture 6 Synapse formation & refinement.
How facilitation influences an attractor model of decision making Larissa Albantakis.
Activity-Dependent Development II April 25, 2007 Mu-ming Poo 1.The neurotrophin hypothesis 2.Maps in somatic sensory and motor cortices 3.Development of.
1 Activity-dependent Development (2) Hebb’s hypothesis Hebbian plasticity in visual system Cellular mechanism of Hebbian plasticity.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Critical periods A time period when environmental factors have especially strong influence in a particular behavior. –Language fluency –Birds- Are you.
Lecture 09 Clustering-based Learning
Unsupervised learning
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
THE ROLE OF NEURONS IN PERCEPTION Basic Question How can the messages sent by neurons represent objects in the environment?
Artificial Neural Network Unsupervised Learning
Critical periods in development - “nature” vs. “nurture”
1 Computational Vision CSCI 363, Fall 2012 Lecture 3 Neurons Central Visual Pathways See Reading Assignment on "Assignments page"
Molecular mechanisms of memory. How does the brain achieve Hebbian plasticity? How is the co-activity of presynaptic and postsynaptic cells registered.
synaptic plasticity is the ability of the connection, or synapse, between two neurons to change in strength in response to either use or disuse of transmission.
THE VISUAL SYSTEM: EYE TO CORTEX Outline 1. The Eyes a. Structure b. Accommodation c. Binocular Disparity 2. The Retina a. Structure b. Completion c. Cone.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Its all physical!.  Basic structure of the NS is set before birth  Neurons are however flexible living cells that can grow new connections  The ability.
The primate visual systemHelmuth Radrich, The primate visual system 1.Structure of the eye 2.Neural responses to light 3.Brightness perception.
Lecture 9: Introduction to Neural Networks Refs: Dayan & Abbott, Ch 7 (Gerstner and Kistler, Chs 6, 7) D Amit & N Brunel, Cerebral Cortex 7, (1997)
Unit 4 Psychology Learning: Neural Pathways, Synapse Formation & the Role of Neurotransmitters.
Chapter 3: Neural Processing and Perception. Neural Processing and Perception Neural processing is the interaction of signals in many neurons.
1960s, 1970s, converging evidence from cognitive neuropsychology, psychology, neurobiology support the view of Multiple memory systems, efforts to experimentally.
Version 0.10 (c) 2007 CELEST VISI  N BRIGHTNESS CONTRAST: ADVANCED MODELING CLASSROOM PRESENTATION.
1 Financial Informatics –XVII: Unsupervised Learning 1 Khurshid Ahmad, Professor of Computer Science, Department of Computer Science Trinity College, Dublin-2,
Strong claim: Synaptic plasticity is the only game in town. Weak Claim: Synaptic plasticity is a game in town. Biophysics class: section III The synaptic.
Trends in Biomedical Science Making Memory. The following slides are mostly derived from The Brain from Top to Bottom, an Interactive Website about the.
James L. McClelland Stanford University
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Storage capacity: consider the neocortex ~20*10^9 cells, 20*10^13 synapses.
Sensory Neural Systems 5 February 2008 Rachel L. León
Synaptic Plasticity Synaptic efficacy (strength) is changing with time. Many of these changes are activity-dependent, i.e. the magnitude and direction.
Neurons and neural pathways
Jochen Triesch, UC San Diego, 1 Part 3: Hebbian Learning and the Development of Maps Outline: kinds of plasticity Hebbian.
Learning in the brain and in one-layer neural networks
Simple learning in connectionist networks
Developmental neuroplasticity
Learning in the brain and in one-layer neural networks
Volume 56, Issue 2, Pages (October 2007)
James L. McClelland Center for the Neural Basis of Cognition
Presented by Rhee, Je-Keun
Mind, Brain & Behavior Wednesday February 12, 2003.
Volume 40, Issue 6, Pages (December 2003)
Learning in the brain and in one-layer neural networks
Simple learning in connectionist networks
Synaptic Transmission and Integration
Construction of Complex Receptive Fields in Cat Primary Visual Cortex
Volume 27, Issue 2, Pages (August 2000)
Volume 56, Issue 2, Pages (October 2007)
Presentation transcript:

How the Brain Learns: Rules and Outcomes Psychology 209 January 28, 2013

The brain is highly plastic and changes in response to experience Alteration of experience leads to alterations of neural representations in the brain. What neurons represent, and how precisely they represent it, are strongly affected by experience. We allocate more of our brain to things we have the most experience with.

The Sensory Homunculus

Monkey Somatosensory Cortex

Merzenich’s Joined Finger Experiment Receptive fields after fingers were sown together Control receptive fields

Merzenich’s Rotating Disk Experiment

Merzenich’s Rotating Disk Experiment: Redistribution and Shrinkage of Fields

Merzenich’s Rotating Disk Experiment: Expansion of Sensory Representation

Temporal Sharpening

Synaptic Transmission and Learning Learning may occur by changing the strengths of connections. Addition and deletion of synapses, as well as larger changes in dendritic and axonal arbors, also occur in response to experience. Recent evidence suggests that neurons may be added or deleted in some cases as well. (This occurs in a specialized sub region of the hippocampus, perhaps elsewhere as well neocortex.) Pre Post

Hebb’s Postulate “When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficiency, as one of the cells firing B, is increased.” D. O. Hebb, Organization of Behavior, 1949 In other words: “Cells that fire together wire together.” Unknown Mathematically, this is often taken as: w ba = a b a a (Generally you have to subtract something to achieve stability)

a1a1 a2a2 b

The Molecular Basis of Hebbian Learning (Short Course!) Glutamate ejected from the pre-synaptic terminal activates AMPA receptors, exciting the post-synaptic neuron. Glutamate also binds to the NMDA receptor, but it only opens when the level of depolarization on the post-synaptic side exceeds a threshold. When the NMDA receptor opens, Ca++ flows in, triggering a biochemical cascade that results in an increase in AMPA receptors. The increase in AMPA receptors means that an the same amount of transmitter release at a later time will cause a stronger post- synaptic effect (LTP).

Linsker’s papers (PNAS, 1986a,b,c) Showed how a simple Hebbian learning rule can give rise to: –Center-surround receptive fields at initial stages of visual processing –Oriented edge detectors at later stages Extended the idea to consider why topographic maps may form.

Emergence of Center- Surround Organization Assume a first-layer (A) of independent randomly spiking neurons project to a second layer of neurons that receive inputs according to a Gaussian function of distance from the corresponding point in the layer below. Then neighboring neurons in layer B will have a tendency to have slightly correlated activations, due to the fact that they tend to receive input from some of the same neurons. Now feed activation forward to layer C from layer B. Activation follows a linear activation rule, and learning occurs according to the Hebbian rule at right. Additional assumptions: Weights (w rs ) are bounded in [-1,+1]. Alternatively there could be two types of weights, some in [0,1] and others in [-1,0]. So, when a weight reaches its min or max value, it cannot change further in that direction. d represents a decay term tending to cause an overall negative drift. The result of running this for a long time with small  is center surround organization as shown (small dots = excitatory connections, large dots = inhibitory connections). This comes from the fact that inputs near the center of the receptive field tend to be more correlated with the other inputs to the cell than the ones near the edges.  w rs =  (a r -  r )(a s -  s ) - d a r =  s w rs a s + b r

1 2 3 input units Activation rule: a r =  s a s w rs Initial weights all =.1 Learning Rule: w rs = a r a s – d = 1.0 d =.075 How Hebbian Learning Plus Weight Decay Strengthens Correlated Inputs and Weakens Isolated Inputs to a Receiving Neuron Final weight Units 1 & 2 active Unit 3 active alone unit r This works because inputs correlated with other inputs are associated with stronger activation of the receiving unit that inputs that occur on their own. (2x) ara3ara3

Linsker’s rule maximizes a ‘correlatedness of inputs’ measure What Linsker calls‘Hebb Optimal’ weights are those maximizing the sum of the pair-wise correlations among a unit’s inputs subject to constraints on the sum of the weights:  i  j Q ij w ri w rj Where When, and, the learning rule shown previously tends to maximize the sum of the Q ij ’s subject to a constraint against large weights (enforced by having a non-zero value for d).

When applied to inputs with ‘mexican hat’ receptive fields, Linsker’s rule tends to produce edge detectors! Two representative simulated receptive fields for units at a layer receiving inputs from a layer below consisting of center-surround units are shown. The Mexican Hat receptive field shape implies a Mexican Hat Q function, and this favors each point being like its near- neighbors and un-like its midrange neighbors. “Each point cannot be the center of a ‘like island’ in an ‘unlike sea’”. Whether you get blobs or edge detectors depends on many factors…

Miller, Keller, and Stryker (Science, 1989) model ocular dominance column development using Hebbian learning Architecture: L and R LGN layers and a cortical layer containing 25x25 simple neuron-like units. Each neuron in each LGN has an initial weak projection that forms a Gaussian hump (illustrated with disc) at the corresponding location in the Cortex, but with some noise around it. In the cortex, there are short-range excitatory connections and longer- range inhibitory connections, with the net effect as shown in B (scaled version shown next to cortex to indicate approximate scale).

Simulation of Ocular Dominance Column Development based on Hebbian Learning Experience and Training: Before ‘birth’, random activity occurs in each retina. Due to overlapping projections to LGN, neighboring LGN neurons in the same eye tend to be correlated. No (or less) between-eye correlation is assumed. Learning of weights to cortex from LGN occurs through a Hebbian learning rule: w cl = a c a l – decay (Note that w’s are not allowed to go below 0). Results indicate that ocular dominance columns slowly develop over time. Each panel shows the cortical sheet at a different point in time during development. Strength of R vs L eye input to a given cortical neuron is indicated with light to dark gray as shown). Some studies indicate that ocular dominance columns are present before there is neural activity, but even in that case mechanisms like those considered here may play a role in maintaining and refining the columnar organization.

Competitive Learning and the Self-Organizing Map (Hbk, Ch. 6) Competitive learning: –Units organized into mutually inhibitory clusters –Unit with largest net input is chosen as ‘winner’ –Winner’s weights are updated to align with the input. –Sum or sum of squares of weights coming to the unit are held constant. Self-organizing map: –Very similar but units are arranged in a line, sheet or higher-d space, and: –Weights are updated for the winner and for other units near the winner –Extent of update falls off with distance. –Units then spread out over the input data, and are assigned in proportion to density of data (as in Merzenich experiments). –In simulation result below, units are arranged along a one-d line. Inputs are points in the 2-d space as shown.

Plusses and Minuses of Hebbian Learning Hebbian learning tends to reinforce whatever response occurs to a particular input. –If a neuron becomes activated, the inputs that activate it are strengthened, so the neuron will be even more likely to be activated next time. This may contribute to failures of learning and even to stamping in of bad habits when the response we make to an input is not the best one to make. Possible examples (see articles by me in today’s readings): - Failure of adults to learn new speech sounds - Phobias, racism - Entrenchment of “Habits of Mind”

Learning Depends on More than Mere Stimulation Exposure to task-irrelevant stimuli does not lead to change in studies from the Merzenich lab and other labs. Outcome feedback can lead to enhanced learning. E.g., feedback helps Japanese adults improve their discrimination of /r/ vs /l/ sounds.

How might outcomes shape learning? Learning can be influenced by neuro-modulators such as ACH or dopamine. –One Merzenich expt. showed that an entire cortical area can come to respond to a particular tone played repeatedly, if a sub-cortical nucleus that releases the modulator ACH is continually stimulated while the tone is being presented. ACH may enhance connection weight changes, and level of ACH release may depend on the animal’s attentional state. Reinforcement learning: –Release of the modulator dopamine may be triggered by occurrence of rewards, and dopamine too may modulate the size of connection weight changes. w rs = Ra r a s – d R = reinforcement –This provides a powerful way of shaping Hebb-like learning to strengthen activations that lead to rewards, and not other activations. –Maybe R signal corresponds to level of dopamine (above or below a baseline level). –R may correspond to the extent to which the reward is greater or less than expected (we will return to this idea later). Error-correcting learning: –Learning rules driven by the difference between what a network produces and a ‘teaching signal’ are the main engines of learning in neural networks – we will turn to these next time.