Does the Brain Use Symbols or Distributed Representations? James L. McClelland Department of Psychology and Center for Mind, Brain, and Computation Stanford.

Slides:



Advertisements
Similar presentations
Summer 2011 Tuesday, 8/ No supposition seems to me more natural than that there is no process in the brain correlated with associating or with.
Advertisements

Knowledge Representation
Tuomas Sandholm Carnegie Mellon University Computer Science Department
PDP: Motivation, basic approach. Cognitive psychology or “How the Mind Works”
Machine Learning Neural Networks
Knowledge ß How do we organize our knowledge? ß How do we access our knowledge? ß Do we really use categories?
Emergence in Cognitive Science: Semantic Cognition Jay McClelland Stanford University.
Emergence of Semantic Structure from Experience Jay McClelland Stanford University.
Concepts and Categories. Functions of Concepts By dividing the world into classes of things to decrease the amount of information we need to learn, perceive,
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Chapter Seven The Network Approach: Mind as a Web.
Knowledge information that is gained and retained what someone has acquired and learned organized in some way into our memory.
Categorization  How do we organize our knowledge?  How do we retrieve knowledge when we need it?
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
General Knowledge Dr. Claudia J. Stanny EXP 4507 Memory & Cognition Spring 2009.
Development and Disintegration of Conceptual Knowledge: A Parallel-Distributed Processing Approach Jay McClelland Department of Psychology and Center for.
Dynamics of learning: A case study Jay McClelland Stanford University.
Using Backprop to Understand Apects of Cognitive Development PDP Class Feb 8, 2010.
Representation, Development and Disintegration of Conceptual Knowledge: A Parallel-Distributed Processing Approach James L. McClelland Department of Psychology.
Emergence of Semantic Structure from Experience Jay McClelland Stanford University.
Semantic Memory Psychology Introduction This is our memory for facts about the world This is our memory for facts about the world How do we know.
Conceptual Hierarchies Arise from the Dynamics of Learning and Processing: Insights from a Flat Attractor Network Christopher M. O’ConnorKen McRaeGeorge.
Integrating New Findings into the Complementary Learning Systems Theory of Memory Jay McClelland, Stanford University.
The PDP Approach to Understanding the Mind and Brain J. McClelland Cognitive Core Class Lecture March 7, 2011.
The PDP Approach to Understanding the Mind and Brain Jay McClelland Stanford University January 21, 2014.
Disintegration of Conceptual Knowledge In Semantic Dementia James L. McClelland Department of Psychology and Center for Mind, Brain, and Computation Stanford.
1 Computational Vision CSCI 363, Fall 2012 Lecture 31 Heading Models.
Contrasting Approaches To Semantic Knowledge Representation and Inference Psychology 209 February 15, 2013.
Emergence of Semantic Knowledge from Experience Jay McClelland Stanford University.
PSY 323 – COGNITION Chapter 9: Knowledge.  Categorization ◦ Process by which things are placed into groups  Concept ◦ Mental groupings of similar objects,
The Influence of Feature Type, Feature Structure and Psycholinguistic Parameters on the Naming Performance of Semantic Dementia and Alzheimer’s Patients.
Development, Disintegration, and Neural Basis of Semantic Cognition: A Parallel-Distributed Processing Approach James L. McClelland Department of Psychology.
What is a concept? Part of semantic memory (vs. episodic memory) A class of items that seem to belong together –‘dog’, ‘balloon’, ‘terrorist’ (things)
Emergence of Semantic Structure from Experience Jay McClelland Stanford University.
Similarity and Attribution Contrasting Approaches To Semantic Knowledge Representation and Inference Jay McClelland Stanford University.
What is a concept? Part of semantic memory (vs. episodic memory) A class of items that seem to belong together –‘dog’, ‘balloon’, ‘terrorist’ (things)
Rapid integration of new schema- consistent information in the Complementary Learning Systems Theory Jay McClelland, Stanford University.
Intro. ANN & Fuzzy Systems Lecture 14. MLP (VI): Model Selection.
Semantic Cognition: A Parallel Distributed Processing Approach James L. McClelland Center for the Neural Basis of Cognition and Departments of Psychology.
The PDP Approach to Understanding the Mind and Brain Jay McClelland Stanford University January 21, 2014.
The Origins of Knowledge Debate How do people gain knowledge about the world around them? Are we born with some fundamental knowledge about concepts like.
Origins of Cognitive Abilities Jay McClelland Stanford University.
The Emergent Structure of Semantic Knowledge
Organization of Semantic Memory Typical empirical testing paradigm: propositional verification task – rt to car has four wheels vs. car is a status symbol.
Chapter 6 Neural Network.
Semantic Memory Psychology Introduction This is our memory for facts about the world How do we know that the capital of Viet Nam is Hanoi How is.
Verbal Representation of Knowledge
Emergent Semantics: Meaning and Metaphor Jay McClelland Department of Psychology and Center for Mind, Brain, and Computation Stanford University.
Semantic Knowledge: Its Nature, its Development, and its Neural Basis James L. McClelland Department of Psychology and Center for Mind, Brain, and Computation.
Organization and Emergence of Semantic Knowledge: A Parallel-Distributed Processing Approach James L. McClelland Department of Psychology and Center for.
Development and Disintegration of Conceptual Knowledge: A Parallel-Distributed Processing Approach James L. McClelland Department of Psychology and Center.
Chapter 9 Knowledge. Some Questions to Consider Why is it difficult to decide if a particular object belongs to a particular category, such as “chair,”
Psychology 209 – Winter 2017 January 31, 2017
What is cognitive psychology?
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 2, 2017.
Simple recurrent networks.
Psychology 209 – Winter 2017 Feb 28, 2017
Development and Disintegration of Conceptual Knowledge: A Parallel-Distributed Processing Approach James L. McClelland Department of Psychology and Center.
James L. McClelland SS 100, May 31, 2011
Does the Brain Use Symbols or Distributed Representations?
Emergence of Semantic Structure from Experience
Emergence of Semantics from Experience
Knowledge Pt 2 Chapter 10 Knowledge Pt 2.
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 8, 2018.
Chap. 7 Regularization for Deep Learning (7.8~7.12 )
Learning linguistic structure with simple recurrent neural networks
CLS, Rapid Schema Consistent Learning, and Similarity-weighted Interleaved learning Psychology 209 Feb 26, 2019.
The Network Approach: Mind as a Web
Presentation transcript:

Does the Brain Use Symbols or Distributed Representations? James L. McClelland Department of Psychology and Center for Mind, Brain, and Computation Stanford University

Representation is a pattern of activation distributed over neurons within and across brain areas. Bidirectional propagation of activation underlies the ability to bring these representations to mind from given inputs. The knowledge underlying propagation of activation is in the connections. language Parallel Distributed Processing Approach to Semantic Cognition

Development and Degeneration Learned distributed representations in an appropriately structured distributed connectionist system underlies the development of conceptual knowledge. Gradual degradation of the representations constructed through this developmental process underlies the pattern of semantic disintegration seen in semantic dementia.

Differentiation, ‘Illusory Correlations’, and Overextension of Frequent Names in Development

The Rumelhart Model

The Training Data: All propositions true of items at the bottom level of the tree, e.g.: Robin can {grow, move, fly}

Target output for ‘robin can’ input

ajaj aiai w ij net i =  a j w ij w ki Forward Propagation of Activation

 k ~ (t k -a k ) w ij  i ~   k w ki w ki ajaj Back Propagation of Error () Error-correcting learning: At the output layer:w ki =  k a i At the prior layer: w ij =  j a j … aiai

ExperienceExperience Early Later Later Still

Why Does the Model Show Progressive Differentiation? Learning in the model is sensitive to patterns of coherent covariation of properties Coherent Covariation: –The tendency for properties of objects to co-vary in clusters

Patterns of coherent covariation are reflected in the principal components of the property covariance matrix of the training patterns. Figure shows attribute loadings on the first three principal components: –1. Plants vs. animals –2. Birds vs. fish –3. Trees vs. flowers Same color = features covary in component Diff color = anti-covarying features Patterns of Coherent Covariation in the Training Set

Illusory Correlations Rochel Gelman found that children think that all animals have feet. –Even animals that look like small furry balls and don’t seem to have any feet at all. A tendency to over-generalize properties typical of a superordinate category at an intermediate point in development is characteristic of the PDP network.

A typical property that a particular object lacks e.g., pine has leaves An infrequent, atypical property

PropertyOne-Class Model1 st class in two-class model 2 nd class in two-class model Can Grow1.0 0 Is Living1.0 0 Has Roots Has Leaves Has Branches Has Bark Has Petals Has Gills Has Scales Can Swim Can Fly Has Feathers Has Legs Has Skin Can See A One-Class and a Two-Class Naïve Bayes Classifier Model

Accounting for the network’s representations with classes at different levels of granularity Regression Beta Weight Epochs of Training Living Thing Plant Tree Pine Bias

Overgeneralization of Frequent Names to Similar Objects “dog” “goat” “tree”

Why Does Overgeneralization of Frequent Names Increase and then decrease? In the simulation shown, dogs are experienced 10 times as much as any other animal, and there are 4 other mammals, 8 other animals, and ten plants. In a one-class model, goat is a living thing: –P(name is ‘Dog’|living thing) = 10/32 = ~.3 In a two-class model, goat is an animal: –P(name is ‘Dog’|animal) = 10/22 ~.5 In a five class model, goat is a mammal: –P(name is ‘Dog’|mammal) = 10/15 =.67 In a 23 class model, goat is in a category by itself: –P(name is ‘Dog’|goat) = 0

Sensitivity to Coherence Requires Convergence A A A

Inference and Generalization in the PDP Model A semantic representation for a new item can be derived by error propagation from given information, using knowledge already stored in the weights. Crucially: –The similarity structure, and hence the pattern of generalization depends on the knowledge already stored in the weights.

Start with a neutral representation on the representation units. Use backprop to adjust the representation to minimize the error.

The result is a representation similar to that of the average bird…

Use the representation to infer what this new thing can do.

Differential Importance (Marcario, 1991) 3-4 yr old children see a puppet and are told he likes to eat, or play with, a certain object (e.g., top object at right) –Children then must choose another one that will “be the same kind of thing to eat” or that will be “the same kind of thing to play with”. –In the first case they tend to choose the object with the same color. –In the second case they will tend to choose the object with the same shape.

Adjustments to Training Environment Among the plants: –All trees are large –All flowers are small –Either can be bright or dull Among the animals: –All birds are bright –All fish are dull –Either can be small or large In other words: –Size covaries with properties that differentiate different types of plants –Brightness covaries with properties that differentiate different types of animals

Similarities of Obtained Representations Size is relevant for Plants Brightness is relevant for Animals

Development and Degeneration Sensitivity to coherent covariation in an appropriately structured Parallel Distributed Processing system underlies the development of conceptual knowledge. Gradual degradation of the representations constructed through this developmental process underlies the pattern of semantic disintegration seen in semantic dementia.

Disintegration of Conceptual Knowledge in Semantic Dementia Progressive loss of specific knowledge of concepts, including their names, with preservation of general information Overgeneralization of frequent names Illusory correlations

Picture naming and drawing in Sem. Demantia

Proposed Architecture for the Organization of Semantic Memory color form motion action valance Temporal pole name Medial Temporal Lobe

Rogers et al (2005) model of semantic dementia nameassocfunction temporal pole vision

Severity of DementiaFraction of Neurons Destroyed omissionswithin categ. superord. Patient Data Simulation Results Errors in Naming for As a Function of Severity

Simulation of Delayed Copying Visual input is presented, then removed. After several time steps, pattern is compared to the pattern that was presented initially. nameassocfunction temporal pole vision

Omissions by feature typeIntrusions by feature type IF’s ‘camel’ DC’s ‘swan’ Simulation results

Conclusion Distributed representations gradually differentiate in ways that allow them to capture many phenomena in conceptual development. Their behavior is approximated by a blend of Naïve Bayes classifiers across several levels of granularity, with the blending weights shifting toward finer grain categories as learning progresses. Effects of damage are approximated by a reversal of this tendency: degraded representations retain the coarse-grained level knowledge but loose the finer- grained information. We are currently extending the models to address the sharing of knowledge across structurally related domains, I’ll be glad to discuss this idea in response to questions.