Tom Griffiths CogSci C131/Psych C123 Computational Models of Cognition.

Slides:



Advertisements
Similar presentations
Approaches, Tools, and Applications Islam A. El-Shaarawy Shoubra Faculty of Eng.
Advertisements

Multi-Layer Perceptron (MLP)
An Introduction to Artificial Intelligence Presented by : M. Eftekhari.
AI 授課教師:顏士淨 2013/09/12 1. Part I & Part II 2  Part I Artificial Intelligence 1 Introduction 2 Intelligent Agents Part II Problem Solving 3 Solving Problems.
Probabilistic Models of Cognition Conceptual Foundations Chater, Tenenbaum, & Yuille TICS, 10(7), (2006)
A Brief History of Artificial Intelligence
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
The dynamics of iterated learning Tom Griffiths UC Berkeley with Mike Kalish, Steve Lewandowsky, Simon Kirby, and Mike Dowman.
Bayesian Learning Rong Jin. Outline MAP learning vs. ML learning Minimum description length principle Bayes optimal classifier Bagging.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
CS 357 – Intro to Artificial Intelligence  Learn about AI, search techniques, planning, optimization of choice, logic, Bayesian probability theory, learning,
A Bayesian view of language evolution by iterated learning Tom Griffiths Brown University Mike Kalish University of Louisiana.
Neural Networks Marco Loog.
Introduction to Computational Natural Language Learning Linguistics (Under: Topics in Natural Language Processing ) Computer Science (Under:
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Marakas: Decision Support Systems, 2nd Edition © 2003, Prentice-Hall Chapter Chapter 7: Expert Systems and Artificial Intelligence Decision Support.
Overview and History of Cognitive Science. How do minds work? What would an answer to this question look like? What is a mind? What is intelligence? How.
COMP 3009 Introduction to AI Dr Eleni Mangina
Revealing inductive biases with Bayesian models Tom Griffiths UC Berkeley with Mike Kalish, Brian Christian, and Steve Lewandowsky.
Vision in Man and Machine. STATS 19 SEM Talk 2. Alan L. Yuille. UCLA. Dept. Statistics and Psychology.
Analyzing iterated learning Tom Griffiths Brown University Mike Kalish University of Louisiana.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Models of Generative Grammar Smriti Singh. Generative Grammar  A Generative Grammar is a set of formal rules that can generate an infinite set of sentences.
Artificial Intelligence
The Computational Theory of Mind. COMPUTATION Functions.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
Bayesian approaches to cognitive sciences. Word learning Bayesian property induction Theory-based causal inference.
Ch1 AI: History and Applications Dr. Bernard Chen Ph.D. University of Central Arkansas Spring 2011.
1 AI and Agents CS 171/271 (Chapters 1 and 2) Some text and images in these slides were drawn from Russel & Norvig’s published material.
Modelling Language Evolution Lecture 2: Learning Syntax Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.
Artificial Intelligence Dr. Paul Wagner Department of Computer Science University of Wisconsin – Eau Claire.
Computer Science and Engineering
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
Optimal predictions in everyday cognition Tom Griffiths Josh Tenenbaum Brown University MIT Predicting the future Optimality and Bayesian inference Results.
Chapter 9 Neural Network.
Cognitive Psychology: Thinking, Intelligence, and Language
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Graphical Models in Vision. Alan L. Yuille. UCLA. Dept. Statistics.
Artificial Intelligence By Michelle Witcofsky And Evan Flanagan.
© 2008 The McGraw-Hill Companies, Inc. Chapter 8: Cognition and Language.
Chapter 13 Artificial Intelligence and Expert Systems.
I Robot.
1 Introduction to Artificial Intelligence (Lecture 1)
1 CS 385 Fall 2006 Chapter 1 AI: Early History and Applications.
PSY270 Michaela Porubanova. Language  a system of communication using sounds or symbols that enables us to express our feelings, thoughts, ideas, and.
CS Inductive Bias1 Inductive Bias: How to generalize on novel data.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
Introduction to Artificial Intelligence CS 438 Spring 2008 Today –AIMA, Chapter 1 –Defining AI Next Tuesday –Intelligent Agents –AIMA, Chapter 2 –HW: Problem.
Introduction Chapter 1 Foundations of statistical natural language processing.
Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at.
Epistemology (How do you know something?)  How do you know your science textbook is true?  How about your history textbook?  How about what your parents.
CS1001 Lecture 25. Overview Homework 4 Homework 4 Artificial Intelligence Artificial Intelligence Database Systems Database Systems.
1 ARTIFICIAL INTELLIGENCE Gilles BÉZARD Version 3.16.
1 Artificial Intelligence & Prolog Programming CSL 302.
Artificial Intelligence Hossaini Winter Outline book : Artificial intelligence a modern Approach by Stuart Russell, Peter Norvig. A Practical Guide.
CS-424 Gregory Dudek Lecture 14 Learning –Inductive inference –Probably approximately correct learning.
Chapter 3 Language Acquisition: A Linguistic Treatment Jang, HaYoung Biointelligence Laborotary Seoul National University.
Artificial Intelligence
CHAPTER 5 Handling Uncertainty BIC 3337 EXPERT SYSTEM.
What is cognitive psychology?
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 2, 2017.
Done Done Course Overview What is AI? What are the Major Challenges?
Chapter 11: Learning Introduction
Knowledge Representation
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 8, 2018.
AI and Agents CS 171/271 (Chapters 1 and 2)
Learning linguistic structure with simple recurrent neural networks
Lecture 14 Learning Inductive inference
CO Games Development 2 Week 22 Machine Learning
Presentation transcript:

Tom Griffiths CogSci C131/Psych C123 Computational Models of Cognition

Tom Griffiths CogSci C131/Psych C123 Computational Models of Cognition

ComputationCognition

Cognitive science The study of intelligent systems Cognition as information processing inputoutputinputoutput

computation Computational modeling Look for principles that characterize both computation and cognition inputoutputinputoutput

Two goals Cognition: –explain human cognition (and behavior) in terms of the underlying computation Computation: –gain insight into how to solve some challenging computational problems

Computational problems Easy: –arithmetic, algebra, chess Difficult: –learning and using language –sophisticated senses: vision, hearing –similarity and categorization –representing the structure of the world –scientific investigation human cognition sets the standard

Three approaches Rules and symbols Networks, features, and spaces Probability and statistics

Three approaches Rules and symbols Networks, features, and spaces Probability and statistics

Logic All As are Bs All Bs are Cs All As are Cs Aristotle ( BC)

Mechanical reasoning ( )

The mathematics of reason Thomas Hobbes ( ) Rene Descartes ( ) Gottfried Leibniz ( )

Modern logic George Boole ( ) Friedrich Frege ( ) PQPQPQPQ

Computation Alan Turing ( )

The World Logic P  Q P Q Facts Inference Rules

Categorization

cat  small  furry  domestic  carnivore

The World Logic P  Q P Q Facts Inference Rules

The World Early AI systems… Workspace Operations Actions Observations Facts Goals Operations

Rules and symbols Perhaps we can consider thought a set of rules, applied to symbols… –generating infinite possibilities with finite means –characterizing cognition as a “formal system” This idea was applied to: –deductive reasoning (logic) –language (generative grammar) –problem solving and action (production systems)

Language as a formal system Noam Chomsky

Language “a set (finite or infinite) of sentences, each finite in length and constructed out of a finite set of elements” all sequences L This is a good sentence1 Sentence bad this is0 linguistic analysis aims to separate the grammatical sequences which are sentences of L from the ungrammatical sequences which are not

A context free grammar S  NP VP NP  T N VP  V NP T  the N  man, ball, … V  hit, took, … S NP VP T N V NP T N the man hit the ball

Rules and symbols Perhaps we can consider thought a set of rules, applied to symbols… –generating infinite possibilities with finite means –characterizing cognition as a “formal system” This idea was applied to: –deductive reasoning (logic) –language (generative grammar) –problem solving and action (production systems) Big question: what are the rules of cognition?

Computational problems Easy: –arithmetic, algebra, chess Difficult: –learning and using language –sophisticated senses: vision, hearing –similarity and categorization –representing the structure of the world –scientific investigation human cognition sets the standard

Inductive problems Drawing conclusions that are not fully justified by the available data –e.g. detective work Much more challenging than deduction! “In solving a problem of this sort, the grand thing is to be able to reason backward. That is a very useful accomplishment, and a very easy one, but people do not practice it much.”

Challenges for symbolic approaches Learning systems of rules and symbols is hard! –some people who think of human cognition in these terms end up arguing against learning…

The poverty of the stimulus The rules and principles that constitute the mature system of knowledge of language are actually very complicated There isn’t enough evidence to identify these principles in the data available to children Therefore Acquisition of these rules and principles must be a consequence of the genetically determined structure of the language faculty

The poverty of the stimulus Learning language requires strong constraints on the set of possible languages These constraints are “Universal Grammar”

Challenges for symbolic approaches Learning systems of rules and symbols is hard! –some people who think of human cognition in these terms end up arguing against learning… Many human concepts have fuzzy boundaries –notions of similarity and typicality are hard to reconcile with binary rules Solving inductive problems requires dealing with uncertainty and partial knowledge

Three approaches Rules and symbols Networks, features, and spaces Probability and statistics

Similarity What determines similarity?

Representations What kind of representations are used by the human mind?

Semantic networksSemantic spaces Representations How can we capture the meaning of words?

Categorization

Computing with spaces x 1 x 2 y perceptual features +1 = cat, -1 = dog x1x1 x2x2 y dog cat error:

Networks, features, and spaces Artificial neural networks can represent any continuous function…

Problems with simple networks x 1 x 2 x1x1 x2x2 y Some kinds of data are not linearly separable x1x1 x2x2 AND x1x1 x2x2 OR x1x1 x2x2 XOR

A solution: multiple layers z 1 z 2 y x 1 x 2 y z1z1 z2z2 x1x1 x2x2 hidden layer input layer output layer

Networks, features, and spaces Artificial neural networks can represent any continuous function… Simple algorithms for learning from data –fuzzy boundaries –effects of typicality

E (error) w ij (  is learning rate) General-purpose learning mechanisms

The Delta Rule x 1 x 2 y +1 = cat, -1 = dog output error influence of input for any function g with derivative g perceptual features

Networks, features, and spaces Artificial neural networks can represent any continuous function… Simple algorithms for learning from data –fuzzy boundaries –effects of typicality A way to explain how people could learn things that look like rules and symbols…

Simple recurrent networks z 1 z 2 x 1 x 2 hidden layer input layer output layer context units input (Elman, 1990) x 2 x 1 copy

Hidden unit activations after 6 iterations of 27,500 words (Elman, 1990)

Networks, features, and spaces Artificial neural networks can represent any continuous function… Simple algorithms for learning from data –fuzzy boundaries –effects of typicality A way to explain how people could learn things that look like rules and symbols… Big question: how much of cognition can be explained by the input data?

Challenges for neural networks Being able to learn anything can make it harder to learn specific things –this is the “bias-variance tradeoff”

Bias-variance tradeoff

What about generalization?

What happened? The set of 8th degree polynomials contains almost all functions through 10 points Our data are some true function, plus noise Fitting the noise gives us the wrong function This is called overfitting –while it has low bias, this class of functions results in an algorithm that has high variance (i.e. is strongly affected by the observed data)

The moral General purpose learning mechanisms do not work well with small amounts of data (the most flexible algorithm isn’t always the best) To make good predictions from small amounts of data, you need algorithms with bias that matches the problem being solved This suggests a different approach to studying induction… –(what people learn as n  0, rather than n   )

Challenges for neural networks Being able to learn anything can make it harder to learn specific things –this is the “bias-variance tradeoff” Neural networks allow us to encode constraints on learning in terms of neurons, weights, and architecture, but is this always the right language?

Three approaches Rules and symbols Networks, features, and spaces Probability and statistics

Probability Gerolamo Cardano ( )

Probability Thomas Bayes ( ) Pierre-Simon Laplace ( )

Bayes’ rule Posterior probability LikelihoodPrior probability Sum over space of hypotheses h: hypothesis d: data How rational agents should update their beliefs in the light of data

Cognition as statistical inference Bayes’ theorem tells us how to combine prior knowledge with data –a different language for describing the constraints on human inductive inference

Prior over functions k = 8,  = 5,  = 1 k = 8,  = 5,  = 0.1 k = 8,  = 5,  = 0.3 k = 8,  = 5,  = 0.01

Maximum a posteriori (MAP) estimation

Cognition as statistical inference Bayes’ theorem tells us how to combine prior knowledge with data –a different language for describing the constraints on human inductive inference Probabilistic approaches also help to describe learning

Probabilistic context free grammars S  NP VP1.0 NP  T N0.7 NP  N 0.3 VP  V NP1.0 T  the0.8 T  a0.2 N  man0.5 N  ball0.5 V  hit0.6 V  took0.4 S NP VP 1.0 T N 0.7 V NP 1.0 the 0.8 man 0.5 hit 0.6 the 0.8 ball 0.5 T N 0.7 P(tree) = 1.0  0.7  1.0  0.8  0.5  0.6  0.7  0.8  0.5

Probability and learnability Any probabilistic context free grammar can be learned from a sample from that grammar as the sample size becomes infinite Priors trade off with the amount of data that needs to be seen to believe a hypothesis

Cognition as statistical inference Bayes’ theorem tells us how to combine prior knowledge with data –a language for describing the constraints on human inductive inference Probabilistic approaches also help to describe learning Big question: what do the constraints on human inductive inference look like?

Challenges for probabilistic approaches Computing probabilities is hard… how could brains possibly do that? How well do the “rational” solutions from probability theory describe how people think in everyday life?

Three approaches Rules and symbols Networks, features, and spaces Probability and statistics