Can NLP Systems be a Cognitive Black Box? March 2006 Jerry T. Ball Senior Research Psychologist Air Force Research Laboratory Mesa, AZ.

Slides:



Advertisements
Similar presentations
Artificial Intelligence
Advertisements

Cognitive Systems, ICANN panel, Q1 What is machine intelligence, as beyond pattern matching, classification and prediction. What is machine intelligence,
ARCHITECTURES FOR ARTIFICIAL INTELLIGENCE SYSTEMS
Chapter 2: Marr’s theory of vision. Cognitive Science  José Luis Bermúdez / Cambridge University Press 2010 Overview Introduce Marr’s distinction between.
Statistical Methods and Linguistics - Steven Abney Thur. POSTECH Computer Science NLP Lab Shim Jun-Hyuk.
INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING NLP-AI IIIT-Hyderabad CIIL, Mysore ICON DECEMBER, 2003.
A Brief History of Artificial Intelligence
1 OSCAR: An Architecture for Generally Intelligent Agents John L. Pollock Philosophy and Cognitive Science University of Arizona
Testing 09. Problems and Directions Past experience with language help us come to the conclusion that we should keep a balance between linguistic and.
CORNELL UNIVERSITY CS 764 Seminar in Computer Vision Ramin Zabih Fall 1998.
Tom Griffiths CogSci C131/Psych C123 Computational Models of Cognition.
Topic: Theoretical Bases for Cognitive Method Objectives Trainees will be able to give reasons for the design and procedures of the Cognitive Method.
Symbolic Encoding of Neural Networks using Communicating Automata with Applications to Verification of Neural Network Based Controllers* Li Su, Howard.
The Importance of Architecture for Achieving Human-level AI John Laird University of Michigan June 17, th Soar Workshop
Overview and History of Cognitive Science. How do minds work? What would an answer to this question look like? What is a mind? What is intelligence? How.
PSY 369: Psycholinguistics Some basic linguistic theory part3.
Polyscheme John Laird February 21, Major Observations Polyscheme is a FRAMEWORK not an architecture – Explicitly does not commit to specific primitives.
CPSC 322 Introduction to Artificial Intelligence September 10, 2004.
Science and Engineering Practices
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Cognitive Robots © 2014, SNU CSE Biointelligence Lab.,
THEORIES OF MIND: AN INTRODUCTION TO COGNITIVE SCIENCE Jay Friedenberg and Gordon Silverman.
Essay Writing in Philosophy
Communicative Language Teaching (CLT)
1 7-Speech Recognition (Cont’d) HMM Calculating Approaches Neural Components Three Basic HMM Problems Viterbi Algorithm State Duration Modeling Training.
Yvonne M. Hansen Visualization for Thinking, Planning, and Problem Solving Simple, graphic shapes, the building blocks of a graphical language, play an.
C. 2008, Pearson Allyn & Bacon Introduction to Cognition Chapter 1.
Artificial Intelligence CIS 479/579 Bruce R. Maxim UM-Dearborn.
Chapter 14: Artificial Intelligence Invitation to Computer Science, C++ Version, Third Edition.
计算机科学概述 Introduction to Computer Science 陆嘉恒 中国人民大学 信息学院
The Newell Test for a Theory of Mind Anderson, John R., & Lebiere, Christian (forthcoming), “The Newell Test for a Theory of Mind”, Behavioral & Brain.
Higher-Level Cognitive Processes
Guide to Simulation Run Graphic: The simulation runs show ME (memory element) activation, production matching and production firing during activation of.
Design Science Method By Temtim Assefa.
TEA Science Workshop #3 October 1, 2012 Kim Lott Utah State University.
7-Speech Recognition Speech Recognition Concepts
Ecological Interface Design
Artificial Intelligence Introductory Lecture Jennifer J. Burg Department of Mathematics and Computer Science.
Cognitive Systems Foresight Language and Speech. Cognitive Systems Foresight Language and Speech How does the human system organise itself, as a neuro-biological.
Cognitive Psychology: Thinking, Intelligence, and Language
Advanced Decision Architectures Collaborative Technology Alliance A Computational Model of Naturalistic Decision Making and the Science of Simulation Walter.
An Instructable Connectionist/Control Architecture: Using Rule-Based Instructions to Accomplish Connectionist Learning in a Human Time Scale Presented.
Introduction to CL & NLP CMSC April 1, 2003.
Major objective of this course is: Design and analysis of modern algorithms Different variants Accuracy Efficiency Comparing efficiencies Motivation thinking.
The Two (Computational) Faces of AI David Davenport Computer Engineering Dept., Bilkent University Ankara – TURKEY PT-AI.
인공지능 연구실 황명진 FSNLP Introduction. 2 The beginning Linguistic science 의 4 부분 –Cognitive side of how human acquire, produce, and understand.
How Solvable Is Intelligence? A brief introduction to AI Dr. Richard Fox Department of Computer Science Northern Kentucky University.
The Next Generation Science Standards: 4. Science and Engineering Practices Professor Michael Wysession Department of Earth and Planetary Sciences Washington.
University of Windsor School of Computer Science Topics in Artificial Intelligence Fall 2008 Sept 11, 2008.
Human Factors In Visualization Research Melanie Tory and Torsten Moller Ajith Radhakrishnan Nandu C Nair.
ECE450 - Software Engineering II1 ECE450 – Software Engineering II Today: Introduction to Software Architecture.
Artificial intelligence
Human Abilities 2 How do people think? 1. Agenda Memory Cognitive Processes – Implications Recap 2.
Introduction to Artificial Intelligence CS 438 Spring 2008 Today –AIMA, Ch. 25 –Robotics Thursday –Robotics continued Home Work due next Tuesday –Ch. 13:
1 Viewing Vision-Language Integration as a Double-Grounding case Katerina Pastra Department of Computer Science, Natural Language Processing Group, University.
Introduction Chapter 1 Foundations of statistical natural language processing.
A Roadmap towards Machine Intelligence
RULES Patty Nordstrom Hien Nguyen. "Cognitive Skills are Realized by Production Rules"
What It Is To Be Conscious: Exploring the Plausibility of Consciousness in Deep Learning Computers Senior Project – Philosophy and Computer Science ID.
Cognitive Architectures and General Intelligent Systems Pay Langley 2006 Presentation : Suwang Jang.
Chapter 15. Cognitive Adequacy in Brain- Like Intelligence in Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans Cinarel, Ceyda.
COURSE AND SYLLABUS DESIGN
1 7-Speech Recognition Speech Recognition Concepts Speech Recognition Approaches Recognition Theories Bayse Rule Simple Language Model P(A|W) Network Types.
Cognitive Modeling Cogs 4961, Cogs 6967 Psyc 4510 CSCI 4960 Mike Schoelles
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 2, 2017.
Cognitive Language Processing for Rosie
Theories of Language Development
Introduction Artificial Intelligent.
Learning linguistic structure with simple recurrent neural networks
Toward a Great Class Project: Discussion of Stoianov & Zorzi’s Numerosity Model Psych 209 – 2019 Feb 14, 2019.
Presentation transcript:

Can NLP Systems be a Cognitive Black Box? March 2006 Jerry T. Ball Senior Research Psychologist Air Force Research Laboratory Mesa, AZ

2 Can NLP Systems be a Cognitive Black Box? Is Cognitive Science Relevant to AI Hard Problems? Is it sufficient to model input/output behavior using computational techniques which bear little resemblance to how humans process language? Is it necessary to model the internals of human language processing behavior in NLP systems?

3 Can NLP Systems be a Cognitive Black Box? Yes Is Cognitive Science Relevant to AI Hard Problems? Is it sufficient to model input/output behavior using computational techniques which bear little resemblance to how humans process language? Is it necessary to model the internals of human language processing behavior in NLP systems?

4 Can NLP Systems be a Cognitive Black Box? Yes No Is Cognitive Science Relevant to AI Hard Problems? Is it sufficient to model input/output behavior using computational techniques which bear little resemblance to how humans process language? Is it necessary to model the internals of human language processing behavior in NLP systems?

5 Can NLP Systems be a Cognitive Black Box? Yes No Yes Is Cognitive Science Relevant to AI Hard Problems? Is it sufficient to model input/output behavior using computational techniques which bear little resemblance to how humans process language? Is it necessary to model the internals of human language processing behavior in NLP systems? – To some level of abstraction below input/output behavior

6 Theoretical Arguments for Looking Inside the Black Box

7 Green’s Argument that AI is not THE Right Method for Cognitive Science Green, C. (2000). Is AI the Right Method for Cognitive Science? Psycholoquy 11 (061) Fodor’s “Disneyland” view of AI – AI is to psychology as Disneyland is to physics Denying the importance of looking inside the black box harks back to behaviorism But we don’t know enough about psychology to be able to program psychologically well-founded systems Need to get our cognitive ontology straightened out before we are likely to make much progress – What is a thought? AI is the attempt to simulate something that is not, at present, at all well understood

8 Green Backtracks Considerably Vicente, K. & Burns, C. (2000). Overcoming the Conceptual Muddle: a Little Help from Systems Theory. Psycoloquy 11 (070). In conclusion, AI may provide a rich source of models and techniques, but until these models are tested against psychological evidence and under realistic psychological constraints, they cannot claim to have any relevance for cognitive scientists. Therefore, AI alone cannot be the method for cognitive science. – The simple response is to do just that! Green’s response: I am in basic accord with their conclusion that “AI ALONE cannot be the method for cognitive science”… I simply think there are some basic problems in psychology that programming is not equipped to solve…

9 Harnad’s Argument for a Convergence between Cognitive Science and AI Harnad, S. (2000). The Convergence Argument in Mind-Modelling. Psycoloquy: 100 (078) There are multiple ways to model small, arbitrary subsets of human functional capacity. Arbitrarily many ways to capture – Calculating skills – Chess playing skills – Scene describing skills “Many ways to skin a cat”

10 Harnad’s Argument for a Convergence between Cognitive Science and AI There are fewer and fewer ways to capture all these skills in the same system – The degrees of freedom for skinning ALL possible cats with the SAME resources are much narrower than those for skinning just one with ANY resources As we scale up from toy tasks to our full performance capacity…the degree of underdetermination of our models will shrink to the normal levels of underdetermination of scientific theories by empirical data

11 Relevance for this Workshop If Harnad is right, then AI and Cognitive Science converge on Hard Problems AI + Hard Problems + White Box = Cognitive Science Human-Centric HLI (human-level intelligence) Computational

12 Relevance for this Workshop NLP is the quintessential AI Hard Problem But Harnad suggests that the Turing Test which is a purely symbol manipulation NLP task, isn’t hard enough – It’s too underdetermined, allowing for multiple possible solutions – It’s ungrounded Harnad’s Total Turing Test requires a robot to interact with an environment and communicate with humans over the timescale of a human life – The symbolic representations the system acquires become grounded in experience – Robot Functionalism

13 My Big Theoretical Claim Ungrounded symbol systems are inadequate to represent the meaning of linguistic expressions – “pilot” does not mean PILOT – “pilot” does not mean pilot_n_1 – “pilot” means something like: Language must be grounded in non-linguistic representations of the objects, events and states of affairs the language describes and/or references (as interpreted by our visual system)

14 My Big Theoretical Claim “pilot” PILOT Real WorldMental Box Real World perception LOT The Wrong ApproachA Better Approach grounding perception

15 My Big Theoretical Claim “pilot” PILOT Real WorldMental Box Real World perception LOT The Wrong ApproachA Better Approach grounding perception

16 My Big Theoretical Claim Full-scale NLP needs perceptual grounding whether or not NLP systems must actually function in the world Barsalou’s Perceptual Symbol Systems hypothesis provides a cognitive basis for grounding linguistic expressions Zwann’s Situation Model experiments show close interaction between language and experienced world We may not need the Total Turing Test—Robot Functionalism—for AI and Cognitive Science to converge We do need to agree to apply Cognitive Science principles to the solution of AI Hard Problems like NLP

17 Practical Arguments for Looking Inside the Black Box

18 Sentences (Normal) Humans Can’t Process (Normally) The horse raced past the barn fell – Many normal humans can’t make sense of this sentence – Humans don’t appear to use exhaustive search and algorithmic backtracking to understand Garden Path sentences The mouse the cat the dog chased bit ate the cheese – Humans are unable to understand multiply center embedded sentences despite the fact that a simple stack mechanism makes them easy for parsers – Humans are very bad at processing truly recursive structures While Mary dressed the baby spit up on the bed – 40% of humans conclude in a post test that Mary dressed the baby – Humans often can’t ignore locally coherent meanings, despite their global incoherence (despite the claims of Chomsky and collaborators)

19 Sentences (Normal) Humans Can Process (Normally) The horse that was ridden past the barn, fell – Given enough grammatical cues, humans have little difficulty making sense of linguistic expressions The dog chased the cat, that bit the mouse, that ate the cheese – Humans can process “right embedded” expressions – These sentences appear to be processed iteratively rather than recursively – AI/Computer Science provides a model for coverting recursive processes into iterative processes which require less memory – Inability of humans to process recursive structures is likely due to short-term working memory limitations Does not appear to be a perceptual limitation!

20 Practical Arguments for Looking Inside the Black Box Impact of Short-Term Working Memory (STWM) Limitations Lack of evidence for Algorithmic Backtracking in humans Lots of evidence that human procedural skills are directional in nature – The direction is forward Inability of humans to process recursive structures Inability of humans to ignore locally coherent structures – Humans can’t retract knowledge Lots of evidence that humans combine high-level symbolic knowledge and serial processing with low-level subsymbolic (statistical) knowledge and massively parallel processing – Symbolic Focus of Attention and STWM – Subsymbolic (statistical) Spreading Activation – Does this help solve the Frame Problem?

21 My Big Practical Claim Paying attention to cognitive science principles may actually facilitate, not hinder, the development of functional NLP systems Full-scale NLP systems which fail to consider human language representation and processing capabilities in sufficient detail are unlikely to be successful

22 Speech Recognition a Counter Example? Current speech recognition systems have achieved phenomenal success Word error rates have been reduced an average of 10% per year for the last decade, and improvements are likely to continue Systems use Hidden Markov Models (HMMs) and algorithms like Viterbi No claims of cognitive plausibility for the mechanisms being used

23 Speech Recognition a Counter Example? Current performance of speech recognition systems is still well below human performance in large vocabulary, speaker independent recognition in noisy environments Some researchers think performance is asymptoting Cognitive scientists claim that various assumptions limit the ultimate performance of such systems – Acoustic model views speech as a sequence of phones – Language models extremely limited Simple Bi-gram or Tri-gram cooccurrence Finite state grammars must be fully expanded to integrate with acoustic model Yet to be seen if such systems will eventually attain (or exceed) human capabilities

24 Speech Recognition a Counter Example? Within Cognitive Science, efforts are underway to improve performance of speech recognition systems by adopting cognitive principles – Adding syllables and other perceptually salient units – Integrating higher-level linguistic knowledge Efforts are also underway to integrate speech recognition front- ends into theoretically motivated computational systems which heretofore overlooked the raw acoustic signal Currently, performance is well below that of AI systems Within AI, efforts are also underway to add higher-level language capabilities – Microsoft wants to combine their speech recognition system with their NLP system Unfortunately, the NLP system processes input right to left!

25 Speech Recognition Systems Viewed Cognitively Highly interactive – Recent psycholinguistic evidence supports interactivity Probabilistic with discrete symbols – Hybrid symbolic, probabilistic systems are now the norm Feedforward only – No feedback loops: some cognitive scientists argue that feedback can’t improve perception Systems sum evidence without competition – Not clear (to me) what competition gives you Beam search limits number of competing alternatives – Integration of parallel-like processing and memory limitations – Not cognitively plausible in its current implementation

26 Wrap-Up

27 Converging AI and Computational Cognitive Science Cognitive Architectures – Development environment of choice for building computational cognitive models – Need to scale up these architectures DARPA BICA is leading the way – Need to build larger-scale models less tied to small empirical data sets – Need to get AI researchers interested in using Cognitive Architectures Reusable components Better development tools Need more symposiums like this one and more opportunities for publishing research

28 Conclusions Full-Scale, functional NLP Systems developed as Cognitive Black Boxes are unlikely to be successful Human capabilities and limitations are too important to be ignored in this highly human-centric domain Need to consider the internals of how humans process language to some level of abstraction below input/output behavior What that level is, is an important topic for discussion… – Neuronal – I hope not! – Hybrid Symbolic/Subsymbolic – My bets are here! – Symbolic – Already tried that!

29 The End

30 Arguments From Connectionism Parallel Distributed Processing (PDP) approaches look inside the black box PDP approaches have highlighted many of the shortcoming of symbolic AI arguing that these shortcomings result from an inappropriate cognitive architecture PDP systems attempt to specify what a cognitive system would be composed of at a level of abstraction somewhere above the neuronal level, but definitely inside the black box PDP systems capture important elements of perception and cognition Many AI systems are now hybrid symbolic/subsymbolic systems

31 Arguments For NLP systems must deal with – Noisy input – Lexical and Grammatical Ambiguity – Non-Literal use of language These aspects call out for adoption of techniques explored in connectionist and statistical approaches – Latent Semantic Analysis (LSA)

32 Latent Semantic Analysis (LSA) Statistical technique for determining similarity of the meaning of words – Based on co-occurrence of words in texts – Submitted to Singular Value Decomposition (SVD) – Identifies latent semantic relationships between words, even words which do not co-occur together Possibility of dealing with intractable problems in meaning representation Determine similarity of meaning without requiring discrete word senses If successful, an avalanche of NLP research on word sense disambiguation will need to be revisited

33 WordNet Psycholinguistically motivated network of word associations Words grouped into synonymn sets (synsets) Various associations between synsets identified – Hypernymns (type – subtype) – Meronymns (part – whole) AI researchers using WordNet as a resource for NLP without buying into its psycholinguistic underpinnings

34 Autonomy of Symbolism Non-linguistic representations are analogical representations of experience They are symbols in the sense that they are in the brain, not the external world Wilks’ “Autonomy of Symbolism” is adopted in this sense

35 My Big Claim Attempts to solve AI hard problems without applying Cognitive Science principles are likely to fail – Especially in AI systems that mimic human cognitive capabilities – Especially in full-scale NLP systems

36 Practical Implications Language generation systems should avoid producing linguistic expressions which humans will have difficulty understanding One way of achieving this is not to rely on processing mechanisms like stacks, recursion and algorithmic backtracking which humans fail to show evidence for