LING 388: Language and Computers Sandiway Fong Lecture 22: 11/10.

Slides:



Advertisements
Similar presentations
Michael W. Smith & Jeffrey D. Wilhelm
Advertisements

TEST-TAKING STRATEGIES FOR THE OHIO ACHIEVEMENT READING ASSESSMENT
Intro to Linguistics Class # 2 Chapter 1: What is Language?
Psych 156A/ Ling 150: Acquisition of Language II Lecture 12 Poverty of the Stimulus I.
AVRAM NOAM CHOMSKY Biography Criticisms and problems
Lytle ISD 17 years LJH = 10 years CLL = 7 years San Antonio Writing Project TC since 2007 Write For Texas Group.
Semantics Semantics is the branch of linguistics that deals with the study of meaning, changes in meaning, and the principles that govern the relationship.
Introduction: The Chomskian Perspective on Language Study.
Week 3a. UG and L2A: Background, principles, parameters CAS LX 400 Second Language Acquisition.
Section 4: Language and Intelligence Overview Instructor: Sandiway Fong Department of Linguistics Department of Computer Science.
Linguistic Theory Lecture 8 Meaning and Grammar. A brief history In classical and traditional grammar not much distinction was made between grammar and.
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
August 23, 2010 Grammars and Lexicons How do linguists study grammar?
LING 388: Language and Computers Sandiway Fong Lecture 28: 12/6.
Computational Intelligence 696i Language Lecture 3 Sandiway Fong.
LING 438/538 Computational Linguistics Sandiway Fong Lecture 20: 11/8.
Computational Intelligence 696i Language Lecture 2 Sandiway Fong.
LING 388 Language and Computers Take-Home Final Examination 12/9/03 Sandiway FONG.
IGERT External Advisory Board Meeting Wednesday, March 14, 2007 INSTITUTE FOR COGNITIVE SCIENCES University of Pennsylvania.
Computational Intelligence 696i Language Lecture 4 Sandiway Fong.
1 Introduction to Computational Natural Language Learning Linguistics (Under: Topics in Natural Language Processing ) Computer Science (Under:
LING 581: Advanced Computational Linguistics Lecture Notes January 12th.
LING 388: Language and Computers Sandiway Fong Lecture 1: 8/23.
LING 438/538 Computational Linguistics Sandiway Fong Lecture 19: 10/31.
Meaning and Language Part 1.
TRANSFORMATIONAL GRAMMAR An introduction. LINGUISTICS Linguistics Traditional Before 1930 Structural 40s -50s Transformational ((Chomsky 1957.
Syntax.
Lecture 1 Introduction: Linguistic Theory and Theories
Linguistic Theory Lecture 2 Phrase Structure. What was there before structure? Classical studies: Classical studies: –Languages such as Latin Rich morphology.
Linguistic Theory Lecture 3 Movement. A brief history of movement Movements as ‘special rules’ proposed to capture facts that phrase structure rules cannot.
LING/C SC/PSYC 438/538 Lecture 2 Sandiway Fong. Today’s Topics Did you read Chapter 1 of JM? – Short Homework 1 (submit by midnight Saturday) Some slides.
Standards  Writing  1.0 Writing Strategies: Students write coherent and focused texts that convey a well-defined perspective and tightly reasoned argument.
1 LIN 1310B Introduction to Linguistics Prof: Nikolay Slavkov TA: Qinghua Tang CLASS 14, Feb 27, 2007.
LING/C SC/PSYC 438/538 Lecture 23 Sandiway Fong. Administrivia Homework 4 – out today – due next Wednesday – (recommend you attempt it early) Reading.
LING 388: Language and Computers Sandiway Fong Lecture 27.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
1 Words and rules Linguistics lecture #2 October 31, 2006.
1 What does “meaning” mean? Linguistics lecture #3 November 2, 2006.
October 15, 2007 Non-finite clauses and control : Grammars and Lexicons Lori Levin.
Levels of Language 6 Levels of Language. Levels of Language Aspect of language are often referred to as 'language levels'. To look carefully at language.
Artificial Intelligence: Natural Language
Semantic Construction lecture 2. Semantic Construction Is there a systematic way of constructing semantic representation from a sentence of English? This.
CSA2050 Introduction to Computational Linguistics Lecture 1 What is Computational Linguistics?
LING/C SC/PSYC 438/538 Lecture 22 Sandiway Fong. Last Time Gentle introduction to probability Important notions: –sample space –events –rule of counting.
Rules, Movement, Ambiguity
Linguistic Anthropology Bringing Back the Brain. What Bloomfield Got “Right” Emphasized spoken language rather than written language The role of the linguist.
Introduction Chapter 1 Foundations of statistical natural language processing.
Topic and the Representation of Discourse Content
Programming Errors. Errors of different types Syntax errors – easiest to fix, found by compiler or interpreter Semantic errors – logic errors, found by.
Sight Words.
The Writing Process Basic Sentence Structure Complete Sentences Types of Sentences Fragments Run-Ons Paragraphs Elements of a Paragraph Outlining.
SYNTAX.
1 Some English Constructions Transformational Framework October 2, 2012 Lecture 7.
Further criticisms of Concept Empiricism Focus: To consider further criticisms of Concept Empiricism, alongside the criticism from Innatism.
English II—February 4, 2015 Bell work: Have you ever visited a new place? How did it make you feel? Was it what you expected? Homework: – Study Island.
“The pen is the tongue of the mind.”
End of the beginning Let’s wrap up some details and be sure we are all on the same page Good way to make friends and be popular.
Learning Through Failure. Reflect O Take a few moments to write down your answers to the following questions: O What was your reaction to the video? O.
Syntax By WJQ. Syntax : Syntax is the study of the rules governing the way words are combined to form sentences in a language, or simply, the study of.
PHILOSOPHY OF LANGUAGE Some topics and historical issues of the 20 th century.
LING/C SC 581: Advanced Computational Linguistics Lecture Notes Feb 3 rd.
Universal Grammar Chomsky and his followers no longer use the term LAD, but refer to the child’s innate endowment as Universal Grammar (UG). UG is a theory.
Chapter 11 Language. Some Questions to Consider How do we understand individual words, and how are words combined to create sentences? How can we understand.
Chapter 3 Language Acquisition: A Linguistic Treatment Jang, HaYoung Biointelligence Laborotary Seoul National University.
“They Say, I Say” How to enter into an argument. “…to give writing the most important thing of all -- namely, a point -- a writer needs to indicate clearly.
Lecture 3: Functional Phrases
SYNTAX.
Introduction to Linguistics
LING/C SC 581: Advanced Computational Linguistics
Traditional Grammar VS. Generative Grammar
Presentation transcript:

LING 388: Language and Computers Sandiway Fong Lecture 22: 11/10

Last Time Homework 4: Translator –today: some time-saving hints for using the debugger

Prolog Debugger Hints sometimes, tracing step-by-step is both tedious and unnecessary you can make use of other tracing commands see the HELP system example –?- debug. puts the Prolog interpreter into debug (but not step-by-step tracing) mode –?- spy predicate/n. –e.g. ?- spy sbar/3. tells Prolog (in debug mode) to stop at sbar/3 –debug command l (Leap) tells Prolog to continue (non-step-by-step) execution

Prolog Debugger Hints example | ?- debug. | ?- spy mapPA/2. | ?- translate(X,[taroo,ga,hon,o,katta]). silently runs sbar/3 generating English predicate-argument structure Call: mapPA(buy(who,who),_689) ? l ? Exit: mapPA(buy(who,who),katta(dare,dare)) ? l Redo: mapPA(buy(who,who),katta(dare,dare)) ? l Fail: mapPA(buy(who,who),_689) ? l Call: mapPA(buy(what,who),_689) ? l ? Exit: mapPA(buy(what,who),katta(nani,dare)) ? l Redo: mapPA(buy(what,who),katta(nani,dare)) ? l Fail: mapPA(buy(what,who),_689) ? l Call: mapPA(buy(john,who),_689) ? l ? Exit: mapPA(buy(john,who),katta(taroo,dare)) ? l Redo: mapPA(buy(john,who),katta(taroo,dare)) ? l Fail: mapPA(buy(john,who),_689) ? l Call: mapPA(buy(bucket,who),_689) ? l Fail: mapPA(buy(bucket,who),_689) ?

Moving on to Other Topics Natural Language Processing is a very broad field –the next area we’ll be exploring will be word semantics –but first a brief look at the inherent complexity of human language

Computational Implementation good news! practical knowledge –you now have practical knowledge on how to write grammars to account for various syntactic phenomena e.g. we’ve covered... –passivization, progressive aspect, determiner-noun/subject-verb agreement, wh-questions, yes/no-questions, different word orders etc. at this point in the course –you have all the basic tools and grammar programming techniques for implementing almost anything in the sense –you could (in principle) implement all the constructions people can document questions –what would that grammar look like? –is it a good model?

The Puzzle of Language the bad news language is a complex system –in terms of shades of meaning –in terms of the syntax –in terms of what is allowed and what is not language is part of a generative system –you can compose constructions and create new sentences –people can have razor-sharp judgments about data they have never encountered before –not just in terms of grammaticality/ungrammaticality –but also in terms of semantic interpretation

The Puzzle of Language compositionality of constructions –active: the militia arrested John –passive: John was arrested –simple: John is sad –raising: John seems to be sad –raising+passive: John seems to have been arrested –*passive+raising: *John was seemed to be arrested

The Puzzle of Language what’s allowed and what’s not –subject relative clause: the man that knows me (is not a liar) –object relative clause: the man that I know (is not a liar) omission of the relative pronoun (that) –subject relative clause: *the man knows me (is not a liar) –object relative clause: the man I know (is not a liar) why? –is it an arbitrary rule? Who came up with it? Why do we all agree?

The Puzzle of Language (The King’s English: Fowler 1908) –the omission of the relative in isolated clauses (as opposed to coordinates) is a question not of correctness but of taste, so far as there is any question at all. [...] –the omission of a defining relative subject is often effective in verse, but in prose is either an archaism or a provincialism. It may, moreover, result in obscurity... Now it would be some fresh insect won its way to a temporary fatal new development. (H. G. Wells) –but when the defining relative is object, or has a preposition, there is no limit to the omission...

The Puzzle of Language 2nd language learners of English worry about these rules a lot –this is the student did it –‘zero’-subject relatives common in Hong Kong English (Gisborne 2000) –this is the student who did it

The Puzzle of Language for semantics –we’re not just talking about (famous) sentences like colorless green ideas sleep furiously (Chomsky 1957) –but also many sentences for which we take the rules of interpretation for granted perhaps suggests that we’re operating with rules or principles which we’re not conscious or aware of

An Example consider the object wh-question –Which report did you file without reading? 1

An Example the object wh-question –Which report did you file without reading? is actually a pretty complicated sentence for any computer grammar writer to deal with let’s look at one problem necessary for interpretation gap-filling –file is a verb, there is a filer and something being filed file(filer,something) –the thing being filed is the report in question file(filer,[which report])

An Example consider the object wh-question –Which report did you file without reading? also –read is a verb, there is a reader and something being read read(reader,something) also implicit –the reader must be the same person referred to by the pronoun you file(filer[you],[which report]) & read(reader,something) & filer=reader –the thing being read must be the same thing being filed, which must be the report in question file(filer[you],[which report]) & read(reader,something) & filer=reader & something = [which report] there are no other possible interpretations – in this case

An Example consider the object wh-question –Which report did you file without reading? there are no other possible interpretations –meaning (for example) that we cannot be asking about some report that you filed but someone else read

An Example –Which report did you file without reading? so the only interpretation is –Which report did you file [the report] without [you] reading [the report]? can be viewed as a form of “compression” –Which report did you file [the report] without [you] reading [the report]? in other words –there is an understanding between speaker and hearer that the hearer can decode and recover the missing bits because they share the same “grammar” recall –the grammar with traces in earlier lectures...

An Example –Which report did you file without reading? only interpretation is –Which report did you file [the report] without [you] reading [the report]? a computer program has to know the rules of gap filling –(for this so-called parasitic gap sentence) questions –what are the rules of gap filling? –were you taught these rules in school? –say you wanted to implement them, can you find them in a grammar book?

An Example rules of gap filling must account for –Which report did you file without reading? –*Which book did you file the report without reading –imagine: you file the book without reading which book –*The report was filed without reading –*The report was filed after Bill read –compare with: The report was filed without being read –These papers are easy to file without reading –This book is not worth reading without attempting to analyze deeply given partial data –can you be confident of coming up with the right rules or mechanism?

The “Rules” what do the rules look like? –surely not Prolog DCG are we sure we covered all the cases? how about –*Who left without insulting? –who left without insulting John? debate –How come “everyone” acquired the same rule systems? –Are these rules innate knowledge or learnt?

The “Rules” How is the knowledge of language acquired? From (Chomsky 1986) Standard belief 30+ years ago –language acquisition is a case of “overlearning” –language is a habit system assumed to be overdetermined by available evidence Plato’s Problem –the problem of “poverty of stimulus” –accounting for the richness, complexity and specificity of shared knowledge given the limitations of the data available –poverty of evidence

The “Rules” idea then that –we’re pre-wired to learn language –data like the sentences we’ve been looking at are (in part) determined by the architecture and machinery of the language faculty –we’re not acquiring these rules from scratch –the pre-wiring is part of our genetic endowment –reasonable to assume what is pre-wired must be universal –if so, the pre-wiring must be flexible enough to account for language variation –yet reduce the learning burden

The “Rules” Minimalist Program (MP) current linguistic technology (research area) language is a computational system even fewer mechanisms Principles-and-Parameters Framework (GB) reduction of construction rules to fundamental principles (the atoms of theory) explanatory adequacy Rule-based systems construction-based monostratal, e.g. context-free grammars multiple levels. e.g. transformational grammars

example –colorless green ideas sleep furiously First hit Interesting things to Google

example –colorless green ideas sleep furiously first hit –compositional semantics – a green idea is, according to well established usage of the word "green" is one that is an idea that is new and untried. –again, a colorless idea is one without vividness, dull and unexciting. –so it follows that a colorless green idea is a new, untried idea that is without vividness, dull and unexciting. –to sleep is, among other things, is to be in a state of dormancy or inactivity, or in a state of unconsciousness. –to sleep furiously may seem a puzzling turn of phrase but one reflects that the mind in sleep often indeed moves furiously with ideas and images flickering in and out.

Interesting things to Google example –colorless green ideas sleep furiously another hit: (a story) –"So this is our ranking system," said Chomsky. "As you can see, the highest rank is yellow." –"And the new ideas?" –"The green ones? Oh, the green ones don't get a color until they've had some seasoning. These ones, anyway, are still too angry. Even when they're asleep, they're furious. We've had to kick them out of the dormitories - they're just unmanageable." –"So where are they?" –"Look," said Chomsky, and pointed out of the window. There below, on the lawn, the colorless green ideas slept, furiously.

Interesting things to Google examples –(1) colorless green ideas sleep furiously –(2) furiously sleep ideas green colorless Chomsky (1957): –... It is fair to assume that neither sentence (1) nor (2) (nor indeed any part of these sentences) has ever occurred in an English discourse. Hence, in any statistical model for grammaticalness, these sentences will be ruled out on identical grounds as equally `remote' from English. Yet (1), though nonsensical, is grammatical, while (2) is not. idea –(1) is syntactically valid, (2) is word salad Statistical Experiment (Pereira 2002)

Interesting things to Google examples –(1) colorless green ideas sleep furiously –(2) furiously sleep ideas green colorless Statistical Experiment (Pereira 2002) wiwi w i-1 bigram language model